You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Apr 11, 2023. It is now read-only.
Update submission instructions, and fix predict.py
- Move hint for users following quickstart to top
- Get rid of example script for submitting custom model predictions, it
wouldn't have worked well anyway. We should make a separate simple
script that does this.
- More helpful text for finding the submit button in the UI
Use wandb internal API to avoid resuming, which breaks run duration and
config.
Copy file name to clipboardExpand all lines: BENCHMARK.md
+5-11Lines changed: 5 additions & 11 deletions
Original file line number
Diff line number
Diff line change
@@ -13,6 +13,8 @@ There are a few requirements for submitting a model to the benchmark.
13
13
14
14
### Submission format
15
15
16
+
*To submit from our baseline model, skip to the [training the baseline model](#training-the-baseline-model-optional) section below.*
17
+
16
18
A valid submission to the CodeSeachNet Challenge requires a file named **model_predictions.csv** with the following fields: `query`, `language`, `identifier`, and `url`:
17
19
18
20
*`query`: the textual representation of the query, e.g. "int to string" .
@@ -24,8 +26,6 @@ For further background and instructions on the submission process, see [the root
24
26
25
27
The row order corresponds to the result ranking in the search task. For example, if in row 5 there is an entry for the Python query "read properties file", and in row 60 another result for the Python query "read properties file", then the URL in row 5 is considered to be ranked higher than the URL in row 60 for that query and language.
26
28
27
-
The script we used to create the baseline submission is [src/predict.py](src/predict.py). You are not required to use this script to produce your submission file; we only provide it for reference.
28
-
29
29
Here is an example:
30
30
31
31
| query | language | identifier | url |
@@ -49,13 +49,7 @@ You can submit your results to the benchmark as follows:
49
49
2. Generate your own file of model predictions following the format above and name it \`model_predictions.csv\`.
50
50
3. Upload a run to wandb with this \`model_predictions.csv\` file attached.
51
51
52
-
Our example script [src/predict.py](src/predict.py) takes care of steps 2 and 3 for a model training run that has already been logged to W&B, given the corresponding W&B run id, which you can find on the /overview page in the browser or by clicking the 'info' icon on a given run.
53
-
54
-
Here is a short example script that will create a run in W&B and perform the upload (step 3) for a local file of predictions:
55
-
```python
56
-
import wandb
57
-
wandb.init(project="codesearchnet")
58
-
wandb.save('model_predictions.csv')
52
+
Our example script [src/predict.py](src/predict.py) takes care of steps 2 and 3 for a model whose training run has been logged to W&B, given the corresponding W&B run id, which you can find on the /overview page in the browser or by clicking the 'info' icon on a given run.
59
53
```
60
54
61
55
### Publishing your submission
@@ -65,7 +59,7 @@ You've now generated all the content required to submit a run to the CodeSearchN
65
59
You can submit your runs by visiting the run page and clicking on the overview tab:
@@ -75,6 +69,6 @@ Once you upload your \`model_predictions.csv\` file, W&B will compute the normal
75
69
76
70
### Training the baseline model (optional)
77
71
78
-
Replicating our results for the CodeSearchNet baseline is optional, as we encourage the community to create their own models and methods for ranking search results. To replicate our baseline submission, you can start with the instructions in the [CodeSearchNet GitHub repository](https://github.com/github/CodeSearchNet). This baseline model uses [src/predict.py](src/predict.py) to generate the submission file.
72
+
Replicating our results for the CodeSearchNet baseline is optional, as we encourage the community to create their own models and methods for ranking search results. To replicate our baseline submission, you can start with the "Quickstart" instructions in the [CodeSearchNet GitHub repository](https://github.com/github/CodeSearchNet). This baseline model uses [src/predict.py](src/predict.py) to generate the submission file.
79
73
80
74
Your run will be logged to W&B, within a project that will be automatically linked to this benchmark.
0 commit comments