Skip to content

Commit b15e610

Browse files
authored
Some fixes that slipped through (#717)
* fix the last missing original task flags * fix missing metrics and answer_choices
1 parent 3614e44 commit b15e610

File tree

11 files changed

+73
-37
lines changed

11 files changed

+73
-37
lines changed

promptsource/templates/app_reviews/templates.yaml

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,8 +30,10 @@ templates:
3030
{{review}}'
3131
metadata: !TemplateMetadata
3232
choices_in_prompt: null
33-
metrics: []
34-
original_task: null
33+
metrics:
34+
- Accuracy
35+
- Spearman Correlation
36+
original_task: false
3537
name: generate_review
3638
reference: Generate a review from the rating.
3739
9746ce4b-ac58-4dfb-9783-d77c95cb62cf: !Template

promptsource/templates/dbpedia_14/templates.yaml

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,8 @@ templates:
3030
{{title}} - {{content}} ||| {{ answer_choices[label] }}
3131
metadata: !TemplateMetadata
3232
choices_in_prompt: true
33-
metrics: []
33+
metrics:
34+
- Accuracy
3435
original_task: true
3536
name: pick_one_category_for_the_following_text
3637
reference: ''
@@ -45,7 +46,8 @@ templates:
4546
refers to which one? ||| {{ answer_choices[label] }}'
4647
metadata: !TemplateMetadata
4748
choices_in_prompt: true
48-
metrics: []
49+
metrics:
50+
- Accuracy
4951
original_task: true
5052
name: 'given_a_choice_of_categories '
5153
reference: ''

promptsource/templates/glue/ax/templates.yaml

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -68,7 +68,8 @@ templates:
6868
{{answer_choices[label]}}'
6969
metadata: !TemplateMetadata
7070
choices_in_prompt: true
71-
metrics: []
71+
metrics:
72+
- Accuracy
7273
original_task: true
7374
name: does_this_imply
7475
reference: ''
@@ -90,7 +91,8 @@ templates:
9091
{{answer_choices[label]}}'
9192
metadata: !TemplateMetadata
9293
choices_in_prompt: true
93-
metrics: []
94+
metrics:
95+
- Accuracy
9496
original_task: true
9597
name: does_this_support
9698
reference: ''

promptsource/templates/jigsaw_unintended_bias/templates.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -131,7 +131,7 @@ templates:
131131
choices_in_prompt: null
132132
metrics:
133133
- Other
134-
original_task: null
134+
original_task: true
135135
name: threatening
136136
reference: ''
137137
71508d0b-7fd9-4810-933c-c3c59f0c5329: !Template

promptsource/templates/mdd/task2_recs/templates.yaml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,8 @@ templates:
4343
{{dialogue_turns.utterance[1]}}'
4444
metadata: !TemplateMetadata
4545
choices_in_prompt: false
46-
metrics: []
46+
metrics:
47+
- Other
4748
original_task: false
4849
name: recommend_movies_dialogue
4950
reference: Given the likes, recommend a movie as a dialogue

promptsource/templates/scitail/snli_format/templates.yaml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,8 @@ templates:
1919
{% endif %}'
2020
metadata: !TemplateMetadata
2121
choices_in_prompt: true
22-
metrics: []
22+
metrics:
23+
- Accuracy
2324
original_task: true
2425
name: Another Yes/No Entailment Framing
2526
reference: ''

promptsource/templates/tydiqa/primary_task/templates.yaml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -150,7 +150,8 @@ templates:
150150
\ | capitalize}}\n {% endif %}\n{% endif %}"
151151
metadata: !TemplateMetadata
152152
choices_in_prompt: false
153-
metrics: []
153+
metrics:
154+
- Accuracy
154155
original_task: true
155156
name: based_on_the_text
156157
reference: Binary question without mentioning KB

promptsource/templates/tydiqa/secondary_task/templates.yaml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,9 @@ templates:
5151
{% endif %}'
5252
metadata: !TemplateMetadata
5353
choices_in_prompt: null
54-
metrics: []
54+
metrics:
55+
- BLEU
56+
- ROUGE
5557
original_task: false
5658
name: title_generation
5759
reference: Generate title from a passage

promptsource/templates/wiki_bio/templates.yaml

Lines changed: 12 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,10 @@ templates:
2222
{{target_text}}'
2323
metadata: !TemplateMetadata
2424
choices_in_prompt: false
25-
metrics: []
25+
metrics:
26+
- BLEU
27+
- ROUGE
28+
- Other
2629
original_task: true
2730
name: who
2831
reference: ''
@@ -40,7 +43,8 @@ templates:
4043
content\"][n] }}\n{% endif %}\n{% endfor %}\n"
4144
metadata: !TemplateMetadata
4245
choices_in_prompt: null
43-
metrics: []
46+
metrics:
47+
- Other
4448
original_task: false
4549
name: comprehension
4650
reference: ''
@@ -54,7 +58,8 @@ templates:
5458
_\",\" \") }} \n{% endif %}\n{% endfor %}"
5559
metadata: !TemplateMetadata
5660
choices_in_prompt: null
57-
metrics: []
61+
metrics:
62+
- Other
5863
original_task: false
5964
name: what_content
6065
reference: ''
@@ -81,7 +86,8 @@ templates:
8186
'
8287
metadata: !TemplateMetadata
8388
choices_in_prompt: null
84-
metrics: []
89+
metrics:
90+
- Other
8591
original_task: false
8692
name: guess_person
8793
reference: ''
@@ -106,7 +112,8 @@ templates:
106112
{% endfor %}'
107113
metadata: !TemplateMetadata
108114
choices_in_prompt: null
109-
metrics: []
115+
metrics:
116+
- Other
110117
original_task: false
111118
name: key_content
112119
reference: ''

promptsource/templates/wiki_hop/masked/templates.yaml

Lines changed: 15 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@ dataset: wiki_hop
22
subset: masked
33
templates:
44
08f2d1cf-c026-4b65-96d0-a28ff91affb5: !Template
5-
answer_choices: null
5+
answer_choices: '{{candidates | join("|||")}}'
66
id: 08f2d1cf-c026-4b65-96d0-a28ff91affb5
77
jinja: '{% set question_split = question.split('' '') %}
88
@@ -52,7 +52,8 @@ templates:
5252
{% endif %}'
5353
metadata: !TemplateMetadata
5454
choices_in_prompt: null
55-
metrics: []
55+
metrics:
56+
- Accuracy
5657
original_task: false
5758
name: Indirect Question about Birthplace / Citizenship / Place of Death
5859
reference: Ask about place of birth, citizenship, or place of death for the subject
@@ -75,13 +76,14 @@ templates:
7576
{{ question_split[0] | replace("_", " ") }}'
7677
metadata: !TemplateMetadata
7778
choices_in_prompt: null
78-
metrics: []
79+
metrics:
80+
- Other
7981
original_task: false
8082
name: Explain Relation
8183
reference: Given information, explain the relation between the subject entity
8284
and the object entity in a fact triple.
8385
3181f711-a376-4d6e-9fca-a34e1d048585: !Template
84-
answer_choices: null
86+
answer_choices: '{{candidates | join("|||")}}'
8587
id: 3181f711-a376-4d6e-9fca-a34e1d048585
8688
jinja: 'Information:
8789
@@ -99,20 +101,22 @@ templates:
99101
{{answer}}'
100102
metadata: !TemplateMetadata
101103
choices_in_prompt: null
102-
metrics: []
104+
metrics:
105+
- Accuracy
103106
original_task: false
104107
name: Generate Object Answer
105108
reference: Given information, generate the best object entity for the fact triple.
106109
639fa83f-14fd-457a-886e-a65334cb7e66: !Template
107-
answer_choices: null
110+
answer_choices: '{{candidates | join("|||")}}'
108111
id: 639fa83f-14fd-457a-886e-a65334cb7e66
109112
jinja: "Information:\n- {{ supports | join(\"\\n- \") }}\n\n{% set question_split\
110113
\ = question.split(' ') %}\nQuestion: ({{ question_split[1:] | join(\" \")}},\
111114
\ {{ question_split[0] | replace(\"_\", \" \") }}, ?)\n\nCandidate Answers:\
112115
\ \n- {{ candidates | join(\"\\n- \") }}\n|||\n{{answer}}"
113116
metadata: !TemplateMetadata
114117
choices_in_prompt: null
115-
metrics: []
118+
metrics:
119+
- Accuracy
116120
original_task: true
117121
name: Choose Best Object Candidate
118122
reference: Given information and possible object candidates, choose the best object
@@ -136,7 +140,8 @@ templates:
136140
") }}, {{answer}})'
137141
metadata: !TemplateMetadata
138142
choices_in_prompt: null
139-
metrics: []
143+
metrics:
144+
- Other
140145
original_task: false
141146
name: Generate Fact Triple
142147
reference: Given information, generate a fact triple.
@@ -158,7 +163,8 @@ templates:
158163
{{ question_split[1:] | join(" ")}}'
159164
metadata: !TemplateMetadata
160165
choices_in_prompt: null
161-
metrics: []
166+
metrics:
167+
- Other
162168
original_task: false
163169
name: Generate Subject Answer
164170
reference: Given information, generate the best subject entity for the fact triple.

0 commit comments

Comments
 (0)