Skip to content

Wikiann Prompts #772

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 4 commits into
base: eval-hackathon
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 4 additions & 1 deletion promptsource/__init__.py
Original file line number Diff line number Diff line change
@@ -1 +1,4 @@
DEFAULT_PROMPTSOURCE_CACHE_HOME = "~/.cache/promptsource"
from pathlib import Path


DEFAULT_PROMPTSOURCE_CACHE_HOME = str(Path("~/.cache/promptsource").expanduser())
35 changes: 28 additions & 7 deletions promptsource/app.py
Original file line number Diff line number Diff line change
@@ -1,30 +1,37 @@
import argparse
import functools
import multiprocessing
import os
import textwrap
from hashlib import sha256
from multiprocessing import Manager, Pool

import pandas as pd
import plotly.express as px
import streamlit as st
from datasets import get_dataset_infos
from datasets.info import DatasetInfosDict
from pygments import highlight
from pygments.formatters import HtmlFormatter
from pygments.lexers import DjangoLexer
from templates import INCLUDED_USERS

from promptsource import DEFAULT_PROMPTSOURCE_CACHE_HOME
from promptsource.session import _get_state
from promptsource.templates import DatasetTemplates, Template, TemplateCollection
from promptsource.templates import INCLUDED_USERS, DatasetTemplates, Template, TemplateCollection
from promptsource.utils import (
get_dataset,
get_dataset_confs,
list_datasets,
removeHyphen,
renameDatasetColumn,
render_features,
linearizeDataset
)


DATASET_INFOS_CACHE_DIR = os.path.join(DEFAULT_PROMPTSOURCE_CACHE_HOME, "DATASET_INFOS")
os.makedirs(DATASET_INFOS_CACHE_DIR, exist_ok=True)

# Python 3.8 switched the default start method from fork to spawn. OS X also has
# some issues related to fork, eee, e.g., https://github.com/bigscience-workshop/promptsource/issues/572
# so we make sure we always use spawn for consistency
Expand All @@ -38,7 +45,17 @@ def get_infos(all_infos, d_name):
:param all_infos: multiprocess-safe dictionary
:param d_name: dataset name
"""
all_infos[d_name] = get_dataset_infos(d_name)
d_name_bytes = d_name.encode("utf-8")
d_name_hash = sha256(d_name_bytes)
foldername = os.path.join(DATASET_INFOS_CACHE_DIR, d_name_hash.hexdigest())
if os.path.isdir(foldername):
infos_dict = DatasetInfosDict.from_directory(foldername)
else:
infos = get_dataset_infos(d_name)
infos_dict = DatasetInfosDict(infos)
os.makedirs(foldername)
infos_dict.write_to_directory(foldername)
all_infos[d_name] = infos_dict


# add an argument for read-only
Expand Down Expand Up @@ -181,11 +198,13 @@ def show_text(t, width=WIDTH, with_markdown=False):
else:
subset_infos = infos[subset_name]

split_sizes = {k: v.num_examples for k, v in subset_infos.splits.items()}
try:
split_sizes = {k: v.num_examples for k, v in subset_infos.splits.items()}
except Exception:
# Fixing bug in some community datasets.
# For simplicity, just filling `split_sizes` with nothing, so the displayed split sizes will be 0.
split_sizes = {}
else:
# Zaid/coqa_expanded and Zaid/quac_expanded don't have dataset_infos.json
# so infos is an empty dic, and `infos[list(infos.keys())[0]]` raises an error
# For simplicity, just filling `split_sizes` with nothing, so the displayed split sizes will be 0.
split_sizes = {}

# Collect template counts, original task counts and names
Expand Down Expand Up @@ -299,6 +318,8 @@ def show_text(t, width=WIDTH, with_markdown=False):
split = st.sidebar.selectbox("Split", splits, key="split_select", index=index)
dataset = dataset[split]
dataset = renameDatasetColumn(dataset)
if dataset_key in ("wikiann"):
dataset = linearizeDataset(dataset)

#
# Loads template data
Expand Down
189 changes: 189 additions & 0 deletions promptsource/templates/wikiann/ace/templates.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,189 @@
dataset: wikiann
subset: ace
templates:
141c35d1-27fd-4f0a-937b-72e6db7cb7d7: !Template
answer_choices: person ||| organization ||| location
id: 141c35d1-27fd-4f0a-937b-72e6db7cb7d7
jinja: 'Given, "{{'' ''.join(tokens)}}", is {{spans[0].split('':'')[1]}} a person,
organization, or location?


|||

{% if spans[0][:3] == ''PER'' %}

person

{% elif spans[0][:3] == ''LOC'' %}

location

{%else%}

organization

{% endif %}'
metadata: !TemplateMetadata
choices_in_prompt: false
metrics:
- Accuracy
original_task: true
name: multiple_choice
reference: ''
18d8b432-c366-49f1-9851-b429fea19f87: !Template
answer_choices: Yes ||| No
id: 18d8b432-c366-49f1-9851-b429fea19f87
jinja: 'Given, "{{'' ''.join(tokens)}}", is {{spans[0].split('':'')[1]}} a person?


|||

{% if spans[0][:3] == ''PER'' %}

Yes

{% else %}

No

{% endif %}'
metadata: !TemplateMetadata
choices_in_prompt: false
metrics:
- Other
- AUC
- COQA F1
- Accuracy
original_task: true
name: is_person
reference: formatting version one to ask if the highlighted string is a person
2b52620e-d916-4b1b-8913-50453fb9e861: !Template
answer_choices: True || False
id: 2b52620e-d916-4b1b-8913-50453fb9e861
jinja: 'Given the following information:


{{'' ''.join(tokens)}}


True or false: is {{spans[0].split('':'')[1]}} a person?

|||

{{spans[0][:3] == "PER"}}'
metadata: !TemplateMetadata
choices_in_prompt: false
metrics:
- AUC
- BLEU
- COQA F1
- Accuracy
- Other
original_task: true
name: is_person_bool
reference: alt. is person question
3f3ddb73-b6e6-4cf1-b1f5-061edec3fcef: !Template
answer_choices: True ||| False
id: 3f3ddb73-b6e6-4cf1-b1f5-061edec3fcef
jinja: 'Given the following information:


{{'' ''.join(tokens)}}


True or false: is {{spans[0].split('':'')[1]}} a location?

|||

{{spans[0][:3] == "LOC"}}'
metadata: !TemplateMetadata
choices_in_prompt: false
metrics:
- AUC
- Accuracy
- COQA F1
- Other
original_task: true
name: is_location_bool
reference: ''
6a9f8e1f-a8c9-468d-8982-350b4e1fec14: !Template
answer_choices: Yes ||| No
id: 6a9f8e1f-a8c9-468d-8982-350b4e1fec14
jinja: 'Given, "{{'' ''.join(tokens)}}", is {{spans[0].split('':'')[1]}} a location?


|||

{% if spans[0][:3] == ''LOC'' %}

Yes

{% else %}

No

{% endif %}'
metadata: !TemplateMetadata
choices_in_prompt: false
metrics:
- AUC
- COQA F1
- Other
- Accuracy
original_task: true
name: is_location
reference: ''
7ad06be3-7499-4305-9090-6bea1b351807: !Template
answer_choices: Yes ||| No
id: 7ad06be3-7499-4305-9090-6bea1b351807
jinja: 'Given, "{{'' ''.join(tokens)}}", is {{spans[0].split('':'')[1]}} an organization?


|||

{% if spans[0][:3] == ''ORG'' %}

Yes

{% else %}

No

{% endif %}'
metadata: !TemplateMetadata
choices_in_prompt: false
metrics:
- AUC
- Accuracy
- COQA F1
- Other
original_task: true
name: is_organization
reference: ''
a1f1dde8-9a22-46b5-aa83-bb9fe57c4cf7: !Template
answer_choices: True || False
id: a1f1dde8-9a22-46b5-aa83-bb9fe57c4cf7
jinja: 'Given the following information:


{{'' ''.join(tokens)}}


True or false: is {{spans[0].split('':'')[1]}} an organization?


|||

{{spans[0][:3] == ''ORG''}}

'
metadata: !TemplateMetadata
choices_in_prompt: false
metrics:
- AUC
- COQA F1
- Accuracy
- Other
original_task: true
name: is_organization_bool
reference: ''
Loading