Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Create Evaluation Datasets #29

Closed
1 of 2 tasks
j2whiting opened this issue Apr 11, 2024 · 0 comments
Closed
1 of 2 tasks

Create Evaluation Datasets #29

j2whiting opened this issue Apr 11, 2024 · 0 comments
Assignees

Comments

@j2whiting
Copy link
Contributor

j2whiting commented Apr 11, 2024

  • Config from Document
  • Config from Dataset

We have created a config from document dataset with approximately 900 pairs of documents and AMRs. For evaluation, mask out the values for parameters and initials and then compare prediction against ground truth. Evaluate using precision, recall, f1. The dataset is uploaded to the shared drive.

Config from dataset strategy is TBD. Perhaps we can use this same dataset, and then map values from the existing AMRs into tabular format. Evaluate the model's ability to map the values in the tables back into the AMR.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants