Question and answer for model evaluation in Amazon Bedrock
Question and answer is used for tasks including generating automatic help-desk responses, information retrieval, and e-learning. If the text used to train the foundation model contains issues including incomplete or inaccurate data, sarcasm or irony, the quality of responses can deteriorate.
Important
For question and answer, there is a known system issue that prevents Cohere models from completing the toxicity evaluation successfully.
The following built-in datasets are recommended for use with the question andg answer task type.
- BoolQ
-
BoolQ is a dataset consisting of yes/no question and answer pairs. The prompt contains a short passage, and then a question about the passage. This dataset is recommended for use with question and answer task type.
- Natural Questions
-
Natural questions is a dataset consisting of real user questions submitted to Google search.
- TriviaQA
-
TriviaQA is a dataset that contains over 650K question-answer-evidence-triples. This dataset is used in question and answer tasks.
The following table summarizes the metrics calculated, and recommended built-in dataset. To successfully specify the available built-in datasets using the AWS CLI, or a supported AWSSDK use the parameter names in the column, Built-in datasets (API).
Task type | Metric | Built-in datasets (console) | Built-in datasets (API) | Computed metric |
---|---|---|---|---|
Question and answer | Accuracy | BoolQ |
Builtin.BoolQ |
NLP-F1 |
NaturalQuestions |
Builtin.NaturalQuestions |
|||
TriviaQA |
Builtin.TriviaQa |
|||
Robustness | BoolQ |
Builtin.BoolQ |
F1 and deltaF1 |
|
NaturalQuestions |
Builtin.NaturalQuestions |
|||
TriviaQA |
Builtin.TriviaQa |
|||
Toxicity | BoolQ |
Builtin.BoolQ |
Toxicity | |
NaturalQuestions |
Builtin.NaturalQuestions |
|||
TriviaQA |
Builtin.TriviaQa |
To learn more about how the computed metric for each built-in dataset is calculated, see Review model evaluation job reports and metrics in Amazon Bedrock