forked from explodinggradients/ragas
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
feat: Factual correctnes metric (explodinggradients#1334)
- Loading branch information
1 parent
4676184
commit b845a62
Showing
2 changed files
with
420 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,150 @@ | ||
# Factual Correctness | ||
|
||
Factual correctness is a metric that compares and evaluates the factual accuracy of the generated `response` with the `reference`. This metric is used to determine the extent to which the generated response aligns with the reference. The factual correctness score ranges from 0 to 1, with higher values indicating better performance. To measure the alignment between the response and the reference, the metric uses the LLM for first break down the response and reference into claims and then uses natural language inference to determine the factual overlap between the response and the reference. Factual overlap is quantified using precision, recall, and F1 score, which can be controlled using the `mode` parameter. | ||
|
||
The formula for calculating True Positive (TP), False Positive (FP), and False Negative (FN) is as follows: | ||
|
||
```{math} | ||
\text{True Positive (TP)} = \text{Number of claims in response that are present in reference} | ||
``` | ||
|
||
```{math} | ||
\text{False Positive (FP)} = \text{Number of claims in response that are not present in reference} | ||
``` | ||
|
||
```{math} | ||
\text{False Negative (FN)} = \text{Number of claims in reference that are not present in response} | ||
``` | ||
|
||
The formula for calculating precision, recall, and F1 score is as follows: | ||
|
||
```{math} | ||
\text{Precision} = {TP \over (TP + FP)} | ||
``` | ||
|
||
```{math} | ||
\text{Recall} = {TP \over (TP + FN)} | ||
``` | ||
|
||
```{math} | ||
\text{F1 Score} = {2 \times \text{Precision} \times \text{Recall} \over (\text{Precision} + \text{Recall})} | ||
``` | ||
|
||
## Example | ||
|
||
```{code-block} python | ||
from ragas.dataset_schema import SingleTurnSample | ||
from ragas.metrics._factual_correctness import FactualCorrectness | ||
sample = SingleTurnSample( | ||
response="The Eiffel Tower is located in Paris.", | ||
reference="The Eiffel Tower is located in Paris. I has a height of 1000ft." | ||
) | ||
scorer = FactualCorrectness() | ||
scorer.llm = openai_model | ||
await scorer.single_turn_ascore(sample) | ||
``` | ||
|
||
By default, the mode is set to `F1`, you can change the mode to `precision` or `recall` by setting the `mode` parameter. | ||
|
||
```{code-block} python | ||
scorer = FactualCorrectness(mode="precision") | ||
``` | ||
|
||
## Controlling the Number of Claims | ||
|
||
Each sentence in the response and reference can be broken down into one or more claims. The number of claims that are generated from a single sentence is determined by the level of `atomicity` and `coverage` required for your application. | ||
|
||
### Understanding Atomicity and Coverage | ||
|
||
In claim decomposition, two important parameters influence the output: | ||
|
||
1. **Atomicity** | ||
2. **Coverage** | ||
|
||
These parameters help control the granularity and completeness of the generated claims. | ||
|
||
### Atomicity | ||
|
||
**Atomicity** refers to how much a sentence is broken down into its smallest, meaningful components. It can be adjusted based on whether you need highly detailed claims or a more consolidated view. | ||
|
||
- **High Atomicity**: The sentence is broken down into its fundamental, indivisible claims. This results in multiple, smaller claims, each representing a distinct piece of information. | ||
|
||
**Example:** | ||
- Original Sentence: | ||
- "Albert Einstein was a German theoretical physicist who developed the theory of relativity and contributed to quantum mechanics." | ||
- Decomposed Claims: | ||
- "Albert Einstein was a German theoretical physicist." | ||
- "Albert Einstein developed the theory of relativity." | ||
- "Albert Einstein contributed to quantum mechanics." | ||
|
||
- **Low Atomicity**: The sentence is kept more intact, resulting in fewer claims that may contain multiple pieces of information. | ||
|
||
**Example:** | ||
- Original Sentence: | ||
- "Albert Einstein was a German theoretical physicist who developed the theory of relativity and contributed to quantum mechanics." | ||
- Decomposed Claims: | ||
- "Albert Einstein was a German theoretical physicist who developed the theory of relativity and contributed to quantum mechanics." | ||
|
||
### Coverage | ||
|
||
**Coverage** refers to how comprehensively the claims represent the information in the original sentence. It can be adjusted to either include all details or to generalize the content. | ||
|
||
- **High Coverage**: The decomposed claims capture all the information present in the original sentence, preserving every detail. | ||
|
||
**Example:** | ||
- Original Sentence: | ||
- "Marie Curie was a Polish and naturalized-French physicist and chemist who conducted pioneering research on radioactivity." | ||
- Decomposed Claims: | ||
- "Marie Curie was a Polish physicist." | ||
- "Marie Curie was a naturalized-French physicist." | ||
- "Marie Curie was a chemist." | ||
- "Marie Curie conducted pioneering research on radioactivity." | ||
|
||
- **Low Coverage**: The decomposed claims cover only the main points, omitting some details to provide a more generalized view. | ||
|
||
**Example:** | ||
- Original Sentence: | ||
- "Marie Curie was a Polish and naturalized-French physicist and chemist who conducted pioneering research on radioactivity." | ||
- Decomposed Claims: | ||
- "Marie Curie was a physicist." | ||
- "Marie Curie conducted research on radioactivity." | ||
|
||
### Combining Atomicity and Coverage | ||
|
||
By adjusting both atomicity and coverage, you can customize the level of detail and completeness to meet the needs of your specific use case. | ||
|
||
- **High Atomicity & High Coverage**: Produces highly detailed and comprehensive claims that cover all aspects of the original sentence. | ||
|
||
**Example:** | ||
- Original Sentence: | ||
- "Charles Babbage was an English mathematician, philosopher, inventor, and mechanical engineer." | ||
- Decomposed Claims: | ||
- "Charles Babbage was an English mathematician." | ||
- "Charles Babbage was a philosopher." | ||
- "Charles Babbage was an inventor." | ||
- "Charles Babbage was a mechanical engineer." | ||
|
||
- **Low Atomicity & Low Coverage**: Produces fewer claims with less detail, summarizing the main idea without going into specifics. | ||
|
||
**Example:** | ||
- Original Sentence: | ||
- "Charles Babbage was an English mathematician, philosopher, inventor, and mechanical engineer." | ||
- Decomposed Claims: | ||
- "Charles Babbage was an English mathematician." | ||
- "Charles Babbage was an inventor." | ||
|
||
### Practical Application | ||
|
||
- Use **High Atomicity and High Coverage** when you need a detailed and comprehensive breakdown for in-depth analysis or information extraction. | ||
- Use **Low Atomicity and Low Coverage** when only the key information is necessary, such as for summarization. | ||
|
||
This flexibility in controlling the number of claims helps ensure that the information is presented at the right level of granularity for your application's requirements. | ||
|
||
## Example | ||
|
||
```{code-block} python | ||
scorer = FactualCorrectness(mode="precision",atomicity="low") | ||
``` |
Oops, something went wrong.