A quick lookup for integrating and interpreting metrics in your workflows
Metric | Enum Value |
---|---|
Action Advancement | GalileoScorers.action_advancement |
Action Completion | GalileoScorers.action_completion |
BLEU | GalileoScorers.bleu |
Chunk Attribution | GalileoScorers.chunk_attribution_utilization |
Chunk Utilization | GalileoScorers.chunk_attribution_utilization |
Completeness | GalileoScorers.completeness |
Context Adherence | GalileoScorers.context_adherence |
Context Relevance (Query Adherence) | GalileoScorers.context_relevance |
Correctness (factuality) | GalileoScorers.correctness |
Ground Truth Adherence | GalileoScorers.ground_truth_adherence |
Instruction Adherence | GalileoScorers.instruction_adherence |
Prompt Injection | GalileoScorers.prompt_injection |
Prompt Perplexity | GalileoScorers.prompt_perplexity |
ROUGE | GalileoScorers.rouge |
Sexism / Bias | GalileoScorers.input_sexism , GalileoScorers.output_sexism |
Tone | GalileoScorers.input_tone , GalileoScorers.output_tone |
Tool Errors | GalileoScorers.tool_error_rate |
Tool Selection Quality | GalileoScorers.tool_selection_quality |
Toxicity | GalileoScorers.input_toxicity , GalileoScorers.output_toxicity |
Metric | Enum Value |
---|---|
Action Advancement | GalileoScorers.action_advancement_luna |
Action Completion | GalileoScorers.action_completion_luna |
Chunk Attribution | GalileoScorers.chunk_attribution_utilization_luna |
Chunk Utilization | GalileoScorers.chunk_attribution_utilization_luna |
Completeness | GalileoScorers.completeness_luna |
Context Adherence | GalileoScorers.context_adherence_luna |
Prompt Injection | GalileoScorers.prompt_injection_luna |
Sexism / Bias | GalileoScorers.input_sexism_luna , GalileoScorers.output_sexism_luna |
Tone | GalileoScorers.input_tone_luna , GalileoScorers.output_tone_luna |
Tool Errors | GalileoScorers.tool_error_rate_luna |
Tool Selection Quality | GalileoScorers.tool_selection_quality_luna |
Toxicity | GalileoScorers.input_toxicity_luna , GalileoScorers.output_toxicity_luna |
run experiment
function (Python, TypeScript) takes a list of metrics as part of its arguments.
run_experiment
function as shown below:
run experiment
, like in the example below.My custom metric
.