Discover Galileo’s Luna-2 Evaluation model, reducing the latency and cost for metric evaluations
Model | Cost/1M token | Accuracy (F1 score) | Latency (avg) | Max tokens |
---|---|---|---|---|
Luna-2 | $0.02 | 0.88 | 152ms | 128k |
GPT 4o | $2.50 | 0.94 | 3,200ms | 128k |
GPT 4o mini | $0.02 | 0.90 | 2,600ms | 128k |
Azure Content Safety | $1.52 | 0.62 | 312ms | 3k |
Model | Small (500 tokens) | Medium (2K tokens) | Large (15K tokens) | Extra Large (100K tokens) |
---|---|---|---|---|
Luna 3B | 63ms | 249ms | 4.1s | 154s |
Luna 8B | 150ms | 580ms | 10.3s | 163s |
Model | Small (500 tokens) | Medium (2K tokens) | Large (15K tokens) | Extra Large (100K tokens) |
---|---|---|---|---|
Luna 3B | 14ms | 47ms | 564ms | 17.2s |
Luna 8B | 31ms | 115ms | 1.06s | 29s |
Model | Small (500 tokens) | Medium (2K tokens) | Large (15K tokens) | Extra Large (100K tokens) |
---|---|---|---|---|
Luna 3B | 19ms | 60ms | 601ms | 12.5s |
Luna 8B | 36ms | 138ms | 1.24s | 21.2s |