A RetroSearch Logo

Home - News ( United States | United Kingdom | Italy | Germany ) - Football scores

Search Query:

Showing content from https://github.com/meta-llama/llama-stack-evals below:

meta-llama/llama-verifications: Functional tests and benchmarks for verifying Llama model providers.

This repository contains a lightweight library to verify model providers via 2 suites,

  1. Functional tests and
  2. Eval benchmarks

Both the suites can be run on any Llama Provider that offers an OpenAI-compliant Inference API.

Type of Verification Description Llama Provider Expectation Functional Tests Testing inference across:
100% pass rate Eval Benchmarks Academic benchmarks by category:
Similar numbers as Llama Model Card

For detailed results, see TESTS_REPORT.md and BENCHMARKS_REPORT.md.

We recommend using uv for this project. Install uv if you don't have it already. See uv for installation instructions.

git clone https://github.com/meta-llama/llama-verifications.git
cd llama-verifications
uv tool install --with-editable . --python 3.12 llama-verifications
uvx llama-verifications list-providers

A new provider can easily be added by adding a new yaml config over here in provider_confs.

List supported models for a Provider
uvx llama-verifications list-models <provider_id>
List Available Benchmarks
uvx llama-verifications list-benchmarks

Set environment variables for API Keys on different endpoints as required. Please note that some benchmarks may require an OpenAI API key.

export CEREBRAS_API_KEY=xxx
export GROQ_API_KEY=xxx
export FIREWORKS_API_KEY=fw_xxx
export LLAMA_API_KEY=xxx
export OPENAI_API_KEY=sk-xxx
export SAMBANOVA_API_KEY=xxx
export TOGETHER_API_KEY=xxx

You can run one or more eval benchmarks against a combination of a provider and model using this simple command.

uvx llama-verifications run-benchmarks \
   --benchmarks <benchmark-id> \
   --provider <provider-name-or-endpoint> \
   --model <model-id> \
   --num-examples 10

💡Pro-Tip: You can control parallelism via --max-parallel-generations 10

You can also run all benchmarks against a specific model or provider using the following script:

./scripts/run_all_benchmarks.sh <model-id> <provider-name-or-endpoint>
Generate Eval Benchmark Report

The latest report can be found at BENCHMARKS_REPORT.md. To update the report, ensure you have the API keys set.

uvx llama-verifications generate-benchmarks-report

The goal for eval benchmarks to get as close as possible (or higher!) to the numbers reported on the "Meta_reference" model card (column 3). Lower numbers mean the deployed model is underperforming on that benchmark, which means degradation in the corresponding capability. We assess the following core capabilities:

To run the functional tests, you will need to set the API keys for the different providers described above.

# to run all tests for a specific model / provider combo
uvx llama-verifications run-tests --model <model-id> --provider <provider>

# to run all tests for a provider across all supported models
uvx llama-verifications run-tests --provider <provider>

# to run all tests across all providers and their respective supported models
uvx llama-verifications run-tests
Generate Functional Test Report

The latest report can be found at TESTS_REPORT.md. To update the report, ensure you have the API keys set.

uvx llama-verifications generate-tests-report

Functional tests check if the inference is being performed properly across a variety of supported use cases, and are expected to pass 100% for a correct implementation.

The easiest way to report test and eval results is to make a new PR with the test and benchmark reports (as well as the provider YAML file, if applicable). This way both Meta teams and the broader community have a transparent way of assessing the provider model performance.

Adding a new provider and/or model

Add new providers and models here: /llama_verifications/provider_confs

To add new test cases, create appropriate YAML files in the /llama_verifications/functional_tests/openai_api/fixtures/test_cases directory following the existing patterns.

To add a new benchmark, you will need to define:

Llama Verifications is MIT licensed, as found in the LICENSE file.


RetroSearch is an open source project built by @garambo | Open a GitHub Issue

Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo

HTML: 3.2 | Encoding: UTF-8 | Version: 0.7.4