Skip to content

NEU-Solution/evaluating_cluster

Repository files navigation

Evaluating Cluster

Architecture Diagrams

Architecture Diagram

Create code to evaluate

The evaluation dataset must be in the following format

[
    {
        "id":1,
        "question": "1 + 1 = ?",
        "choices" : [
            "A. 1",
            "B. 2",
            "C. 3",
            "D. 4"
        ],
        "answer": "D"
    }
]

Test

python -m test.test

Remenber add env file

WANDB_API_KEY=your_api_key
WANDB_PROJECT=mlops
WANDB_ENTITY=neu-solution

Run docker evaluation

# Build the Docker image

# evaluation function
docker build -t evaluate_model -f Dockerfile.eval .

# evaluation server
docker build -t evaluate_model -f Dockerfile .

# Run the container with tests
docker run --gpus all --env-file .env -v ~/.cache/huggingface:/root/.cache/huggingface  evaluate_model 

Run docker api service

docker build -t evaluation-api .

docker run --gpus all --env-file .env -p 23477:23477 -v ~/.cache/huggingface:/root/.cache/huggingface evaluation-api

Run via docker compose

docker-compose up --build -d

Built With

This section lists major frameworks/libraries used to bootstrap the project:

mlflow Apache Airflow bucket db dbms ui GitHub Actions Vercel nVIDIA FastAPI Docker

About

Evaluation service of LLMOps

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors