Thank you for your interest in contributing to the CASE Benchmark!
Evaluated a model? Share your results!
- Run evaluation using the benchmark
- Create a PR adding your results to
results/<model_name>/ - Include:
results.json- Evaluation outputmodel_card.md- Model description
See docs/submission.md for detailed instructions.
Found a bug or have a suggestion?
- Check existing issues
- Open a new issue with:
- Clear description
- Steps to reproduce (for bugs)
- Expected vs actual behavior
Documentation improvements are always welcome:
- Fix typos or clarify explanations
- Add examples or tutorials
- Translate to other languages
Want to add support for a new model?
- Create a new file in
case_benchmark/models/ - Implement the
EmbeddingModelinterface - Add tests in
tests/ - Update documentation
# Clone the repository
git clone https://github.com/gittb/case-benchmark.git
cd case-benchmark
# Create virtual environment
python -m venv venv
source venv/bin/activate # or `venv\Scripts\activate` on Windows
# Install in development mode
pip install -e ".[dev]"
# Run tests
pytest tests/
# Run linting
ruff check .
ruff format .- Use ruff for formatting and linting
- Follow existing code patterns
- Add docstrings to public functions
- Write tests for new functionality
- Fork the repository
- Create a feature branch (
git checkout -b feature/my-feature) - Make your changes
- Run tests and linting
- Commit with clear messages
- Open a pull request
Open a discussion or reach out via issues.