Thanks for taking the time to contribute! This project aims to track and compare LLM coding capabilities through interactive demonstrations.
- Model Outputs:
- Provide only the generated
index.htmlfile in the model subdirectory. - Ensure the output is relatively self-contained.
- If the model generated a single HTML file with CSS and JS included, that's preferred.
- Provide only the generated
- Structure:
- Follow the directory structure:
benchmark/model/index.html. - Include the matching
prompt.txtfor any new benchmark.
- Follow the directory structure:
- Model Identifiers: Use lowercase names for folders (e.g.,
gpt-4o,gemini-1.5-pro). - No Malicious Code: Submissions must not contain any scripts tracking users or conducting malicious activity.
- Fork the repository and create your feature branch.
- Add your model result or benchmark.
- Run
bash create_config.shto updateconfig.json. - Commit your changes (including the updated
config.json). - Test Locally: Open
index.htmlin your browser and verify your new entry appears correctly. - Submit a PR: Provide a brief description of the model or benchmark you added.
We use a simple bash script to generate the site index. Before submitting, ensure:
bash create_config.shruns without errors and the resulting config.json is valid.
If you have questions or want to discuss a specific benchmark, feel free to open an Issue!