Thank you for your interest in contributing to SCBench! This document provides guidelines for creating custom graders and submitting benchmark results.
Note: This repository contains example evaluations only. The full benchmark is maintained separately to prevent overfitting. Contact kenny@latch.bio for information about contributing to the full benchmark.
The complete SCBench comprises evaluations across single-cell platforms:
| Platform | Status |
|---|---|
| 10x Chromium | Available |
| BD Rhapsody | Available |
| Parse | Available |
| Illumina | Available |
| MissionBio | Available |
| CSGenetics | Available |
This repository contains representative examples. The full benchmark is withheld to ensure performance metrics reflect genuine single-cell analysis capabilities.
Graders are maintained in a separate repository. To create a custom grader, see eval-graders.
Help build the leaderboard by benchmarking your model:
scbench batch evals_canonical/ --model your-model-name --output results/scbench leaderboard results/ --output summary.jsonInclude:
- summary.json with results
- Model name and version
- Date of evaluation
- Any relevant notes (hardware, runtime, etc.)
- End files with newlines
- No docstrings or comments (per project style)
- Use
|for Union types:str | None - Imports at top of file
- Use available types (
list,dict) instead of typing module
- 2-space indentation
- No trailing commas
- Sorted keys (for ground truth labels)
- Clear, descriptive commit messages
- One logical change per commit
- Reference issue numbers when applicable
- Automated checks: Tests and linting must pass
- Ground truth review: Verify analysis is correct and reproducible
- Grader review: Ensure grader logic is sound
- Documentation review: Check clarity and completeness
- Approval: Two maintainer approvals required
- Open an issue for questions
- Tag maintainers for urgent reviews
- Join discussions for design decisions
Thank you for contributing to SCBench!