Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,8 @@ tags:
|
|
10 |
|
11 |
# Cell-o1: Training LLMs to Solve Single-Cell Reasoning Puzzles with Reinforcement Learning
|
12 |
|
|
|
|
|
13 |
> [!Note]
|
14 |
> Please refer to our [repository](https://github.com/ncbi-nlp/cell-o1) and [paper](https://www.arxiv.org/abs/2506.02911) for more details.
|
15 |
|
@@ -17,6 +19,9 @@ tags:
|
|
17 |
Cell type annotation is a key task in analyzing the heterogeneity of single-cell RNA sequencing data. Although recent foundation models automate this process, they typically annotate cells independently, without considering batch-level cellular context or providing explanatory reasoning. In contrast, human experts often annotate distinct cell types for different cell clusters based on their domain knowledge.
|
18 |
To mimic this expert behavior, we introduce ***CellPuzzles***—a benchmark requiring unique cell-type assignments across cell batches. Existing LLMs struggle with this task, with the best baseline (OpenAI's o1) achieving only 19.0% batch accuracy. To address this, we present ***Cell-o1***, a reasoning-enhanced language model trained via SFT on distilled expert traces, followed by RL with batch-level rewards. ***Cell-o1*** outperforms all baselines on both cell-level and batch-level metrics, and exhibits emergent behaviors such as self-reflection and curriculum reasoning, offering insights into its interpretability and generalization.
|
19 |
|
|
|
|
|
|
|
20 |
|
21 |
|
22 |
## 🚀 How to Run Inference
|
|
|
10 |
|
11 |
# Cell-o1: Training LLMs to Solve Single-Cell Reasoning Puzzles with Reinforcement Learning
|
12 |
|
13 |
+
|
14 |
+
|
15 |
> [!Note]
|
16 |
> Please refer to our [repository](https://github.com/ncbi-nlp/cell-o1) and [paper](https://www.arxiv.org/abs/2506.02911) for more details.
|
17 |
|
|
|
19 |
Cell type annotation is a key task in analyzing the heterogeneity of single-cell RNA sequencing data. Although recent foundation models automate this process, they typically annotate cells independently, without considering batch-level cellular context or providing explanatory reasoning. In contrast, human experts often annotate distinct cell types for different cell clusters based on their domain knowledge.
|
20 |
To mimic this expert behavior, we introduce ***CellPuzzles***—a benchmark requiring unique cell-type assignments across cell batches. Existing LLMs struggle with this task, with the best baseline (OpenAI's o1) achieving only 19.0% batch accuracy. To address this, we present ***Cell-o1***, a reasoning-enhanced language model trained via SFT on distilled expert traces, followed by RL with batch-level rewards. ***Cell-o1*** outperforms all baselines on both cell-level and batch-level metrics, and exhibits emergent behaviors such as self-reflection and curriculum reasoning, offering insights into its interpretability and generalization.
|
21 |
|
22 |
+
<p align="center">
|
23 |
+
<img src="assets/overview.png" alt="CellPuzzles Overview" width="95%">
|
24 |
+
</p>
|
25 |
|
26 |
|
27 |
## 🚀 How to Run Inference
|