ncbi
/

Safetensors
qwen2
biology
bioinformatics
single-cell
Fangyinfff commited on
Commit
d188c2c
·
verified ·
1 Parent(s): c39bb74

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -10,6 +10,8 @@ tags:
10
 
11
  # Cell-o1: Training LLMs to Solve Single-Cell Reasoning Puzzles with Reinforcement Learning
12
 
 
 
13
  > [!Note]
14
  > Please refer to our [repository](https://github.com/ncbi-nlp/cell-o1) and [paper](https://www.arxiv.org/abs/2506.02911) for more details.
15
 
@@ -17,6 +19,9 @@ tags:
17
  Cell type annotation is a key task in analyzing the heterogeneity of single-cell RNA sequencing data. Although recent foundation models automate this process, they typically annotate cells independently, without considering batch-level cellular context or providing explanatory reasoning. In contrast, human experts often annotate distinct cell types for different cell clusters based on their domain knowledge.
18
  To mimic this expert behavior, we introduce ***CellPuzzles***—a benchmark requiring unique cell-type assignments across cell batches. Existing LLMs struggle with this task, with the best baseline (OpenAI's o1) achieving only 19.0% batch accuracy. To address this, we present ***Cell-o1***, a reasoning-enhanced language model trained via SFT on distilled expert traces, followed by RL with batch-level rewards. ***Cell-o1*** outperforms all baselines on both cell-level and batch-level metrics, and exhibits emergent behaviors such as self-reflection and curriculum reasoning, offering insights into its interpretability and generalization.
19
 
 
 
 
20
 
21
 
22
  ## 🚀 How to Run Inference
 
10
 
11
  # Cell-o1: Training LLMs to Solve Single-Cell Reasoning Puzzles with Reinforcement Learning
12
 
13
+
14
+
15
  > [!Note]
16
  > Please refer to our [repository](https://github.com/ncbi-nlp/cell-o1) and [paper](https://www.arxiv.org/abs/2506.02911) for more details.
17
 
 
19
  Cell type annotation is a key task in analyzing the heterogeneity of single-cell RNA sequencing data. Although recent foundation models automate this process, they typically annotate cells independently, without considering batch-level cellular context or providing explanatory reasoning. In contrast, human experts often annotate distinct cell types for different cell clusters based on their domain knowledge.
20
  To mimic this expert behavior, we introduce ***CellPuzzles***—a benchmark requiring unique cell-type assignments across cell batches. Existing LLMs struggle with this task, with the best baseline (OpenAI's o1) achieving only 19.0% batch accuracy. To address this, we present ***Cell-o1***, a reasoning-enhanced language model trained via SFT on distilled expert traces, followed by RL with batch-level rewards. ***Cell-o1*** outperforms all baselines on both cell-level and batch-level metrics, and exhibits emergent behaviors such as self-reflection and curriculum reasoning, offering insights into its interpretability and generalization.
21
 
22
+ <p align="center">
23
+ <img src="assets/overview.png" alt="CellPuzzles Overview" width="95%">
24
+ </p>
25
 
26
 
27
  ## 🚀 How to Run Inference