jiagengwu commited on
Commit
2a585ed
Β·
verified Β·
1 Parent(s): 5f81501

Update docs.md

Browse files
Files changed (1) hide show
  1. docs.md +79 -28
docs.md CHANGED
@@ -24,49 +24,100 @@
24
 
25
 
26
  <h2>πŸ“œ Background</h2>
27
- <p>Recent advances in <strong>Large Language Models (LLMs)</strong> have demonstrated transformative potential in improving healthcare delivery and clinical research. By combining extensive pretraining with supervised instruction tuning across diverse tasks, LLMs excel in natural language understanding, generation, and reasoning. These capabilities allow LLMs to serve as versatile, general-purpose medical assistants.</p>
28
- <p>Despite this promise, concerns remain around the <strong>reliability and clinical validity</strong> of LLM-generated outputs. Real-world contexts often involve unstructured, multilingual text from <strong>electronic health records (EHRs)</strong>, and require support for tasks like phenotype identification and event extraction that remain underexplored. Moreover, the scarcity of <strong>multilingual benchmarks</strong> further limits the global applicability of LLMs in medicine.</p>
29
- <p>To address these challenges, we introduce the <em>largest multilingual clinical benchmark</em> to date, <strong>BRIDGE (Benchmarking Large Language Models in Multilingual Real-world Clinical Text Understanding)</strong>, evaluating 52 LLMs on:</p>
 
 
 
 
 
30
  <ul>
31
- <li><strong>87 clinical tasks</strong></li>
32
- <li><strong>9 languages</strong></li>
33
- <li><strong>1M+ clinical samples</strong></li>
34
  </ul>
 
35
 
36
  <div style="text-align: center;">
37
- <img src="https://cdn-uploads.huggingface.co/production/uploads/67a040fb6934f9aa1c866f99/2fh-jETNSL9iXJXTT-fdN.png" style="width: 50%;" alt="BRIDGE benchmark graphic">
38
  </div>
39
 
 
40
  <h2>🌍 Key Features</h2>
41
- <p>Our benchmark spans a wide range of document types and clinical tasks, including classification, event extraction, and generation. It further supports three inference strategies: <strong>zero-shot</strong>, <strong>few-shot</strong>, and <strong>chain-of-thought (CoT)</strong> prompting. We evaluated <strong>52 LLMs</strong>, including general-purpose, open-source, proprietary, and medical-domain models.</p>
42
  <ul>
43
- <li><strong>Multilingual Data</strong>: Clinical tasks in <strong>9 languages</strong> for global relevance.</li>
44
- <li><strong>Diverse Clinical Documents</strong>: Notes, summaries, radiology reports, and more.</li>
45
- <li><strong>Multiple NLP Tasks</strong>: Classification, extraction, QA, summarization, etc.</li>
46
- <li><strong>Evaluation Modes</strong>:
 
47
  <ul>
48
- <li><strong>Zero-shot</strong></li>
49
- <li><strong>Few-shot</strong></li>
50
- <li><strong>Chain-of-Thought (CoT)</strong> reasoning</li>
 
51
  </ul>
52
- </li>
53
  </ul>
 
54
 
 
 
 
 
55
 
56
- <h2>πŸ† BRIDGE Leaderboard</h2>
57
- <p>To support ongoing evaluation, we introduce our <strong>BRIDGE Leaderboard</strong>, which provides:</p>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
  <ul>
59
- <li>Easy visualizations</li>
60
- <li>Side-by-side comparisons</li>
61
- <li>Continuous tracking of LLM performance across tasks, languages, and evaluation strategies</li>
 
 
 
 
 
 
 
62
  </ul>
63
- <p>This leaderboard empowers researchers and clinicians to make informed decisions and track model progress over time.</p>
64
 
65
  <h2>πŸ“š Citation</h2>
66
- <pre><code>@article{BRIDGE2025,
67
- title = {PAPER TITLE},
68
- author = {Your Name and Contributors},
69
- year = {2025},
70
- journal = {Your Journal or Conference},
 
 
 
 
 
 
 
 
 
 
 
 
 
 
71
  }
72
- </code></pre>
 
24
 
25
 
26
  <h2>πŸ“œ Background</h2>
27
+ <p>Recent advances in <strong>Large Language Models (LLMs)</strong> have demonstrated transformative potential in <strong>healthcare</strong>, yet concerns remain around their reliability and clinical validity across diverse clinical tasks, specialties, and languages. To support timely and trustworthy evaluation, building upon our <a href="https://ai.nejm.org/doi/full/10.1056/AIra2400012">systematic review</a> of global clinical text resources, we introduce <a href="https://arxiv.org/abs/2504.19467">BRIDGE</a>, <strong>a multilingual benchmark that comprises 87 real-world clinical text tasks spanning nine languages and more than one million samples</strong>. Furthermore, we construct this leaderboard of LLM in clinical text understanding by systematically evaluating <strong>52 state-of-the-art LLMs</strong> (by 2024/04/29).</p>
28
+
29
+ <div style="text-align: center;">
30
+ <img src="figure/dataset.png" style="width: 100%;" alt="BRIDGE benchmark graphic">
31
+ </div>
32
+
33
+ <h2>πŸ† BRIDGE Leaderboard</h2>
34
+ <p>BRIDGE features three leaderboards, each evaluating LLM performance in clinical text tasks under a distinct inference strategy:</p>
35
  <ul>
36
+ <li><strong>Zero shot</strong>: Only the task instructions and input data are provided. The LLM is prompted to directly produce the target answer without any support.</li>
37
+ <li><strong>Chain-of-Thought (CoT)</strong>: Task instructions explicitly guide the LLM to generate a step-by-step explanation of its reasoning process before providing the final answer, enhancing interpretability and reasoning transparency.</li>
38
+ <li><strong>Few-shot</strong>: Five independent samples serve as examples, which leverage the LLM's capability of in-context learning to guide the model to conduct tasks. </li>
39
  </ul>
40
+ <p>In addition, BRIDGE offers multiple <strong>model filters</strong> and <strong>task filters</strong> to enable users to explore LLM performance across <strong>different clinical contexts</strong>, empowering researchers and clinicians to make informed decisions and track model advancements over time.</p>
41
 
42
  <div style="text-align: center;">
43
+ <img src="figure/model.png" style="width: 100%;" alt="BRIDGE benchmark graphic">
44
  </div>
45
 
46
+
47
  <h2>🌍 Key Features</h2>
 
48
  <ul>
49
+ <li><strong>Real-world Clinical Text</strong>: All tasks are sourced from real-world medical settings, such as electronic health records (EHRs), clinical case reports, or healthcare consultations</li>
50
+ <li><strong>Multilingual Context</strong>: 9 languages: English, Chinese, Spanish, Japanese, German, Russian, French, Norwegian, and Portuguese</li>
51
+ <li><strong>Diverse Task Types</strong>: 8 task types: Text classification, Semantic similarity, Normalization and Coding, Named entity recognition (NER), Event extraction, Question answering (QA), and Text Summarization</li>
52
+ <li><strong>Broad Clinical Applications</strong>: 14 Clinical specialties, 7 Clinical document types, 20 Clinical applications covering 6 clinical stages of patient care</li>
53
+ <li><strong>Advanced LLMs (52 models)</strong>:
54
  <ul>
55
+ <li><strong>Proprietary models</strong>: GPT-4o, GPT-3.5, Gemini-2.0-Flash, Gemini-1.5-Pro ...</li>
56
+ <li><strong>Open-source models</strong>: Llama 3/4, QWEN2.5, Mistral, Gemma ...</li>
57
+ <li><strong>Medical models</strong>: Baichuan-M1-14B, meditron, MeLLaMA... </li>
58
+ <li><strong>Reasoning models</strong>: Deepseek-R1(671B), QWQ-32B, Deepseek-R1-Distll-Qwen/Llama ...</li>
59
  </ul>
60
+ </li>
61
  </ul>
62
+ More Details can be found in our <a href="https://arxiv.org/abs/2504.19467">BRIDGE paper</a> and <a href="https://ai.nejm.org/doi/full/10.1056/AIra2400012">systematic review</a>.
63
 
64
+ <h2>πŸ› οΈ How to Evaluate Your Model on BRIDGE ?</h2>
65
+ <h4>πŸ“‚ Dataset Access</h4>
66
+ <p>All fully open-access datasets in BRIDGE are available in <a href="https://huggingface.co/datasets/YLab-Open/BRIDGE-Open">BRIDGE-Open</a>. To ensure fairness of this leaderboard, we publicly release the following data for each task:
67
+ Five completed samples serving as few-shot examples, and all testing samples with instruction and input information.</p>
68
 
69
+ <p>Due to privacy and security considerations of clinical data, regulated-access datasets can not be directly published. However, all detailed task descriptions and their corresponding data sources are available in our <a href="https://arxiv.org/abs/2504.19467">BRIDGE paper</a>.
70
+ Importantly, all 87 datasets have been verified to be either fully open-access or publicly accessible via reasonable request.</p>
71
+
72
+ <h4>πŸ”₯ Result Submission and Evaluation</h4>
73
+ If you would like to submit your model results to BRIDGE and demonstrate its performance, please follow these steps:
74
+ <ul>
75
+ <li><strong>Run local inference:</strong> Download the <a href="https://huggingface.co/datasets/YLab-Open/BRIDGE-Open">BRIDGE-Open</a> dataset and perform inference locally. Save the generated output of each sample in its "pred" field for each dataset file.</li>
76
+ <li><strong>Prepare your submission:</strong> Send your results to us (see the following Contact Information). Please clearly describe your model details and experimental settings, including the inference strategy used (zero-shot, few-shot, or CoT), and note any additional relevant configurations if applicable.</li>
77
+ <li><strong>Code Reference:</strong> About LLM inference, result extraction, and evaluation scheme, please refer to our <a href="https://github.com/YLab-Open/BRIDGE">BRIDGE GitHub repo</a>.</li>
78
+ </ul>
79
+ We will review and evaluate your submission and update the leaderboard accordingly.
80
+
81
+
82
+ <h2>🀝 Contributing</h2>
83
+ <p>We welcome and greatly value contributions and collaborations from the community!
84
+ If you have clinical text datasets that you would like to share for broader exploration, please contact us!</p>
85
+ <p>We are committed to expanding BRIDGE while strictly adhering to appropriate data use agreements and ethical guidelines. Let's work together to advance the responsible application of LLMs in medicine!</p>
86
+
87
+
88
+ <h2>πŸ“’ Updates</h2>
89
  <ul>
90
+ <li>πŸ—“οΈ 2025/04/28: BRIDGE Leaderboard V1.0.0 is now live!</li>
91
+ <li>πŸ—“οΈ 2025/04/28: Our paper <a href="https://arxiv.org/abs/2504.19467">BRIDGE</a> is now available on arXiv!</li>
92
+ </ul>
93
+
94
+ <h2>πŸ“¬ Contact Information</h2>
95
+ If you have any questions about BRIDGE or the leaderboard, feel free to reach out!
96
+ <ul>
97
+ <li><strong>Leaderboard Managers</strong>: Jiageng Wu ([email protected]), Kevin Xie ([email protected])</li>
98
+ <li><strong>Benchmark Managers</strong>: Bowen Gu ([email protected]), Jiageng Wu ([email protected])</li>
99
+ <li><strong>Program Lead</strong>: Jie Yang ([email protected])</li>
100
  </ul>
 
101
 
102
  <h2>πŸ“š Citation</h2>
103
+ <pre><code>
104
+ @article{BRIDGE-benchmark,
105
+ title={BRIDGE: Benchmarking Large Language Models for Understanding Real-world Clinical Practice Text},
106
+ author={Wu, Jiageng and Gu, Bowen and Zhou, Ren and Xie, Kevin and Snyder, Doug and Jiang, Yixing and Carducci, Valentina and Wyss, Richard and Desai, Rishi J and Alsentzer, Emily and Celi, Leo Anthony and Rodman, Adam and Schneeweiss, Sebastian and Chen, Jonathan H. and Romero-Brufau, Santiago and Lin, Kueiyu Joshua and Yang, Jie},
107
+ year={2025},
108
+ journal={arXiv preprint arXiv: 2504.19467},
109
+ archivePrefix={arXiv},
110
+ primaryClass={cs.CL},
111
+ url={https://arxiv.org/abs/2504.19467},
112
+ }
113
+ @article{clinical-text-review,
114
+ title={Clinical text datasets for medical artificial intelligence and large language modelsβ€”a systematic review},
115
+ author={Wu, Jiageng and Liu, Xiaocong and Li, Minghui and Li, Wanxin and Su, Zichang and Lin, Shixu and Garay, Lucas and Zhang, Zhiyun and Zhang, Yujie and Zeng, Qingcheng and Shen, Jie and Yuan, Changzheng and Yang, Jie},
116
+ journal={NEJM AI},
117
+ volume={1},
118
+ number={6},
119
+ pages={AIra2400012},
120
+ year={2024},
121
+ publisher={Massachusetts Medical Society}
122
  }
123
+ </code></pre>