Contextual AI Reranker v2 Family of instruction-following multilingual rerankers on the cost/performance Pareto frontier across public and customer benchmarks ContextualAI/ctxl-rerank-v2-instruct-multilingual-6b Text Ranking • 7B • Updated Sep 3 • 99 • 13 ContextualAI/ctxl-rerank-v2-instruct-multilingual-6b-nvfp4 Text Ranking • 4B • Updated Sep 3 • 76 • 2 ContextualAI/ctxl-rerank-v2-instruct-multilingual-2b Text Ranking • 3B • Updated Sep 3 • 296 • 1 ContextualAI/ctxl-rerank-v2-instruct-multilingual-2b-nvfp4 Text Ranking • 2B • Updated Sep 3 • 17
ContextualAI/ctxl-rerank-v2-instruct-multilingual-6b-nvfp4 Text Ranking • 4B • Updated Sep 3 • 76 • 2
LMUnit Models & Datasets for LMunit evaluation ContextualAI/LFQA Viewer • Updated Jul 19 • 260 • 4 ContextualAI/reward-bench-filtered Viewer • Updated Jul 19 • 2.99k • 11 ContextualAI/BiGGenBench Viewer • Updated Jul 20 • 3.2k • 6 ContextualAI/Flask Viewer • Updated Jul 18 • 2k • 21
Archangel Archangel is a suite of human feedback-aligned LLMs, released as part of the Human-Aware Loss Functions (HALOs) project by Ethayarajh et al. (2024). ContextualAI/archangel_sft_llama7b Text Generation • 7B • Updated Jan 11, 2024 • 10 • 1 ContextualAI/archangel_kto_llama30b Text Generation • 33B • Updated Jan 11, 2024 • 2 • 2 ContextualAI/archangel_sft-kto_llama13b Text Generation • 13B • Updated Jan 11, 2024 • 1.16k • 3 ContextualAI/archangel_sft-kto_llama30b Text Generation • 33B • Updated Jan 11, 2024 • 4 • 2
CLAIR and APO Data and Models for the paper "Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment" ContextualAI/ultrafeedback_clair_32k Viewer • Updated Sep 17, 2024 • 29.1k • 137 • 5 ContextualAI/ultrafeedback_rlaif_32k Viewer • Updated Aug 13, 2024 • 31.8k • 18 • 1 ContextualAI/ultrafeedback_rlaif-offpolicy_32k Viewer • Updated Aug 13, 2024 • 29.1k • 14 • 1 ContextualAI/ultrafeedback_stronger-preferred_32k Viewer • Updated Aug 13, 2024 • 29.1k • 14 • 1
Contextual AI Instruction-Following Retrieval Evals Instruction-following retrieval evaluation datasets ContextualAI/recency-based-instructions Viewer • Updated Aug 12 • 268 • 31 ContextualAI/source-based-instructions Viewer • Updated Aug 12 • 270 • 30 • 1 ContextualAI/complex-instructions Viewer • Updated Aug 12 • 271 • 14
Contextual-SQL Text-to-SQL models ContextualAI/ctx-bird-reward-250121 Text Generation • 33B • Updated Jul 3 • 11 • 3
Zephyr KTO Aligned models based on Zephyr-SFT from Table 2 and 3 in the KTO paper by Ethayarajh et al. (2024) (https://arxiv.org/pdf/2402.01306). ContextualAI/zephyr_sft_kto Text Generation • Updated May 5, 2024 • 1 • 1 ContextualAI/zephyr_sft_kto_unary Text Generation • Updated May 5, 2024 ContextualAI/zephyr_sft_dpo Text Generation • Updated May 5, 2024
Contextual AI Reranker v2 Family of instruction-following multilingual rerankers on the cost/performance Pareto frontier across public and customer benchmarks ContextualAI/ctxl-rerank-v2-instruct-multilingual-6b Text Ranking • 7B • Updated Sep 3 • 99 • 13 ContextualAI/ctxl-rerank-v2-instruct-multilingual-6b-nvfp4 Text Ranking • 4B • Updated Sep 3 • 76 • 2 ContextualAI/ctxl-rerank-v2-instruct-multilingual-2b Text Ranking • 3B • Updated Sep 3 • 296 • 1 ContextualAI/ctxl-rerank-v2-instruct-multilingual-2b-nvfp4 Text Ranking • 2B • Updated Sep 3 • 17
ContextualAI/ctxl-rerank-v2-instruct-multilingual-6b-nvfp4 Text Ranking • 4B • Updated Sep 3 • 76 • 2
Contextual AI Instruction-Following Retrieval Evals Instruction-following retrieval evaluation datasets ContextualAI/recency-based-instructions Viewer • Updated Aug 12 • 268 • 31 ContextualAI/source-based-instructions Viewer • Updated Aug 12 • 270 • 30 • 1 ContextualAI/complex-instructions Viewer • Updated Aug 12 • 271 • 14
LMUnit Models & Datasets for LMunit evaluation ContextualAI/LFQA Viewer • Updated Jul 19 • 260 • 4 ContextualAI/reward-bench-filtered Viewer • Updated Jul 19 • 2.99k • 11 ContextualAI/BiGGenBench Viewer • Updated Jul 20 • 3.2k • 6 ContextualAI/Flask Viewer • Updated Jul 18 • 2k • 21
Contextual-SQL Text-to-SQL models ContextualAI/ctx-bird-reward-250121 Text Generation • 33B • Updated Jul 3 • 11 • 3
Archangel Archangel is a suite of human feedback-aligned LLMs, released as part of the Human-Aware Loss Functions (HALOs) project by Ethayarajh et al. (2024). ContextualAI/archangel_sft_llama7b Text Generation • 7B • Updated Jan 11, 2024 • 10 • 1 ContextualAI/archangel_kto_llama30b Text Generation • 33B • Updated Jan 11, 2024 • 2 • 2 ContextualAI/archangel_sft-kto_llama13b Text Generation • 13B • Updated Jan 11, 2024 • 1.16k • 3 ContextualAI/archangel_sft-kto_llama30b Text Generation • 33B • Updated Jan 11, 2024 • 4 • 2
Zephyr KTO Aligned models based on Zephyr-SFT from Table 2 and 3 in the KTO paper by Ethayarajh et al. (2024) (https://arxiv.org/pdf/2402.01306). ContextualAI/zephyr_sft_kto Text Generation • Updated May 5, 2024 • 1 • 1 ContextualAI/zephyr_sft_kto_unary Text Generation • Updated May 5, 2024 ContextualAI/zephyr_sft_dpo Text Generation • Updated May 5, 2024
CLAIR and APO Data and Models for the paper "Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment" ContextualAI/ultrafeedback_clair_32k Viewer • Updated Sep 17, 2024 • 29.1k • 137 • 5 ContextualAI/ultrafeedback_rlaif_32k Viewer • Updated Aug 13, 2024 • 31.8k • 18 • 1 ContextualAI/ultrafeedback_rlaif-offpolicy_32k Viewer • Updated Aug 13, 2024 • 29.1k • 14 • 1 ContextualAI/ultrafeedback_stronger-preferred_32k Viewer • Updated Aug 13, 2024 • 29.1k • 14 • 1