Ashish Tanwer
ashishtanwer
AI & ML interests
None yet
Recent Activity
liked
a model
about 17 hours ago
openai/gpt-oss-120b
liked
a model
about 18 hours ago
Qwen/Qwen-Image
liked
a model
1 day ago
OmniAvatar/OmniAvatar-14B
Organizations
RAG
DataLabelling
LLM
-
Running2.49k2.49k
Anycoder
π’Generate modern web applications with Svelte or transformers.js
-
Runtime error274274
Qwen2.5 Coder Artifacts
π’Generate application code with Qwen2.5-Coder-32B
-
Running922922
QwQ-32B-Preview
πQwQ-32B-Preview
-
Running on CPU Upgrade13.4k13.4k
Open LLM Leaderboard
πTrack, rank and evaluate open LLMs and chatbots
Evals
ClassicalML
Paper and resources for Classical ML
InfraML
Agents
Transformer
-
sentence-transformers/all-mpnet-base-v2
Sentence Similarity β’ 0.1B β’ Updated β’ 18.9M β’ β’ 1.12k -
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Paper β’ 1910.10683 β’ Published β’ 14 -
google-t5/t5-base
Translation β’ 0.2B β’ Updated β’ 1.79M β’ β’ 733 -
Attention Is All You Need
Paper β’ 1706.03762 β’ Published β’ 76
DataCleaning
Dataset
-
The RefinedWeb Dataset for Falcon LLM: Outperforming Curated Corpora with Web Data, and Web Data Only
Paper β’ 2306.01116 β’ Published β’ 38 -
HuggingFaceFW/fineweb
Viewer β’ Updated β’ 52.5B β’ 537k β’ 2.28k -
tiiuae/falcon-refinedweb
Viewer β’ Updated β’ 968M β’ 12.8k β’ 865 -
cerebras/SlimPajama-627B
Preview β’ Updated β’ 63.3k β’ 485
Training
-
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Paper β’ 1910.10683 β’ Published β’ 14 -
AutoTrain: No-code training for state-of-the-art models
Paper β’ 2410.15735 β’ Published β’ 60 -
LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report
Paper β’ 2405.00732 β’ Published β’ 122 -
LoRA: Low-Rank Adaptation of Large Language Models
Paper β’ 2106.09685 β’ Published β’ 44
Diffusion
DataCrawling
Agents
RAG
Transformer
-
sentence-transformers/all-mpnet-base-v2
Sentence Similarity β’ 0.1B β’ Updated β’ 18.9M β’ β’ 1.12k -
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Paper β’ 1910.10683 β’ Published β’ 14 -
google-t5/t5-base
Translation β’ 0.2B β’ Updated β’ 1.79M β’ β’ 733 -
Attention Is All You Need
Paper β’ 1706.03762 β’ Published β’ 76
DataLabelling
DataCleaning
LLM
-
Running2.49k2.49k
Anycoder
π’Generate modern web applications with Svelte or transformers.js
-
Runtime error274274
Qwen2.5 Coder Artifacts
π’Generate application code with Qwen2.5-Coder-32B
-
Running922922
QwQ-32B-Preview
πQwQ-32B-Preview
-
Running on CPU Upgrade13.4k13.4k
Open LLM Leaderboard
πTrack, rank and evaluate open LLMs and chatbots
Dataset
-
The RefinedWeb Dataset for Falcon LLM: Outperforming Curated Corpora with Web Data, and Web Data Only
Paper β’ 2306.01116 β’ Published β’ 38 -
HuggingFaceFW/fineweb
Viewer β’ Updated β’ 52.5B β’ 537k β’ 2.28k -
tiiuae/falcon-refinedweb
Viewer β’ Updated β’ 968M β’ 12.8k β’ 865 -
cerebras/SlimPajama-627B
Preview β’ Updated β’ 63.3k β’ 485
Evals
Training
-
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Paper β’ 1910.10683 β’ Published β’ 14 -
AutoTrain: No-code training for state-of-the-art models
Paper β’ 2410.15735 β’ Published β’ 60 -
LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report
Paper β’ 2405.00732 β’ Published β’ 122 -
LoRA: Low-Rank Adaptation of Large Language Models
Paper β’ 2106.09685 β’ Published β’ 44
ClassicalML
Paper and resources for Classical ML
Diffusion
InfraML