-
meta-llama/Llama-3.1-8B-Instruct
Text Generation • 8B • Updated • 5.01M • • 5.04k -
meta-llama/Llama-3.2-3B-Instruct
Text Generation • 3B • Updated • 1.62M • • 1.84k -
meta-llama/Llama-3.3-70B-Instruct
Text Generation • 71B • Updated • 421k • • 2.58k -
meta-llama/Llama-4-Scout-17B-16E-Instruct
Any-to-Any • 109B • Updated • 206k • 1.15k
Satya Saurabh Mishra
saurabhmishra9
·
AI & ML interests
Data Science, Machine Learning, AI etc
Organizations
Prompting and RAG
-
Don't Do RAG: When Cache-Augmented Generation is All You Need for Knowledge Tasks
Paper • 2412.15605 • Published • 2 -
Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection
Paper • 2310.11511 • Published • 78 -
Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity
Paper • 2403.14403 • Published • 7 -
Compressed Chain of Thought: Efficient Reasoning Through Dense Representations
Paper • 2412.13171 • Published • 35
LLM Models
-
meta-llama/Llama-3.1-8B-Instruct
Text Generation • 8B • Updated • 5.01M • • 5.04k -
meta-llama/Llama-3.2-3B-Instruct
Text Generation • 3B • Updated • 1.62M • • 1.84k -
meta-llama/Llama-3.3-70B-Instruct
Text Generation • 71B • Updated • 421k • • 2.58k -
meta-llama/Llama-4-Scout-17B-16E-Instruct
Any-to-Any • 109B • Updated • 206k • 1.15k
Prompting and RAG
-
Don't Do RAG: When Cache-Augmented Generation is All You Need for Knowledge Tasks
Paper • 2412.15605 • Published • 2 -
Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection
Paper • 2310.11511 • Published • 78 -
Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity
Paper • 2403.14403 • Published • 7 -
Compressed Chain of Thought: Efficient Reasoning Through Dense Representations
Paper • 2412.13171 • Published • 35
models
0
None public yet
datasets
0
None public yet