-
Quantifying Language Models' Sensitivity to Spurious Features in Prompt Design or: How I learned to start worrying about prompt formatting
Paper • 2310.11324 • Published • 1 -
Flaw or Artifact? Rethinking Prompt Sensitivity in Evaluating LLMs
Paper • 2509.01790 • Published • 7 -
POSIX: A Prompt Sensitivity Index For Large Language Models
Paper • 2410.02185 • Published -
A Survey on Evaluation of Large Language Models
Paper • 2307.03109 • Published • 43
Collections
Discover the best community collections!
Collections including paper arxiv:2307.03109
-
A Survey on Evaluation of Large Language Models
Paper • 2307.03109 • Published • 43 -
SemScore: Automated Evaluation of Instruction-Tuned LLMs based on Semantic Textual Similarity
Paper • 2401.17072 • Published • 25 -
LLM Comparator: Visual Analytics for Side-by-Side Evaluation of Large Language Models
Paper • 2402.10524 • Published • 23
-
Levels of AGI for Operationalizing Progress on the Path to AGI
Paper • 2311.02462 • Published • 36 -
Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
Paper • 2206.04615 • Published • 6 -
A Survey on Evaluation of Large Language Models
Paper • 2307.03109 • Published • 43 -
Bring Your Own Data! Self-Supervised Evaluation for Large Language Models
Paper • 2306.13651 • Published • 16
-
Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning
Paper • 2211.04325 • Published • 1 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 27 -
On the Opportunities and Risks of Foundation Models
Paper • 2108.07258 • Published • 2 -
Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks
Paper • 2204.07705 • Published • 2
-
Instruction-Following Evaluation for Large Language Models
Paper • 2311.07911 • Published • 22 -
HuggingFaceH4/mt_bench_prompts
Viewer • Updated • 80 • 7.32k • 25 -
vectara/hallucination_evaluation_model
Text Classification • Updated • 96.3k • 350 -
GAIA: a benchmark for General AI Assistants
Paper • 2311.12983 • Published • 248
-
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20 -
Learning to summarize from human feedback
Paper • 2009.01325 • Published • 4 -
Training language models to follow instructions with human feedback
Paper • 2203.02155 • Published • 24
-
Quantifying Language Models' Sensitivity to Spurious Features in Prompt Design or: How I learned to start worrying about prompt formatting
Paper • 2310.11324 • Published • 1 -
Flaw or Artifact? Rethinking Prompt Sensitivity in Evaluating LLMs
Paper • 2509.01790 • Published • 7 -
POSIX: A Prompt Sensitivity Index For Large Language Models
Paper • 2410.02185 • Published -
A Survey on Evaluation of Large Language Models
Paper • 2307.03109 • Published • 43
-
Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning
Paper • 2211.04325 • Published • 1 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 27 -
On the Opportunities and Risks of Foundation Models
Paper • 2108.07258 • Published • 2 -
Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks
Paper • 2204.07705 • Published • 2
-
A Survey on Evaluation of Large Language Models
Paper • 2307.03109 • Published • 43 -
SemScore: Automated Evaluation of Instruction-Tuned LLMs based on Semantic Textual Similarity
Paper • 2401.17072 • Published • 25 -
LLM Comparator: Visual Analytics for Side-by-Side Evaluation of Large Language Models
Paper • 2402.10524 • Published • 23
-
Instruction-Following Evaluation for Large Language Models
Paper • 2311.07911 • Published • 22 -
HuggingFaceH4/mt_bench_prompts
Viewer • Updated • 80 • 7.32k • 25 -
vectara/hallucination_evaluation_model
Text Classification • Updated • 96.3k • 350 -
GAIA: a benchmark for General AI Assistants
Paper • 2311.12983 • Published • 248
-
Levels of AGI for Operationalizing Progress on the Path to AGI
Paper • 2311.02462 • Published • 36 -
Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
Paper • 2206.04615 • Published • 6 -
A Survey on Evaluation of Large Language Models
Paper • 2307.03109 • Published • 43 -
Bring Your Own Data! Self-Supervised Evaluation for Large Language Models
Paper • 2306.13651 • Published • 16
-
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20 -
Learning to summarize from human feedback
Paper • 2009.01325 • Published • 4 -
Training language models to follow instructions with human feedback
Paper • 2203.02155 • Published • 24