-
OpenCodeInstruct: A Large-scale Instruction Tuning Dataset for Code LLMs
Paper • 2504.04030 • Published • 1 -
KodCode: A Diverse, Challenging, and Verifiable Synthetic Dataset for Coding
Paper • 2503.02951 • Published • 33 -
BigCodeBench: Benchmarking Code Generation with Diverse Function Calls and Complex Instructions
Paper • 2406.15877 • Published • 48 -
Magicoder: Source Code Is All You Need
Paper • 2312.02120 • Published • 82
Renat
u-brixton
·
AI & ML interests
None yet
Organizations
emlnp 2023 tbd
-
Can Retriever-Augmented Language Models Reason? The Blame Game Between the Retriever and the Language Model
Paper • 2212.09146 • Published • 3 -
RaLLe: A Framework for Developing and Evaluating Retrieval-Augmented Large Language Models
Paper • 2308.10633 • Published • 1 -
MemeCap: A Dataset for Captioning and Interpreting Memes
Paper • 2305.13703 • Published -
Contrastive Learning for Inference in Dialogue
Paper • 2310.12467 • Published
foundation_models
-
Apple Intelligence Foundation Language Models
Paper • 2407.21075 • Published • 5 -
The Llama 3 Herd of Models
Paper • 2407.21783 • Published • 117 -
Nemotron-4 340B Technical Report
Paper • 2406.11704 • Published -
Gemma 2: Improving Open Language Models at a Practical Size
Paper • 2408.00118 • Published • 78
monte_carlo_24_best
-
Planning Like Human: A Dual-process Framework for Dialogue Planning
Paper • 2406.05374 • Published -
Plug-and-Play Policy Planner for Large Language Model Powered Dialogue Agents
Paper • 2311.00262 • Published -
Strength Lies in Differences! Towards Effective Non-collaborative Dialogues via Tailored Strategy Planning
Paper • 2403.06769 • Published -
Prompt-Based Monte-Carlo Tree Search for Goal-Oriented Dialogue Policy Planning
Paper • 2305.13660 • Published
emnlp 2023
-
DSI++: Updating Transformer Memory with New Documents
Paper • 2212.09744 • Published • 1 -
Where to start? Analyzing the potential value of intermediate models
Paper • 2211.00107 • Published -
INSTRUCTSCORE: Explainable Text Generation Evaluation with Finegrained Feedback
Paper • 2305.14282 • Published -
G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment
Paper • 2303.16634 • Published • 3
math
-
Why do Learning Rates Transfer? Reconciling Optimization and Scaling Limits for Deep Learning
Paper • 2402.17457 • Published -
Curvature-Informed SGD via General Purpose Lie-Group Preconditioners
Paper • 2402.04553 • Published -
TextGrad: Automatic "Differentiation" via Text
Paper • 2406.07496 • Published • 31 -
Surge Phenomenon in Optimal Learning Rate and Batch Size Scaling
Paper • 2405.14578 • Published • 1
alignment_24_best
-
KTO: Model Alignment as Prospect Theoretic Optimization
Paper • 2402.01306 • Published • 21 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 64 -
SimPO: Simple Preference Optimization with a Reference-Free Reward
Paper • 2405.14734 • Published • 12 -
Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment
Paper • 2408.06266 • Published • 10
sft_24_best
-
Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning
Paper • 2402.04833 • Published • 5 -
A Closer Look at the Limitations of Instruction Tuning
Paper • 2402.05119 • Published • 5 -
STaR-GATE: Teaching Language Models to Ask Clarifying Questions
Paper • 2403.19154 • Published -
The Good, The Bad, and The Greedy: Evaluation of LLMs Should Not Ignore Non-Determinism
Paper • 2407.10457 • Published • 24
code_rlcef
-
OpenCodeInstruct: A Large-scale Instruction Tuning Dataset for Code LLMs
Paper • 2504.04030 • Published • 1 -
KodCode: A Diverse, Challenging, and Verifiable Synthetic Dataset for Coding
Paper • 2503.02951 • Published • 33 -
BigCodeBench: Benchmarking Code Generation with Diverse Function Calls and Complex Instructions
Paper • 2406.15877 • Published • 48 -
Magicoder: Source Code Is All You Need
Paper • 2312.02120 • Published • 82
emnlp 2023
-
DSI++: Updating Transformer Memory with New Documents
Paper • 2212.09744 • Published • 1 -
Where to start? Analyzing the potential value of intermediate models
Paper • 2211.00107 • Published -
INSTRUCTSCORE: Explainable Text Generation Evaluation with Finegrained Feedback
Paper • 2305.14282 • Published -
G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment
Paper • 2303.16634 • Published • 3
emlnp 2023 tbd
-
Can Retriever-Augmented Language Models Reason? The Blame Game Between the Retriever and the Language Model
Paper • 2212.09146 • Published • 3 -
RaLLe: A Framework for Developing and Evaluating Retrieval-Augmented Large Language Models
Paper • 2308.10633 • Published • 1 -
MemeCap: A Dataset for Captioning and Interpreting Memes
Paper • 2305.13703 • Published -
Contrastive Learning for Inference in Dialogue
Paper • 2310.12467 • Published
math
-
Why do Learning Rates Transfer? Reconciling Optimization and Scaling Limits for Deep Learning
Paper • 2402.17457 • Published -
Curvature-Informed SGD via General Purpose Lie-Group Preconditioners
Paper • 2402.04553 • Published -
TextGrad: Automatic "Differentiation" via Text
Paper • 2406.07496 • Published • 31 -
Surge Phenomenon in Optimal Learning Rate and Batch Size Scaling
Paper • 2405.14578 • Published • 1
foundation_models
-
Apple Intelligence Foundation Language Models
Paper • 2407.21075 • Published • 5 -
The Llama 3 Herd of Models
Paper • 2407.21783 • Published • 117 -
Nemotron-4 340B Technical Report
Paper • 2406.11704 • Published -
Gemma 2: Improving Open Language Models at a Practical Size
Paper • 2408.00118 • Published • 78
alignment_24_best
-
KTO: Model Alignment as Prospect Theoretic Optimization
Paper • 2402.01306 • Published • 21 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 64 -
SimPO: Simple Preference Optimization with a Reference-Free Reward
Paper • 2405.14734 • Published • 12 -
Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment
Paper • 2408.06266 • Published • 10
monte_carlo_24_best
-
Planning Like Human: A Dual-process Framework for Dialogue Planning
Paper • 2406.05374 • Published -
Plug-and-Play Policy Planner for Large Language Model Powered Dialogue Agents
Paper • 2311.00262 • Published -
Strength Lies in Differences! Towards Effective Non-collaborative Dialogues via Tailored Strategy Planning
Paper • 2403.06769 • Published -
Prompt-Based Monte-Carlo Tree Search for Goal-Oriented Dialogue Policy Planning
Paper • 2305.13660 • Published
sft_24_best
-
Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning
Paper • 2402.04833 • Published • 5 -
A Closer Look at the Limitations of Instruction Tuning
Paper • 2402.05119 • Published • 5 -
STaR-GATE: Teaching Language Models to Ask Clarifying Questions
Paper • 2403.19154 • Published -
The Good, The Bad, and The Greedy: Evaluation of LLMs Should Not Ignore Non-Determinism
Paper • 2407.10457 • Published • 24