zzfive
's Collections
Self-Rewarding Language Models
Paper
•
2401.10020
•
Published
•
142
Orion-14B: Open-source Multilingual Large Language Models
Paper
•
2401.12246
•
Published
•
10
MambaByte: Token-free Selective State Space Model
Paper
•
2401.13660
•
Published
•
49
MM-LLMs: Recent Advances in MultiModal Large Language Models
Paper
•
2401.13601
•
Published
•
44
OLMo: Accelerating the Science of Language Models
Paper
•
2402.00838
•
Published
•
79
Dolma: an Open Corpus of Three Trillion Tokens for Language Model
Pretraining Research
Paper
•
2402.00159
•
Published
•
59
In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs
Miss
Paper
•
2402.10790
•
Published
•
40
DataDreamer: A Tool for Synthetic Data Generation and Reproducible LLM
Workflows
Paper
•
2402.10379
•
Published
•
29
Paper
•
2402.13144
•
Published
•
94
How Easy is It to Fool Your Multimodal LLMs? An Empirical Analysis on
Deceptive Prompts
Paper
•
2402.13220
•
Published
•
12
LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens
Paper
•
2402.13753
•
Published
•
111
OpenCodeInterpreter: Integrating Code Generation with Execution and
Refinement
Paper
•
2402.14658
•
Published
•
82
Linear Transformers are Versatile In-Context Learners
Paper
•
2402.14180
•
Published
•
6
Watermarking Makes Language Models Radioactive
Paper
•
2402.14904
•
Published
•
22
ChunkAttention: Efficient Self-Attention with Prefix-Aware KV Cache and
Two-Phase Partition
Paper
•
2402.15220
•
Published
•
19
Genie: Generative Interactive Environments
Paper
•
2402.15391
•
Published
•
70
ChatMusician: Understanding and Generating Music Intrinsically with LLM
Paper
•
2402.16153
•
Published
•
55
MegaScale: Scaling Large Language Model Training to More Than 10,000
GPUs
Paper
•
2402.15627
•
Published
•
33
MobiLlama: Towards Accurate and Lightweight Fully Transparent GPT
Paper
•
2402.16840
•
Published
•
23
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper
•
2402.17764
•
Published
•
592
Video as the New Language for Real-World Decision Making
Paper
•
2402.17139
•
Published
•
18
Beyond Language Models: Byte Models are Digital World Simulators
Paper
•
2402.19155
•
Published
•
49
Resonance RoPE: Improving Context Length Generalization of Large
Language Models
Paper
•
2403.00071
•
Published
•
22
DenseMamba: State Space Models with Dense Hidden Connection for
Efficient Large Language Models
Paper
•
2403.00818
•
Published
•
14
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Paper
•
2403.03507
•
Published
•
182
Gemini 1.5: Unlocking multimodal understanding across millions of tokens
of context
Paper
•
2403.05530
•
Published
•
59
DeepSeek-VL: Towards Real-World Vision-Language Understanding
Paper
•
2403.05525
•
Published
•
39
Synth^2: Boosting Visual-Language Models with Synthetic Captions and
Image Embeddings
Paper
•
2403.07750
•
Published
•
21
MoAI: Mixture of All Intelligence for Large Language and Vision Models
Paper
•
2403.07508
•
Published
•
75
VisionGPT-3D: A Generalized Multimodal Agent for Enhanced 3D Vision
Understanding
Paper
•
2403.09530
•
Published
•
8
Griffon v2: Advancing Multimodal Perception with High-Resolution Scaling
and Visual-Language Co-Referring
Paper
•
2403.09333
•
Published
•
14
Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for
Large Language Models
Paper
•
2403.12881
•
Published
•
16
HyperLLaVA: Dynamic Visual and Language Expert Tuning for Multimodal
Large Language Models
Paper
•
2403.13447
•
Published
•
17
Mini-Gemini: Mining the Potential of Multi-modality Vision Language
Models
Paper
•
2403.18814
•
Published
•
44
BioMedLM: A 2.7B Parameter Language Model Trained On Biomedical Text
Paper
•
2403.18421
•
Published
•
21
Jamba: A Hybrid Transformer-Mamba Language Model
Paper
•
2403.19887
•
Published
•
103
Direct Preference Optimization of Video Large Multimodal Models from
Language Model Reward
Paper
•
2404.01258
•
Published
•
10
WavLLM: Towards Robust and Adaptive Speech Large Language Model
Paper
•
2404.00656
•
Published
•
9
CodeEditorBench: Evaluating Code Editing Capability of Large Language
Models
Paper
•
2404.03543
•
Published
•
15
LVLM-Intrepret: An Interpretability Tool for Large Vision-Language
Models
Paper
•
2404.03118
•
Published
•
23
MiniGPT4-Video: Advancing Multimodal LLMs for Video Understanding with
Interleaved Visual-Textual Tokens
Paper
•
2404.03413
•
Published
•
25
ReFT: Representation Finetuning for Language Models
Paper
•
2404.03592
•
Published
•
89
Chinese Tiny LLM: Pretraining a Chinese-Centric Large Language Model
Paper
•
2404.04167
•
Published
•
12
LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders
Paper
•
2404.05961
•
Published
•
64
Rho-1: Not All Tokens Are What You Need
Paper
•
2404.07965
•
Published
•
83
RecurrentGemma: Moving Past Transformers for Efficient Open Language
Models
Paper
•
2404.07839
•
Published
•
41
Applying Guidance in a Limited Interval Improves Sample and Distribution
Quality in Diffusion Models
Paper
•
2404.07724
•
Published
•
11
Megalodon: Efficient LLM Pretraining and Inference with Unlimited
Context Length
Paper
•
2404.08801
•
Published
•
62
TriForce: Lossless Acceleration of Long Sequence Generation with
Hierarchical Speculative Decoding
Paper
•
2404.11912
•
Published
•
16
AutoCrawler: A Progressive Understanding Web Agent for Web Crawler
Generation
Paper
•
2404.12753
•
Published
•
41
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your
Phone
Paper
•
2404.14219
•
Published
•
251
How Good Are Low-bit Quantized LLaMA3 Models? An Empirical Study
Paper
•
2404.14047
•
Published
•
43
FlowMind: Automatic Workflow Generation with LLMs
Paper
•
2404.13050
•
Published
•
32
Multi-Head Mixture-of-Experts
Paper
•
2404.15045
•
Published
•
59
WildChat: 1M ChatGPT Interaction Logs in the Wild
Paper
•
2405.01470
•
Published
•
59
Xmodel-VLM: A Simple Baseline for Multimodal Vision Language Model
Paper
•
2405.09215
•
Published
•
18
OpenRLHF: An Easy-to-use, Scalable and High-performance RLHF Framework
Paper
•
2405.11143
•
Published
•
33
Imp: Highly Capable Large Multimodal Models for Mobile Devices
Paper
•
2405.12107
•
Published
•
25
AlignGPT: Multi-modal Large Language Models with Adaptive Alignment
Capability
Paper
•
2405.14129
•
Published
•
12
ConvLLaVA: Hierarchical Backbones as Visual Encoder for Large Multimodal
Models
Paper
•
2405.15738
•
Published
•
43
Stacking Your Transformers: A Closer Look at Model Growth for Efficient
LLM Pre-Training
Paper
•
2405.15319
•
Published
•
25
NV-Embed: Improved Techniques for Training LLMs as Generalist Embedding
Models
Paper
•
2405.17428
•
Published
•
16
Value-Incentivized Preference Optimization: A Unified Approach to Online
and Offline RLHF
Paper
•
2405.19320
•
Published
•
9
Offline Regularised Reinforcement Learning for Large Language Models
Alignment
Paper
•
2405.19107
•
Published
•
13
Show, Don't Tell: Aligning Language Models with Demonstrated Feedback
Paper
•
2406.00888
•
Published
•
30
Xmodel-LM Technical Report
Paper
•
2406.02856
•
Published
•
7
Mixture-of-Agents Enhances Large Language Model Capabilities
Paper
•
2406.04692
•
Published
•
54
Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts
Language Models
Paper
•
2406.06563
•
Published
•
17
Turbo Sparse: Achieving LLM SOTA Performance with Minimal Activated
Parameters
Paper
•
2406.05955
•
Published
•
22
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs
with Nothing
Paper
•
2406.08464
•
Published
•
62
Discovering Preference Optimization Algorithms with and for Large
Language Models
Paper
•
2406.08414
•
Published
•
12
HelpSteer2: Open-source dataset for training top-performing reward
models
Paper
•
2406.08673
•
Published
•
16
Samba: Simple Hybrid State Space Models for Efficient Unlimited Context
Language Modeling
Paper
•
2406.07522
•
Published
•
36
Self-play with Execution Feedback: Improving Instruction-following
Capabilities of Large Language Models
Paper
•
2406.13542
•
Published
•
16
Iterative Length-Regularized Direct Preference Optimization: A Case
Study on Improving 7B Language Models to GPT-4 Level
Paper
•
2406.11817
•
Published
•
13
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper
•
2406.15319
•
Published
•
60
BigCodeBench: Benchmarking Code Generation with Diverse Function Calls
and Complex Instructions
Paper
•
2406.15877
•
Published
•
45
Scaling Laws for Linear Complexity Language Models
Paper
•
2406.16690
•
Published
•
22
Sparser is Faster and Less is More: Efficient Sparse Attention for
Long-Range Transformers
Paper
•
2406.16747
•
Published
•
17
OlympicArena Medal Ranks: Who Is the Most Intelligent AI So Far?
Paper
•
2406.16772
•
Published
•
2
Unlocking Continual Learning Abilities in Language Models
Paper
•
2406.17245
•
Published
•
28
Direct Preference Knowledge Distillation for Large Language Models
Paper
•
2406.19774
•
Published
•
21
AutoRAG-HP: Automatic Online Hyper-Parameter Tuning for
Retrieval-Augmented Generation
Paper
•
2406.19251
•
Published
•
8
RegMix: Data Mixture as Regression for Language Model Pre-training
Paper
•
2407.01492
•
Published
•
33
Step-Controlled DPO: Leveraging Stepwise Error for Enhanced Mathematical
Reasoning
Paper
•
2407.00782
•
Published
•
23
DogeRM: Equipping Reward Models with Domain Knowledge through Model
Merging
Paper
•
2407.01470
•
Published
•
5
Summary of a Haystack: A Challenge to Long-Context LLMs and RAG Systems
Paper
•
2407.01370
•
Published
•
85
MInference 1.0: Accelerating Pre-filling for Long-Context LLMs via
Dynamic Sparse Attention
Paper
•
2407.02490
•
Published
•
23
To Forget or Not? Towards Practical Knowledge Unlearning for Large
Language Models
Paper
•
2407.01920
•
Published
•
13
Eliminating Position Bias of Language Models: A Mechanistic Approach
Paper
•
2407.01100
•
Published
•
6
DotaMath: Decomposition of Thought with Code Assistance and
Self-correction for Mathematical Reasoning
Paper
•
2407.04078
•
Published
•
16
LLaMAX: Scaling Linguistic Horizons of LLM by Enhancing Translation
Capabilities Beyond 100 Languages
Paper
•
2407.05975
•
Published
•
34
InverseCoder: Unleashing the Power of Instruction-Tuned Code LLMs with
Inverse-Instruct
Paper
•
2407.05700
•
Published
•
9
PAS: Data-Efficient Plug-and-Play Prompt Augmentation System
Paper
•
2407.06027
•
Published
•
8
Lookback Lens: Detecting and Mitigating Contextual Hallucinations in
Large Language Models Using Only Attention Maps
Paper
•
2407.07071
•
Published
•
11
AgentInstruct: Toward Generative Teaching with Agentic Flows
Paper
•
2407.03502
•
Published
•
43
Inference Performance Optimization for Large Language Models on CPUs
Paper
•
2407.07304
•
Published
•
52
SpreadsheetLLM: Encoding Spreadsheets for Large Language Models
Paper
•
2407.09025
•
Published
•
125
Human-like Episodic Memory for Infinite Context LLMs
Paper
•
2407.09450
•
Published
•
56
MUSCLE: A Model Update Strategy for Compatible LLM Evolution
Paper
•
2407.09435
•
Published
•
20
Transformer Layers as Painters
Paper
•
2407.09298
•
Published
•
13
H2O-Danube3 Technical Report
Paper
•
2407.09276
•
Published
•
18
Understanding Retrieval Robustness for Retrieval-Augmented Image
Captioning
Paper
•
2406.02265
•
Published
•
6
Characterizing Prompt Compression Methods for Long Context Inference
Paper
•
2407.08892
•
Published
•
9
Paper
•
2407.10671
•
Published
•
154
Learning to Refuse: Towards Mitigating Privacy Risks in LLMs
Paper
•
2407.10058
•
Published
•
29
Q-Sparse: All Large Language Models can be Fully Sparsely-Activated
Paper
•
2407.10969
•
Published
•
20
The Good, The Bad, and The Greedy: Evaluation of LLMs Should Not Ignore
Non-Determinism
Paper
•
2407.10457
•
Published
•
22
Foundational Autoraters: Taming Large Language Models for Better
Automatic Evaluation
Paper
•
2407.10817
•
Published
•
13
MMM: Multilingual Mutual Reinforcement Effect Mix Datasets & Test with
Open-domain Information Extraction Large Language Models
Paper
•
2407.10953
•
Published
•
4
Spectra: A Comprehensive Study of Ternary, Quantized, and FP16 Language
Models
Paper
•
2407.12327
•
Published
•
76
GoldFinch: High Performance RWKV/Transformer Hybrid with Linear Pre-Fill
and Extreme KV-Cache Compression
Paper
•
2407.12077
•
Published
•
52
Patch-Level Training for Large Language Models
Paper
•
2407.12665
•
Published
•
16
The Art of Saying No: Contextual Noncompliance in Language Models
Paper
•
2407.12043
•
Published
•
4
Practical Unlearning for Large Language Models
Paper
•
2407.10223
•
Published
•
4
Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies
Paper
•
2407.13623
•
Published
•
52
Understanding Reference Policies in Direct Preference Optimization
Paper
•
2407.13709
•
Published
•
16
Internal Consistency and Self-Feedback in Large Language Models: A
Survey
Paper
•
2407.14507
•
Published
•
44
SciCode: A Research Coding Benchmark Curated by Scientists
Paper
•
2407.13168
•
Published
•
13
Phi-3 Safety Post-Training: Aligning Language Models with a "Break-Fix"
Cycle
Paper
•
2407.13833
•
Published
•
11
Knowledge Mechanisms in Large Language Models: A Survey and Perspective
Paper
•
2407.15017
•
Published
•
33
Compact Language Models via Pruning and Knowledge Distillation
Paper
•
2407.14679
•
Published
•
35
BOND: Aligning LLMs with Best-of-N Distillation
Paper
•
2407.14622
•
Published
•
17
DDK: Distilling Domain Knowledge for Efficient Large Language Models
Paper
•
2407.16154
•
Published
•
20
Data Mixture Inference: What do BPE Tokenizers Reveal about their
Training Data?
Paper
•
2407.16607
•
Published
•
21
SeaLLMs 3: Open Foundation and Chat Multilingual Large Language Models
for Southeast Asian Languages
Paper
•
2407.19672
•
Published
•
54
Self-Training with Direct Preference Optimization Improves
Chain-of-Thought Reasoning
Paper
•
2407.18248
•
Published
•
30
Mixture of Nested Experts: Adaptive Processing of Visual Tokens
Paper
•
2407.19985
•
Published
•
34
Visual Riddles: a Commonsense and World Knowledge Challenge for Large
Vision and Language Models
Paper
•
2407.19474
•
Published
•
22
ThinK: Thinner Key Cache by Query-Driven Pruning
Paper
•
2407.21018
•
Published
•
30
The Llama 3 Herd of Models
Paper
•
2407.21783
•
Published
•
103
ShieldGemma: Generative AI Content Moderation Based on Gemma
Paper
•
2407.21772
•
Published
•
13
Gemma 2: Improving Open Language Models at a Practical Size
Paper
•
2408.00118
•
Published
•
73
Improving Text Embeddings for Smaller Language Models Using Contrastive
Fine-tuning
Paper
•
2408.00690
•
Published
•
21
Unleashing the Power of Data Tsunami: A Comprehensive Survey on Data
Assessment and Selection for Instruction Tuning of Language Models
Paper
•
2408.02085
•
Published
•
17
Paper
•
2408.02666
•
Published
•
25
Scaling LLM Test-Time Compute Optimally can be More Effective than
Scaling Model Parameters
Paper
•
2408.03314
•
Published
•
33
Transformer Explainer: Interactive Learning of Text-Generative Models
Paper
•
2408.04619
•
Published
•
154
Better Alignment with Instruction Back-and-Forth Translation
Paper
•
2408.04614
•
Published
•
14
Learning to Predict Program Execution by Modeling Dynamic Dependency on
Code Graphs
Paper
•
2408.02816
•
Published
•
4
Gemma Scope: Open Sparse Autoencoders Everywhere All At Once on Gemma 2
Paper
•
2408.05147
•
Published
•
37
ToolSandbox: A Stateful, Conversational, Interactive Evaluation
Benchmark for LLM Tool Use Capabilities
Paper
•
2408.04682
•
Published
•
14
Mutual Reasoning Makes Smaller LLMs Stronger Problem-Solvers
Paper
•
2408.06195
•
Published
•
58
LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
Paper
•
2408.07055
•
Published
•
65
Layerwise Recurrent Router for Mixture-of-Experts
Paper
•
2408.06793
•
Published
•
30
Amuro & Char: Analyzing the Relationship between Pre-Training and
Fine-Tuning of Large Language Models
Paper
•
2408.06663
•
Published
•
15
FuxiTranyu: A Multilingual Large Language Model Trained with Balanced
Data
Paper
•
2408.06273
•
Published
•
9
Paper
•
2408.07410
•
Published
•
13
DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for
Reinforcement Learning and Monte-Carlo Tree Search
Paper
•
2408.08152
•
Published
•
51
I-SHEEP: Self-Alignment of LLM from Scratch through an Iterative
Self-Enhancement Paradigm
Paper
•
2408.08072
•
Published
•
31
Training Language Models on the Knowledge Graph: Insights on
Hallucinations and Their Detectability
Paper
•
2408.07852
•
Published
•
14
FuseChat: Knowledge Fusion of Chat Models
Paper
•
2408.07990
•
Published
•
9
BAM! Just Like That: Simple and Efficient Parameter Upcycling for
Mixture of Experts
Paper
•
2408.08274
•
Published
•
11
Cybench: A Framework for Evaluating Cybersecurity Capabilities and Risk
of Language Models
Paper
•
2408.08926
•
Published
•
4
TableBench: A Comprehensive and Complex Benchmark for Table Question
Answering
Paper
•
2408.09174
•
Published
•
51
To Code, or Not To Code? Exploring Impact of Code in Pre-training
Paper
•
2408.10914
•
Published
•
40
MagicDec: Breaking the Latency-Throughput Tradeoff for Long Context
Generation with Speculative Decoding
Paper
•
2408.11049
•
Published
•
10
LLM Pruning and Distillation in Practice: The Minitron Approach
Paper
•
2408.11796
•
Published
•
53
FocusLLM: Scaling LLM's Context by Parallel Decoding
Paper
•
2408.11745
•
Published
•
23
Hermes 3 Technical Report
Paper
•
2408.11857
•
Published
•
36
ConflictBank: A Benchmark for Evaluating the Influence of Knowledge
Conflicts in LLM
Paper
•
2408.12076
•
Published
•
11
Memory-Efficient LLM Training with Online Subspace Descent
Paper
•
2408.12857
•
Published
•
10
SWE-bench-java: A GitHub Issue Resolving Benchmark for Java
Paper
•
2408.14354
•
Published
•
40
LlamaDuo: LLMOps Pipeline for Seamless Migration from Service LLMs to
Small-Scale Local LLMs
Paper
•
2408.13467
•
Published
•
23
MobileQuant: Mobile-friendly Quantization for On-device Language Models
Paper
•
2408.13933
•
Published
•
13
Efficient Detection of Toxic Prompts in Large Language Models
Paper
•
2408.11727
•
Published
•
11
Writing in the Margins: Better Inference Pattern for Long Context
Retrieval
Paper
•
2408.14906
•
Published
•
138
The Mamba in the Llama: Distilling and Accelerating Hybrid Models
Paper
•
2408.15237
•
Published
•
36
BaichuanSEED: Sharing the Potential of ExtensivE Data Collection and
Deduplication by Introducing a Competitive Large Language Model Baseline
Paper
•
2408.15079
•
Published
•
51
Leveraging Open Knowledge for Advancing Task Expertise in Large Language
Models
Paper
•
2408.15915
•
Published
•
19
Efficient LLM Scheduling by Learning to Rank
Paper
•
2408.15792
•
Published
•
19
Knowledge Navigator: LLM-guided Browsing Framework for Exploratory
Search in Scientific Literature
Paper
•
2408.15836
•
Published
•
11
ReMamba: Equip Mamba with Effective Long-Sequence Modeling
Paper
•
2408.15496
•
Published
•
10
Auxiliary-Loss-Free Load Balancing Strategy for Mixture-of-Experts
Paper
•
2408.15664
•
Published
•
11
SciLitLLM: How to Adapt LLMs for Scientific Literature Understanding
Paper
•
2408.15545
•
Published
•
33
CURLoRA: Stable LLM Continual Fine-Tuning and Catastrophic Forgetting
Mitigation
Paper
•
2408.14572
•
Published
•
7
GIFT-SW: Gaussian noise Injected Fine-Tuning of Salient Weights for LLMs
Paper
•
2408.15300
•
Published
•
3
OLMoE: Open Mixture-of-Experts Language Models
Paper
•
2409.02060
•
Published
•
77
LongRecipe: Recipe for Efficient Long Context Generalization in Large
Languge Models
Paper
•
2409.00509
•
Published
•
38
ContextCite: Attributing Model Generation to Context
Paper
•
2409.00729
•
Published
•
13
PrivacyLens: Evaluating Privacy Norm Awareness of Language Models in
Action
Paper
•
2409.00138
•
Published
•
1
LongCite: Enabling LLMs to Generate Fine-grained Citations in
Long-context QA
Paper
•
2409.02897
•
Published
•
44
Arctic-SnowCoder: Demystifying High-Quality Data in Code Pretraining
Paper
•
2409.02326
•
Published
•
16
Attention Heads of Large Language Models: A Survey
Paper
•
2409.03752
•
Published
•
86
WildVis: Open Source Visualizer for Million-Scale Chat Logs in the Wild
Paper
•
2409.03753
•
Published
•
18
How Do Your Code LLMs Perform? Empowering Code Instruction Tuning with
High-Quality Data
Paper
•
2409.03810
•
Published
•
30
Configurable Foundation Models: Building LLMs from a Modular Perspective
Paper
•
2409.02877
•
Published
•
27
Spinning the Golden Thread: Benchmarking Long-Form Generation in
Language Models
Paper
•
2409.02076
•
Published
•
9
Towards a Unified View of Preference Learning for Large Language Models:
A Survey
Paper
•
2409.02795
•
Published
•
72
MemoRAG: Moving towards Next-Gen RAG Via Memory-Inspired Knowledge
Discovery
Paper
•
2409.05591
•
Published
•
26
Benchmarking Chinese Knowledge Rectification in Large Language Models
Paper
•
2409.05806
•
Published
•
14
GroUSE: A Benchmark to Evaluate Evaluators in Grounded Question
Answering
Paper
•
2409.06595
•
Published
•
37
PingPong: A Benchmark for Role-Playing Language Models with User
Emulation and Multi-Model Evaluation
Paper
•
2409.06820
•
Published
•
59
Gated Slot Attention for Efficient Linear-Time Sequence Modeling
Paper
•
2409.07146
•
Published
•
19
Self-Harmonized Chain of Thought
Paper
•
2409.04057
•
Published
•
16
Source2Synth: Synthetic Data Generation and Curation Grounded in Real
Data Sources
Paper
•
2409.08239
•
Published
•
15
Ferret: Federated Full-Parameter Tuning at Scale for Large Language
Models
Paper
•
2409.06277
•
Published
•
14
On the Diagram of Thought
Paper
•
2409.10038
•
Published
•
10
A Comprehensive Evaluation of Quantized Instruction-Tuned Large Language
Models: An Experimental Analysis up to 405B
Paper
•
2409.11055
•
Published
•
16
Measuring and Enhancing Trustworthiness of LLMs in RAG through Grounded
Attributions and Learning to Refuse
Paper
•
2409.11242
•
Published
•
4
Qwen2.5-Coder Technical Report
Paper
•
2409.12186
•
Published
•
121
LLMs + Persona-Plug = Personalized LLMs
Paper
•
2409.11901
•
Published
•
30
Preference Tuning with Human Feedback on Language, Speech, and Vision
Tasks: A Survey
Paper
•
2409.11564
•
Published
•
18
GRIN: GRadient-INformed MoE
Paper
•
2409.12136
•
Published
•
14
Training Language Models to Self-Correct via Reinforcement Learning
Paper
•
2409.12917
•
Published
•
128
MMSearch: Benchmarking the Potential of Large Models as Multi-modal
Search Engines
Paper
•
2409.12959
•
Published
•
35
Scaling Smart: Accelerating Large Language Model Pre-training with Small
Model Initialization
Paper
•
2409.12903
•
Published
•
20
Language Models Learn to Mislead Humans via RLHF
Paper
•
2409.12822
•
Published
•
9
Fact, Fetch, and Reason: A Unified Evaluation of Retrieval-Augmented
Generation
Paper
•
2409.12941
•
Published
•
20
Hackphyr: A Local Fine-Tuned LLM Agent for Network Security Environments
Paper
•
2409.11276
•
Published
•
6
HelloBench: Evaluating Long Text Generation Capabilities of Large
Language Models
Paper
•
2409.16191
•
Published
•
40
Reward-Robust RLHF in LLMs
Paper
•
2409.15360
•
Published
•
4
Programming Every Example: Lifting Pre-training Data Quality like
Experts at Scale
Paper
•
2409.17115
•
Published
•
58
Boosting Healthcare LLMs Through Retrieved Context
Paper
•
2409.15127
•
Published
•
18
NoTeeline: Supporting Real-Time Notetaking from Keypoints with Large
Language Models
Paper
•
2409.16493
•
Published
•
7
HyperAgent: Generalist Software Engineering Agents to Solve Coding Tasks
at Scale
Paper
•
2409.16299
•
Published
•
9
MaskLLM: Learnable Semi-Structured Sparsity for Large Language Models
Paper
•
2409.17481
•
Published
•
43
The Imperative of Conversation Analysis in the Era of LLMs: A Survey of
Tasks, Techniques, and Trends
Paper
•
2409.14195
•
Published
•
10
Enhancing Structured-Data Retrieval with GraphRAG: Soccer Data Case
Study
Paper
•
2409.17580
•
Published
•
6
Modulated Intervention Preference Optimization (MIPO): Keep the Easy,
Refine the Difficult
Paper
•
2409.17545
•
Published
•
16