qiuwenbo
's Collections
interest_need_read
updated
ProcessBench: Identifying Process Errors in Mathematical Reasoning
Paper
•
2412.06559
•
Published
•
72
Maya: An Instruction Finetuned Multilingual Multimodal Model
Paper
•
2412.07112
•
Published
•
26
Paper
•
2412.16720
•
Published
•
29
Diving into Self-Evolving Training for Multimodal Reasoning
Paper
•
2412.17451
•
Published
•
41
B-STaR: Monitoring and Balancing Exploration and Exploitation in
Self-Taught Reasoners
Paper
•
2412.17256
•
Published
•
44
Multi-LLM Text Summarization
Paper
•
2412.15487
•
Published
•
5
Offline Reinforcement Learning for LLM Multi-Step Reasoning
Paper
•
2412.16145
•
Published
•
37
MegaPairs: Massive Data Synthesis For Universal Multimodal Retrieval
Paper
•
2412.14475
•
Published
•
52
Progressive Multimodal Reasoning via Active Retrieval
Paper
•
2412.14835
•
Published
•
71
Paper
•
2412.15115
•
Published
•
337
VidTok: A Versatile and Open-Source Video Tokenizer
Paper
•
2412.13061
•
Published
•
8
Paper
•
2412.13501
•
Published
•
23
Smarter, Better, Faster, Longer: A Modern Bidirectional Encoder for
Fast, Memory Efficient, and Long Context Finetuning and Inference
Paper
•
2412.13663
•
Published
•
119
Compressed Chain of Thought: Efficient Reasoning Through Dense
Representations
Paper
•
2412.13171
•
Published
•
31
Reliable, Reproducible, and Really Fast Leaderboards with Evalica
Paper
•
2412.11314
•
Published
•
2
The Open Source Advantage in Large Language Models (LLMs)
Paper
•
2412.12004
•
Published
•
9
SPaR: Self-Play with Tree-Search Refinement to Improve
Instruction-Following in Large Language Models
Paper
•
2412.11605
•
Published
•
16
RetroLLM: Empowering Large Language Models to Retrieve Fine-grained
Evidence within Generation
Paper
•
2412.11919
•
Published
•
33
Smaller Language Models Are Better Instruction Evolvers
Paper
•
2412.11231
•
Published
•
27
Apollo: An Exploration of Video Understanding in Large Multimodal Models
Paper
•
2412.10360
•
Published
•
136
Multimodal Latent Language Modeling with Next-Token Diffusion
Paper
•
2412.08635
•
Published
•
42
Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Paper
•
2412.09501
•
Published
•
43
Euclid: Supercharging Multimodal LLMs with Synthetic High-Fidelity
Visual Descriptions
Paper
•
2412.08737
•
Published
•
52
InternLM-XComposer2.5-OmniLive: A Comprehensive Multimodal System for
Long-term Streaming Video and Audio Interactions
Paper
•
2412.09596
•
Published
•
92
Paper
•
2412.08905
•
Published
•
97
Chimera: Improving Generalist Model with Domain-Specific Experts
Paper
•
2412.05983
•
Published
•
9
Evaluating and Aligning CodeLLMs on Human Preference
Paper
•
2412.05210
•
Published
•
47
Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs
Paper
•
2412.21187
•
Published
•
26
CypherBench: Towards Precise Retrieval over Full-scale Modern Knowledge
Graphs in the LLM Era
Paper
•
2412.18702
•
Published
•
5
Next Token Prediction Towards Multimodal Intelligence: A Comprehensive
Survey
Paper
•
2412.18619
•
Published
•
49
HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs
Paper
•
2412.18925
•
Published
•
86
MMFactory: A Universal Solution Search Engine for Vision-Language Tasks
Paper
•
2412.18072
•
Published
•
14
YuLan-Mini: An Open Data-efficient Language Model
Paper
•
2412.17743
•
Published
•
60
Mulberry: Empowering MLLM with o1-like Reasoning and Reflection via
Collective Monte Carlo Tree Search
Paper
•
2412.18319
•
Published
•
34
Bridging the Data Provenance Gap Across Text, Speech and Video
Paper
•
2412.17847
•
Published
•
7
SKETCH: Structured Knowledge Enhanced Text Comprehension for Holistic
Retrieval
Paper
•
2412.15443
•
Published
•
8
Ensembling Large Language Models with Process Reward-Guided Tree Search
for Better Complex Reasoning
Paper
•
2412.15797
•
Published
•
16
OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks
with Reinforcement Fine-Tuning
Paper
•
2412.16849
•
Published
•
7
Outcome-Refining Process Supervision for Code Generation
Paper
•
2412.15118
•
Published
•
19
DRT-o1: Optimized Deep Reasoning Translation via Long Chain-of-Thought
Paper
•
2412.17498
•
Published
•
21