Submitted by akhaliq 86 Eagle: Exploring The Design Space for Multimodal LLMs with Mixture of Encoders · 15 authors 929 3
Submitted by dongguosheng 54 BaichuanSEED: Sharing the Potential of ExtensivE Data Collection and Deduplication by Introducing a Competitive Large Language Model Baseline · 20 authors 4
Submitted by akhaliq 42 Dolphin: Long Context as a New Modality for Energy-Efficient On-Device Language Models · 4 authors 4
Submitted by akhaliq 21 LLaVA-MoD: Making LLaVA Tiny via MoE Knowledge Distillation · 16 authors 220 2
Submitted by yolay 19 Leveraging Open Knowledge for Advancing Task Expertise in Large Language Models · 12 authors 5 4
Submitted by SYZhang0805 16 Distribution Backtracking Builds A Faster Convergence Trajectory for One-step Diffusion Distillation · 9 authors 32 2
Submitted by osanseviero 15 Auxiliary-Loss-Free Load Balancing Strategy for Mixture-of-Experts · 5 authors 3
Submitted by Uri-ka 14 Knowledge Navigator: LLM-guided Browsing Framework for Exploratory Search in Scientific Literature · 3 authors 4
Submitted by akhaliq 8 Towards Realistic Example-based Modeling via 3D Gaussian Stitching · 6 authors 3