DECO: Sparse Mixture-of-Experts with Dense-Comparable Performance on End-Side Devices Paper • 2605.10933 • Published 5 days ago • 2
PipeLLM: Fast and Confidential Large Language Model Services with Speculative Pipelined Encryption Paper • 2411.03357 • Published Nov 4, 2024
SmallThinker: A Family of Efficient Large Language Models Natively Trained for Local Deployment Paper • 2507.20984 • Published Jul 28, 2025 • 58
SmallThinker: A Family of Efficient Large Language Models Natively Trained for Local Deployment Paper • 2507.20984 • Published Jul 28, 2025 • 58
ConPET: Continual Parameter-Efficient Tuning for Large Language Models Paper • 2309.14763 • Published Sep 26, 2023 • 1
ReLU$^2$ Wins: Discovering Efficient Activation Functions for Sparse LLMs Paper • 2402.03804 • Published Feb 6, 2024 • 4
ProSparse: Introducing and Enhancing Intrinsic Activation Sparsity within Large Language Models Paper • 2402.13516 • Published Feb 21, 2024 • 1