Auxiliary-Loss-Free Load Balancing Strategy for Mixture-of-Experts Paper • 2408.15664 • Published Aug 28, 2024 • 11
Auxiliary-Loss-Free Load Balancing Strategy for Mixture-of-Experts Paper • 2408.15664 • Published Aug 28, 2024 • 11
DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence Paper • 2406.11931 • Published Jun 17, 2024 • 58
Calibrating Factual Knowledge in Pretrained Language Models Paper • 2210.03329 • Published Oct 7, 2022 • 1
Why Can GPT Learn In-Context? Language Models Implicitly Perform Gradient Descent as Meta-Optimizers Paper • 2212.10559 • Published Dec 20, 2022
Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning Paper • 2305.14160 • Published May 23, 2023 • 1
Math-Shepherd: Verify and Reinforce LLMs Step-by-step without Human Annotations Paper • 2312.08935 • Published Dec 14, 2023 • 4
On the Representation Collapse of Sparse Mixture of Experts Paper • 2204.09179 • Published Apr 20, 2022 • 1
StableMoE: Stable Routing Strategy for Mixture of Experts Paper • 2204.08396 • Published Apr 18, 2022 • 1
DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model Paper • 2405.04434 • Published May 7, 2024 • 14
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models Paper • 2401.06066 • Published Jan 11, 2024 • 44
DeepSeek LLM: Scaling Open-Source Language Models with Longtermism Paper • 2401.02954 • Published Jan 5, 2024 • 41