Papers
-
WorldVQA: Measuring Atomic World Knowledge in Multimodal Large Language Models
-
Towards Pixel-Level VLM Perception via Simple Points Prediction
-
BabyVision: Visual Reasoning Beyond Language
-
Kimi-Dev: Agentless Training as Skill Prior for SWE-Agents
-
Kimi Linear: An Expressive, Efficient Attention Architecture
-
Seer: Online Context Learning for Fast Synchronous LLM Reinforcement Learning
-
Mooncake: A KVCache-centric Disaggregated Architecture for LLM Serving
-
Kimi K2: Open Agentic Intelligence
-
More is not always better? Enhancing Many-Shot In-Context Learning with Differentiated and Reweighting Objectives
-
G1: Bootstrapping Perception and Reasoning Abilities of Vision-Language Model via Reinforcement Learning
-
Kimi-Audio Technical Report
-
Muon is Scalable for LLM Training
-
MoBA: Mixture of Block Attention for Long-Context LLMs
