0xSero/MiniMax-M2.1-REAP-50-W4A16-REPAIR-IN-PROGRESS Text Generation • 17B • Updated 8 days ago • 2.01k • 27
view reply llama-server by default in most implementation keeps the reasoning content in reasoning_content variable in response attribute. You can get it from there. Otherwise use reasoning-format flag and pass DeepSeek value to get pure tokens
MiniCPM-V 4.5: Cooking Efficient MLLMs via Architecture, Data, and Training Recipe Paper • 2509.18154 • Published Sep 16, 2025 • 52