Files
claude-task-master/.changeset/quiet-rabbits-bathe.md

530 B

task-master-ai
task-master-ai
patch

Fix max_tokens limits for OpenRouter and Groq models

  • Add special handling in config-manager.js for custom OpenRouter models to use a conservative default of 32,768 max_tokens
  • Update qwen/qwen-turbo model max_tokens from 1,000,000 to 32,768 to match OpenRouter's actual limits
  • Fix moonshotai/kimi-k2-instruct max_tokens to 16,384 to match Groq's actual limit (fixes #1028)
  • This prevents "maximum context length exceeded" errors when using OpenRouter models not in our supported models list