.taskmaster/config.json
File (Recommended - New Structure)
.taskmaster/
directory when you run the task-master models --setup
interactive setup or initialize a new project with task-master init
..taskmasterconfig
in the root will continue to work, but should be migrated to the new structure using task-master migrate
.task-master models --setup
command (or models
MCP tool) to interactively create and manage this file. You can also set specific models directly using task-master models --set-<role>=<model_id>
, adding --ollama
or --openrouter
flags for custom models. Manual editing is possible but not recommended unless you understand the structure..taskmasterconfig
File (Backward Compatibility)
task-master migrate
to move this to .taskmaster/config.json
..env
file or MCP env
block - For API Keys Only).env
file in your project root.env
section of your .cursor/mcp.json
file.ANTHROPIC_API_KEY
: Your Anthropic API key.PERPLEXITY_API_KEY
: Your Perplexity API key.OPENAI_API_KEY
: Your OpenAI API key.GOOGLE_API_KEY
: Your Google API key (also used for Vertex AI provider).MISTRAL_API_KEY
: Your Mistral API key.AZURE_OPENAI_API_KEY
: Your Azure OpenAI API key (also requires AZURE_OPENAI_ENDPOINT
).OPENROUTER_API_KEY
: Your OpenRouter API key.XAI_API_KEY
: Your X-AI API key.baseURL
in .taskmasterconfig
: You can add a baseURL
property to any model role (main
, research
, fallback
) to override the default API endpoint for that provider. If omitted, the provider’s standard endpoint is used.<PROVIDER>_BASE_URL
): For greater flexibility, especially with third-party services, you can set an environment variable like OPENAI_BASE_URL
or MISTRAL_BASE_URL
. This will override any baseURL
set in the configuration file for that provider. This is the recommended way to connect to OpenAI-compatible APIs.AZURE_OPENAI_ENDPOINT
: Required if using Azure OpenAI key (can also be set as baseURL
for the Azure model role).OLLAMA_BASE_URL
: Override the default Ollama API URL (Default: http://localhost:11434/api
).VERTEX_PROJECT_ID
: Your Google Cloud project ID for Vertex AI. Required when using the ‘vertex’ provider.VERTEX_LOCATION
: Google Cloud region for Vertex AI (e.g., ‘us-central1’). Default is ‘us-central1’.GOOGLE_APPLICATION_CREDENTIALS
: Path to service account credentials JSON file for Google Cloud auth (alternative to API key for Vertex AI).main
, research
, fallback
), maxTokens
, temperature
, logLevel
, defaultSubtasks
, defaultPriority
, and projectName
are managed in .taskmaster/config.json
(or .taskmasterconfig
for unmigrated projects), not environment variables.
defaultTag
(string): Default tag context for new operations (default: “master”)--from-branch
option:
task-master add-tag --from-branch
to create a tag based on your current git branch name.taskmaster/state.json
to track tagged system runtime information:
currentTag
: Currently active tag contextlastSwitched
: Timestamp of last tag switchmigrationNoticeShown
: Whether migration notice has been displayed.env
File (for API Keys)task-master models --setup
in your project root to create or repair the file..taskmaster/config.json
. For legacy projects, you may want to use task-master migrate
to move to the new structure..env
file (for CLI) or .cursor/mcp.json
(for MCP) and are valid for the providers selected in your config file.task-master init
doesn’t respond:GOOGLE_API_KEY
environment variableGOOGLE_APPLICATION_CREDENTIALS
to point to your service account JSON fileVERTEX_PROJECT_ID
to your Google Cloud project IDVERTEX_LOCATION
to your preferred Google Cloud region (default: us-central1)AZURE_OPENAI_API_KEY
environment variable with your Azure OpenAI API keymodelId
in your configuration should match the deployment name you created in Azure OpenAI Studio, not the underlying model namebaseURL
settings override the global azureBaseURL
settingbaseURL
, use the full path including /openai/deployments
baseURL
includes the full path: https://your-resource-name.openai.azure.com/openai/deployments
modelId
exactly matches what’s configured in Azure OpenAI StudioAZURE_OPENAI_API_KEY
is correct and has not expiredmaxTokens
maintain appropriate Tokens per Minute Rate Limit (TPM) in your deployment.