feat(ai): Enhance Perplexity research calls & fix docs examples
Improves the quality and relevance of research-backed AI operations: - Tweaks Perplexity AI calls to use max input tokens (8700), temperature 0.1, high context size, and day-fresh search recency. - Adds a system prompt to guide Perplexity research output. Docs: - Updates CLI examples in taskmaster.mdc to use ANSI-C quoting ($'...') for multi-line prompts, ensuring they work correctly in bash/zsh.
This commit is contained in:
@@ -709,12 +709,24 @@ Include concrete code examples and technical considerations where relevant.`;
|
||||
const researchResponse = await perplexityClient.chat.completions.create({
|
||||
model: PERPLEXITY_MODEL,
|
||||
messages: [
|
||||
{
|
||||
role: 'system',
|
||||
content: `You are a helpful assistant that provides research on current best practices and implementation approaches for software development.
|
||||
You are given a task and a description of the task.
|
||||
You need to provide a list of best practices, libraries, design patterns, and implementation approaches that are relevant to the task.
|
||||
You should provide concrete code examples and technical considerations where relevant.`
|
||||
},
|
||||
{
|
||||
role: 'user',
|
||||
content: researchQuery
|
||||
}
|
||||
],
|
||||
temperature: 0.1 // Lower temperature for more factual responses
|
||||
temperature: 0.1, // Lower temperature for more factual responses
|
||||
max_tokens: 8700, // Respect maximum input tokens for Perplexity (8719 max)
|
||||
web_search_options: {
|
||||
search_context_size: 'high'
|
||||
},
|
||||
search_recency_filter: 'day' // Filter for results that are as recent as today to capture new releases
|
||||
});
|
||||
|
||||
const researchResult = researchResponse.choices[0].message.content;
|
||||
@@ -814,7 +826,7 @@ Note on dependencies: Subtasks can depend on other subtasks with lower IDs. Use
|
||||
anthropic,
|
||||
{
|
||||
model: session?.env?.ANTHROPIC_MODEL || CONFIG.model,
|
||||
max_tokens: session?.env?.MAX_TOKENS || CONFIG.maxTokens,
|
||||
max_tokens: 8700,
|
||||
temperature: session?.env?.TEMPERATURE || CONFIG.temperature,
|
||||
system: systemPrompt,
|
||||
messages: [{ role: 'user', content: userPrompt }]
|
||||
@@ -1328,7 +1340,12 @@ Include concrete code examples and technical considerations where relevant.`;
|
||||
content: researchQuery
|
||||
}
|
||||
],
|
||||
temperature: 0.1 // Lower temperature for more factual responses
|
||||
temperature: 0.1, // Lower temperature for more factual responses
|
||||
max_tokens: 8700, // Respect maximum input tokens for Perplexity (8719 max)
|
||||
web_search_options: {
|
||||
search_context_size: 'high'
|
||||
},
|
||||
search_recency_filter: 'day' // Filter for results that are as recent as today to capture new releases
|
||||
});
|
||||
|
||||
const researchResult = researchResponse.choices[0].message.content;
|
||||
|
||||
Reference in New Issue
Block a user