mirror of
https://github.com/eyaltoledano/claude-task-master.git
synced 2026-01-06 04:01:34 +00:00
Version Packages (#1367)
Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: Ralph Khreish <35776126+Crunchyman-ralph@users.noreply.github.com>
This commit is contained in:
parent
98df39c014
commit
47c5b1e2ba
@ -1,29 +0,0 @@
|
||||
---
|
||||
"task-master-ai": minor
|
||||
---
|
||||
|
||||
Add support for custom OpenAI-compatible providers, allowing you to connect Task Master to any service that implements the OpenAI API specification
|
||||
|
||||
**How to use:**
|
||||
|
||||
Configure your custom provider with the `models` command:
|
||||
|
||||
```bash
|
||||
task-master models --set-main <your-model-id> --openai-compatible --baseURL <your-api-endpoint>
|
||||
```
|
||||
|
||||
Example:
|
||||
|
||||
```bash
|
||||
task-master models --set-main llama-3-70b --openai-compatible --baseURL http://localhost:8000/v1
|
||||
# Or for an interactive view
|
||||
task-master models --setup
|
||||
```
|
||||
|
||||
Set your API key (if required by your provider) in mcp.json, your .env file or in your env exports:
|
||||
|
||||
```bash
|
||||
OPENAI_COMPATIBLE_API_KEY="your-key-here"
|
||||
```
|
||||
|
||||
This gives you the flexibility to use virtually any LLM service with Task Master, whether it's self-hosted, a specialized provider, or a custom inference server.
|
||||
@ -1,7 +0,0 @@
|
||||
---
|
||||
"task-master-ai": patch
|
||||
---
|
||||
|
||||
Improve parse PRD schema for better llm model compatiblity
|
||||
|
||||
- Fixes #1353
|
||||
@ -1,7 +0,0 @@
|
||||
---
|
||||
"task-master-ai": patch
|
||||
---
|
||||
|
||||
Fix subtask ID display to show full compound notation
|
||||
|
||||
When displaying a subtask via `tm show 104.1`, the header and properties table showed only the subtask's local ID (e.g., "1") instead of the full compound ID (e.g., "104.1"). The CLI now preserves and displays the original requested task ID throughout the display chain, ensuring subtasks are clearly identified with their parent context. Also improved TypeScript typing by using discriminated unions for Task/Subtask returns from `tasks.get()`, eliminating unsafe type coercions.
|
||||
@ -1,30 +0,0 @@
|
||||
---
|
||||
"task-master-ai": minor
|
||||
---
|
||||
|
||||
Add native support for Z.ai (GLM models), giving you access to high-performance Chinese models including glm-4.6 with massive 200K+ token context windows at competitive pricing
|
||||
|
||||
**How to use:**
|
||||
|
||||
1. Get your Z.ai API key from <https://z.ai/manage-apikey/apikey-list>
|
||||
2. Set your API key in .env, mcp.json or in env exports:
|
||||
|
||||
```bash
|
||||
ZAI_API_KEY="your-key-here"
|
||||
```
|
||||
|
||||
3. Configure Task Master to use GLM models:
|
||||
|
||||
```bash
|
||||
task-master models --set-main glm-4.6
|
||||
# Or for an interactive view
|
||||
task-master models --setup
|
||||
```
|
||||
|
||||
**Available models:**
|
||||
|
||||
- `glm-4.6` - Latest model with 200K+ context, excellent for complex projects
|
||||
- `glm-4.5` - Previous generation, still highly capable
|
||||
- Additional GLM variants for different use cases: `glm-4.5-air`, `glm-4.5v`
|
||||
|
||||
GLM models offer strong performance on software engineering tasks, with particularly good results on code generation and technical reasoning. The large context window makes them ideal for analyzing entire codebases or working with extensive documentation.
|
||||
@ -1,7 +0,0 @@
|
||||
---
|
||||
"task-master-ai": patch
|
||||
---
|
||||
|
||||
Fixed MCP server sometimes crashing when getting into the commit step of autopilot
|
||||
|
||||
- autopilot now persists state consistently through the whole flow
|
||||
@ -1,7 +0,0 @@
|
||||
---
|
||||
"task-master-ai": patch
|
||||
---
|
||||
|
||||
Improve gemini cli integration
|
||||
|
||||
When initializing Task Master with the `gemini` profile, you now get properly configured context files tailored specifically for Gemini CLI, including MCP configuration and Gemini-specific features like file references, session management, and headless mode.
|
||||
@ -1,25 +0,0 @@
|
||||
{
|
||||
"mode": "exit",
|
||||
"tag": "rc",
|
||||
"initialVersions": {
|
||||
"task-master-ai": "0.30.2",
|
||||
"@tm/cli": "",
|
||||
"docs": "0.0.8",
|
||||
"extension": "0.25.6",
|
||||
"@tm/mcp": "",
|
||||
"@tm/ai-sdk-provider-grok-cli": "",
|
||||
"@tm/build-config": "",
|
||||
"@tm/claude-code-plugin": "0.0.3",
|
||||
"@tm/bridge": "",
|
||||
"@tm/core": ""
|
||||
},
|
||||
"changesets": [
|
||||
"dirty-walls-ask",
|
||||
"fix-metadata-nullable",
|
||||
"forty-squids-sell",
|
||||
"mighty-pens-ring",
|
||||
"nice-laws-help",
|
||||
"plain-animals-train",
|
||||
"tricky-bats-ring"
|
||||
]
|
||||
}
|
||||
@ -1,23 +0,0 @@
|
||||
---
|
||||
"task-master-ai": minor
|
||||
---
|
||||
|
||||
Add LM Studio integration, enabling you to run Task Master completely offline with local models at zero API cost.
|
||||
|
||||
**How to use:**
|
||||
|
||||
1. Download and install [LM Studio](https://lmstudio.ai/)
|
||||
2. Launch LM Studio and download a model (e.g., Llama 3.2, Mistral, Qwen)
|
||||
3. Optional: Add api key to mcp.json or .env (LMSTUDIO_API_KEY)
|
||||
4. Go to the "Local Server" tab and click "Start Server"
|
||||
5. Configure Task Master:
|
||||
|
||||
```bash
|
||||
task-master models --set-main <model-name> --lmstudio
|
||||
```
|
||||
|
||||
Example:
|
||||
|
||||
```bash
|
||||
task-master models --set-main llama-3.2-3b --lmstudio
|
||||
```
|
||||
90
CHANGELOG.md
90
CHANGELOG.md
@ -1,5 +1,95 @@
|
||||
# task-master-ai
|
||||
|
||||
## 0.31.0
|
||||
|
||||
### Minor Changes
|
||||
|
||||
- [#1360](https://github.com/eyaltoledano/claude-task-master/pull/1360) [`819d5e1`](https://github.com/eyaltoledano/claude-task-master/commit/819d5e1bc5fb81be4b25f1823988a8e20abe8440) Thanks [@Crunchyman-ralph](https://github.com/Crunchyman-ralph)! - Add support for custom OpenAI-compatible providers, allowing you to connect Task Master to any service that implements the OpenAI API specification
|
||||
|
||||
**How to use:**
|
||||
|
||||
Configure your custom provider with the `models` command:
|
||||
|
||||
```bash
|
||||
task-master models --set-main <your-model-id> --openai-compatible --baseURL <your-api-endpoint>
|
||||
```
|
||||
|
||||
Example:
|
||||
|
||||
```bash
|
||||
task-master models --set-main llama-3-70b --openai-compatible --baseURL http://localhost:8000/v1
|
||||
# Or for an interactive view
|
||||
task-master models --setup
|
||||
```
|
||||
|
||||
Set your API key (if required by your provider) in mcp.json, your .env file or in your env exports:
|
||||
|
||||
```bash
|
||||
OPENAI_COMPATIBLE_API_KEY="your-key-here"
|
||||
```
|
||||
|
||||
This gives you the flexibility to use virtually any LLM service with Task Master, whether it's self-hosted, a specialized provider, or a custom inference server.
|
||||
|
||||
- [#1360](https://github.com/eyaltoledano/claude-task-master/pull/1360) [`819d5e1`](https://github.com/eyaltoledano/claude-task-master/commit/819d5e1bc5fb81be4b25f1823988a8e20abe8440) Thanks [@Crunchyman-ralph](https://github.com/Crunchyman-ralph)! - Add native support for Z.ai (GLM models), giving you access to high-performance Chinese models including glm-4.6 with massive 200K+ token context windows at competitive pricing
|
||||
|
||||
**How to use:**
|
||||
1. Get your Z.ai API key from <https://z.ai/manage-apikey/apikey-list>
|
||||
2. Set your API key in .env, mcp.json or in env exports:
|
||||
|
||||
```bash
|
||||
ZAI_API_KEY="your-key-here"
|
||||
```
|
||||
|
||||
3. Configure Task Master to use GLM models:
|
||||
|
||||
```bash
|
||||
task-master models --set-main glm-4.6
|
||||
# Or for an interactive view
|
||||
task-master models --setup
|
||||
```
|
||||
|
||||
**Available models:**
|
||||
- `glm-4.6` - Latest model with 200K+ context, excellent for complex projects
|
||||
- `glm-4.5` - Previous generation, still highly capable
|
||||
- Additional GLM variants for different use cases: `glm-4.5-air`, `glm-4.5v`
|
||||
|
||||
GLM models offer strong performance on software engineering tasks, with particularly good results on code generation and technical reasoning. The large context window makes them ideal for analyzing entire codebases or working with extensive documentation.
|
||||
|
||||
- [#1360](https://github.com/eyaltoledano/claude-task-master/pull/1360) [`819d5e1`](https://github.com/eyaltoledano/claude-task-master/commit/819d5e1bc5fb81be4b25f1823988a8e20abe8440) Thanks [@Crunchyman-ralph](https://github.com/Crunchyman-ralph)! - Add LM Studio integration, enabling you to run Task Master completely offline with local models at zero API cost.
|
||||
|
||||
**How to use:**
|
||||
1. Download and install [LM Studio](https://lmstudio.ai/)
|
||||
2. Launch LM Studio and download a model (e.g., Llama 3.2, Mistral, Qwen)
|
||||
3. Optional: Add api key to mcp.json or .env (LMSTUDIO_API_KEY)
|
||||
4. Go to the "Local Server" tab and click "Start Server"
|
||||
5. Configure Task Master:
|
||||
|
||||
```bash
|
||||
task-master models --set-main <model-name> --lmstudio
|
||||
```
|
||||
|
||||
Example:
|
||||
|
||||
```bash
|
||||
task-master models --set-main llama-3.2-3b --lmstudio
|
||||
```
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- [#1362](https://github.com/eyaltoledano/claude-task-master/pull/1362) [`3e70edf`](https://github.com/eyaltoledano/claude-task-master/commit/3e70edfa3a1f47bd8a6d2d2a30c20c72f5758b9b) Thanks [@Crunchyman-ralph](https://github.com/Crunchyman-ralph)! - Improve parse PRD schema for better llm model compatiblity
|
||||
- Fixes #1353
|
||||
|
||||
- [#1358](https://github.com/eyaltoledano/claude-task-master/pull/1358) [`0c639bd`](https://github.com/eyaltoledano/claude-task-master/commit/0c639bd1db9d2d9b4c2c22ac60b0d875ba75f80e) Thanks [@Crunchyman-ralph](https://github.com/Crunchyman-ralph)! - Fix subtask ID display to show full compound notation
|
||||
|
||||
When displaying a subtask via `tm show 104.1`, the header and properties table showed only the subtask's local ID (e.g., "1") instead of the full compound ID (e.g., "104.1"). The CLI now preserves and displays the original requested task ID throughout the display chain, ensuring subtasks are clearly identified with their parent context. Also improved TypeScript typing by using discriminated unions for Task/Subtask returns from `tasks.get()`, eliminating unsafe type coercions.
|
||||
|
||||
- [#1339](https://github.com/eyaltoledano/claude-task-master/pull/1339) [`3b09b5d`](https://github.com/eyaltoledano/claude-task-master/commit/3b09b5da2a929f260d275f056d35bb6ded54ca6d) Thanks [@Crunchyman-ralph](https://github.com/Crunchyman-ralph)! - Fixed MCP server sometimes crashing when getting into the commit step of autopilot
|
||||
- autopilot now persists state consistently through the whole flow
|
||||
|
||||
- [#1326](https://github.com/eyaltoledano/claude-task-master/pull/1326) [`9d5812b`](https://github.com/eyaltoledano/claude-task-master/commit/9d5812ba6725cfadebb8db8f4aa732cf3cdb3a36) Thanks [@SharifMrCreed](https://github.com/SharifMrCreed)! - Improve gemini cli integration
|
||||
|
||||
When initializing Task Master with the `gemini` profile, you now get properly configured context files tailored specifically for Gemini CLI, including MCP configuration and Gemini-specific features like file references, session management, and headless mode.
|
||||
|
||||
## 0.31.0-rc.0
|
||||
|
||||
### Minor Changes
|
||||
|
||||
@ -11,6 +11,13 @@
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies []:
|
||||
- @tm/core@null
|
||||
|
||||
## null
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- [#1305](https://github.com/eyaltoledano/claude-task-master/pull/1305) [`a98d96e`](https://github.com/eyaltoledano/claude-task-master/commit/a98d96ef0414833b948672f86da4acc11f700ebb) Thanks [@bjcoombs](https://github.com/bjcoombs)! - Fix warning message box width to match dashboard box width for consistent UI alignment
|
||||
|
||||
- Updated dependencies []:
|
||||
|
||||
@ -1,5 +1,7 @@
|
||||
# docs
|
||||
|
||||
## 0.0.9
|
||||
|
||||
## 0.0.8
|
||||
|
||||
## 0.0.7
|
||||
|
||||
@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "docs",
|
||||
"version": "0.0.8",
|
||||
"version": "0.0.9",
|
||||
"private": true,
|
||||
"description": "Task Master documentation powered by Mintlify",
|
||||
"scripts": {
|
||||
|
||||
@ -275,7 +275,7 @@
|
||||
"tailwindcss": "4.1.11",
|
||||
"typescript": "^5.9.2",
|
||||
"@tm/core": "*",
|
||||
"task-master-ai": "0.31.0-rc.0"
|
||||
"task-master-ai": "*"
|
||||
},
|
||||
"overrides": {
|
||||
"glob@<8": "^10.4.5",
|
||||
|
||||
@ -13,3 +13,10 @@
|
||||
|
||||
- Updated dependencies []:
|
||||
- @tm/core@null
|
||||
|
||||
## null
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies []:
|
||||
- @tm/core@null
|
||||
|
||||
8
package-lock.json
generated
8
package-lock.json
generated
@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "task-master-ai",
|
||||
"version": "0.30.2",
|
||||
"version": "0.31.0",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "task-master-ai",
|
||||
"version": "0.30.2",
|
||||
"version": "0.31.0",
|
||||
"license": "MIT WITH Commons-Clause",
|
||||
"workspaces": [
|
||||
"apps/*",
|
||||
@ -217,7 +217,7 @@
|
||||
}
|
||||
},
|
||||
"apps/docs": {
|
||||
"version": "0.0.8",
|
||||
"version": "0.0.9",
|
||||
"devDependencies": {
|
||||
"mintlify": "^4.2.111"
|
||||
}
|
||||
@ -28264,7 +28264,7 @@
|
||||
},
|
||||
"packages/claude-code-plugin": {
|
||||
"name": "@tm/claude-code-plugin",
|
||||
"version": "0.0.3",
|
||||
"version": "0.0.4",
|
||||
"license": "MIT WITH Commons-Clause"
|
||||
},
|
||||
"packages/tm-bridge": {
|
||||
|
||||
@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "task-master-ai",
|
||||
"version": "0.31.0-rc.0",
|
||||
"version": "0.31.0",
|
||||
"description": "A task management system for ambitious AI-driven development that doesn't overwhelm and confuse Cursor.",
|
||||
"main": "index.js",
|
||||
"type": "module",
|
||||
|
||||
@ -7,3 +7,5 @@
|
||||
## null
|
||||
|
||||
## null
|
||||
|
||||
## null
|
||||
|
||||
@ -10,4 +10,6 @@
|
||||
|
||||
## null
|
||||
|
||||
## null
|
||||
|
||||
## 1.0.1
|
||||
|
||||
@ -1,5 +1,7 @@
|
||||
# @tm/claude-code-plugin
|
||||
|
||||
## 0.0.4
|
||||
|
||||
## 0.0.3
|
||||
|
||||
## 0.0.3
|
||||
|
||||
@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@tm/claude-code-plugin",
|
||||
"version": "0.0.3",
|
||||
"version": "0.0.4",
|
||||
"description": "Task Master AI plugin for Claude Code - AI-powered task management with commands, agents, and MCP integration",
|
||||
"type": "module",
|
||||
"private": true,
|
||||
|
||||
@ -6,3 +6,10 @@
|
||||
|
||||
- Updated dependencies []:
|
||||
- @tm/core@null
|
||||
|
||||
## null
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies []:
|
||||
- @tm/core@null
|
||||
|
||||
@ -10,6 +10,8 @@
|
||||
|
||||
## null
|
||||
|
||||
## null
|
||||
|
||||
## 0.26.1
|
||||
|
||||
All notable changes to the @task-master/tm-core package will be documented in this file.
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user