mirror of
https://github.com/danielmiessler/Fabric.git
synced 2026-02-19 10:14:21 -05:00
chore: update model references to latest versions across docs and code
- Update default summarize model to `claude-sonnet-4-5` - Replace `gpt-4o` references with `gpt-5.2` throughout docs - Replace `gpt-4o-mini` references with `gpt-5-mini` - Add MiniMax-M2.5 and M2.5-lightning to static models list - Update image generation warning to suggest `gpt-5.2` - Update OpenAI chat example script to use `gpt-5-mini` - Update REST API docs and examples with current model names
This commit is contained in:
@@ -246,7 +246,7 @@ Set the model via environment variable:
|
||||
```bash
|
||||
export FABRIC_CHANGELOG_SUMMARIZE_MODEL=claude-opus-4
|
||||
# or
|
||||
export FABRIC_CHANGELOG_SUMMARIZE_MODEL=gpt-4
|
||||
export FABRIC_CHANGELOG_SUMMARIZE_MODEL=gpt-5.2
|
||||
```
|
||||
|
||||
AI summaries are cached and only regenerated when:
|
||||
|
||||
@@ -7,7 +7,7 @@ import (
|
||||
"strings"
|
||||
)
|
||||
|
||||
const DefaultSummarizeModel = "claude-sonnet-4-20250514"
|
||||
const DefaultSummarizeModel = "claude-sonnet-4-5"
|
||||
const MinContentLength = 256 // Minimum content length to consider for summarization
|
||||
|
||||
const prompt = `# ROLE
|
||||
|
||||
@@ -16,6 +16,6 @@ notification: true
|
||||
# notificationCommand: '/path/to/custom-notification-script.sh "$1" "$2"'
|
||||
|
||||
# Other common settings
|
||||
model: "gpt-4o"
|
||||
model: "gpt-5.2"
|
||||
temperature: 0.7
|
||||
stream: true
|
||||
|
||||
@@ -81,7 +81,7 @@ Stream AI responses using Server-Sent Events (SSE).
|
||||
{
|
||||
"userInput": "Explain quantum computing",
|
||||
"vendor": "openai",
|
||||
"model": "gpt-4o",
|
||||
"model": "gpt-5.2",
|
||||
"patternName": "explain",
|
||||
"contextName": "",
|
||||
"strategyName": "",
|
||||
@@ -103,7 +103,7 @@ Stream AI responses using Server-Sent Events (SSE).
|
||||
| ------- | ---------- | --------- | ------------- |
|
||||
| `userInput` | **Yes** | - | Your message or question |
|
||||
| `vendor` | **Yes** | - | AI provider: `openai`, `anthropic`, `gemini`, `ollama`, etc. |
|
||||
| `model` | **Yes** | - | Model name: `gpt-4o`, `claude-sonnet-4.5`, `gemini-2.0-flash-exp`, etc. |
|
||||
| `model` | **Yes** | - | Model name: `gpt-5.2`, `claude-sonnet-4.5`, `gemini-2.0-flash-exp`, etc. |
|
||||
| `patternName` | No | `""` | Pattern to apply (from `~/.config/fabric/patterns/`) |
|
||||
| `contextName` | No | `""` | Context to prepend (from `~/.config/fabric/contexts/`) |
|
||||
| `strategyName` | No | `""` | Strategy to use (from `~/.config/fabric/strategies/`) |
|
||||
@@ -151,7 +151,7 @@ curl -X POST http://localhost:8080/chat \
|
||||
"prompts": [{
|
||||
"userInput": "What is Fabric?",
|
||||
"vendor": "openai",
|
||||
"model": "gpt-4o",
|
||||
"model": "gpt-5.2",
|
||||
"patternName": "explain"
|
||||
}]
|
||||
}'
|
||||
@@ -232,9 +232,9 @@ List available AI models.
|
||||
|
||||
```json
|
||||
{
|
||||
"models": ["gpt-4o", "gpt-4o-mini", "claude-sonnet-4.5", "gemini-2.0-flash-exp"],
|
||||
"models": ["gpt-5.2", "gpt-5-mini", "claude-sonnet-4.5", "gemini-2.0-flash-exp"],
|
||||
"vendors": {
|
||||
"openai": ["gpt-4o", "gpt-4o-mini"],
|
||||
"openai": ["gpt-5.2", "gpt-5-mini"],
|
||||
"anthropic": ["claude-sonnet-4.5", "claude-opus-4.5"],
|
||||
"gemini": ["gemini-2.0-flash-exp", "gemini-2.0-flash-thinking-exp"]
|
||||
}
|
||||
@@ -359,7 +359,7 @@ curl -X POST http://localhost:8080/chat \
|
||||
\"prompts\": [{
|
||||
\"userInput\": \"$TRANSCRIPT\",
|
||||
\"vendor\": \"openai\",
|
||||
\"model\": \"gpt-4o\",
|
||||
\"model\": \"gpt-5.2\",
|
||||
\"patternName\": \"youtube_summary\"
|
||||
}]
|
||||
}"
|
||||
@@ -374,7 +374,7 @@ curl -s -X POST http://localhost:8080/youtube/transcript \
|
||||
jq -r '.transcript' | \
|
||||
xargs -I {} curl -X POST http://localhost:8080/chat \
|
||||
-H "Content-Type: application/json" \
|
||||
-d "{\"prompts\":[{\"userInput\":\"{}\",\"vendor\":\"openai\",\"model\":\"gpt-4o\",\"patternName\":\"youtube_summary\"}]}"
|
||||
-d "{\"prompts\":[{\"userInput\":\"{}\",\"vendor\":\"openai\",\"model\":\"gpt-5.2\",\"patternName\":\"youtube_summary\"}]}"
|
||||
```
|
||||
|
||||
#### Alternative: Using a script
|
||||
@@ -398,7 +398,7 @@ curl -X POST "$API_BASE/chat" \
|
||||
\"prompts\": [{
|
||||
\"userInput\": $(echo "$TRANSCRIPT" | jq -Rs .),
|
||||
\"vendor\": \"openai\",
|
||||
\"model\": \"gpt-4o\",
|
||||
\"model\": \"gpt-5.2\",
|
||||
\"patternName\": \"youtube_summary\"
|
||||
}]
|
||||
}"
|
||||
|
||||
@@ -75,7 +75,7 @@ func (o *Client) SetResponsesAPIEnabled(enabled bool) {
|
||||
// checkImageGenerationCompatibility warns if the model doesn't support image generation
|
||||
func checkImageGenerationCompatibility(model string) {
|
||||
if !supportsImageGeneration(model) {
|
||||
fmt.Fprintf(os.Stderr, "Warning: Model '%s' does not support image generation. Supported models: %s. Consider using -m gpt-4o for image generation.\n",
|
||||
fmt.Fprintf(os.Stderr, "Warning: Model '%s' does not support image generation. Supported models: %s. Consider using -m gpt-5.2 for image generation.\n",
|
||||
model, strings.Join(ImageGenerationSupportedModels, ", "))
|
||||
}
|
||||
}
|
||||
|
||||
@@ -521,7 +521,7 @@ func TestCheckImageGenerationCompatibility(t *testing.T) {
|
||||
assert.NotEmpty(t, output, "Expected warning output for unsupported model")
|
||||
assert.Contains(t, output, tt.expectedText, "Warning message should contain model name")
|
||||
assert.Contains(t, output, "Supported models:", "Warning should mention supported models")
|
||||
assert.Contains(t, output, "gpt-4o", "Warning should suggest gpt-4o")
|
||||
assert.Contains(t, output, "gpt-5.2", "Warning should suggest gpt-5.2")
|
||||
} else {
|
||||
assert.Empty(t, output, "No warning expected for supported model")
|
||||
}
|
||||
|
||||
@@ -194,6 +194,8 @@ func (c *Client) getStaticModels(modelsKey string) ([]string, error) {
|
||||
}, nil
|
||||
case "static:minimax":
|
||||
return []string{
|
||||
"MiniMax-M2.5",
|
||||
"MiniMax-M2.5-lightning",
|
||||
"MiniMax-M2",
|
||||
"MiniMax-M2.1",
|
||||
"MiniMax-M2.1-lightning",
|
||||
|
||||
@@ -6,7 +6,7 @@ RESPONSE=$(curl "$OPENAI_API_BASE_URL/chat/completions" \
|
||||
-s -w "\n%{http_code}" \
|
||||
-H "Content-Type: application/json" \
|
||||
-H "Authorization: Bearer $OPENAI_API_KEY" \
|
||||
-d "{\"model\":\"gpt-4o-mini\",\"messages\":[{\"role\":\"user\",\"content\":$INPUT}]}")
|
||||
-d "{\"model\":\"gpt-5-mini\",\"messages\":[{\"role\":\"user\",\"content\":$INPUT}]}")
|
||||
|
||||
HTTP_CODE=$(echo "$RESPONSE" | tail -n1)
|
||||
BODY=$(echo "$RESPONSE" | sed '$d')
|
||||
|
||||
@@ -12,7 +12,7 @@ This example demonstrates how to use pattern variables in REST API calls to the
|
||||
{
|
||||
"userInput": "Hello my name is Kayvan",
|
||||
"patternName": "translate",
|
||||
"model": "gpt-4o",
|
||||
"model": "gpt-5.2",
|
||||
"vendor": "openai",
|
||||
"contextName": "",
|
||||
"strategyName": "",
|
||||
@@ -64,7 +64,7 @@ curl -X POST http://localhost:8080/api/chat \
|
||||
{
|
||||
"userInput": "Hello my name is Kayvan",
|
||||
"patternName": "translate",
|
||||
"model": "gpt-4o",
|
||||
"model": "gpt-5.2",
|
||||
"vendor": "openai",
|
||||
"variables": {
|
||||
"lang_code": "fr"
|
||||
@@ -85,7 +85,7 @@ For patterns that use multiple variables:
|
||||
{
|
||||
"userInput": "Analyze this business model",
|
||||
"patternName": "custom_analysis",
|
||||
"model": "gpt-4o",
|
||||
"model": "gpt-5.2",
|
||||
"variables": {
|
||||
"role": "expert consultant",
|
||||
"experience": "15",
|
||||
|
||||
Reference in New Issue
Block a user