@johnowennixon/diffdash 1.14.0 → 1.15.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -32,9 +32,10 @@ npm install -g @johnowennixon/diffdash
32
32
 
33
33
  ## LLM Models
34
34
 
35
- Currently, for this application, the best LLM model is **gpt-5-mini-minimal** (GPT-5 Mini with reasoning disabled) from OpenAI.
35
+ Currently, for this application, the best LLM model is **gpt-4.1-mini** from OpenAI.
36
36
  It is set as the default model.
37
37
  I can only presume they have done a ton of training on diffs.
38
+ I have tested later GPT models but they don't seem to be as good.
38
39
 
39
40
  ## API Keys
40
41
 
package/dist/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@johnowennixon/diffdash",
3
- "version": "1.14.0",
3
+ "version": "1.15.0",
4
4
  "description": "A command-line tool to generate Git commit messages using AI",
5
5
  "license": "0BSD",
6
6
  "author": "John Owen Nixon",
@@ -1,15 +1,15 @@
1
1
  import { env_get_substitute } from "./lib_env.js";
2
2
  import { llm_model_get_choices, llm_model_get_details } from "./lib_llm_model.js";
3
- const model_name_default = "gpt-5-mini-minimal";
3
+ const model_name_default = "gpt-4.1-mini";
4
4
  const model_name_options = [
5
5
  "claude-haiku-4.5", // fallback
6
6
  "deepseek-chat",
7
7
  "gemini-2.5-flash",
8
8
  "gemini-3-flash-preview-low",
9
- "gpt-4.1-mini", // fallback
9
+ "gpt-4.1-mini", // the best
10
10
  "gpt-4.1-nano",
11
11
  "gpt-5-mini",
12
- "gpt-5-mini-minimal", // the best
12
+ "gpt-5-mini-minimal", // fallback
13
13
  "gpt-5-nano",
14
14
  "gpt-5-nano-minimal",
15
15
  "grok-code-fast-1",
@@ -660,39 +660,13 @@ export const LLM_MODEL_DETAILS = [
660
660
  provider_options: provider_options_openrouter({ only: "mistral" }),
661
661
  },
662
662
  {
663
- llm_model_name: "qwen3-235b-a22b-2507-instruct@cerebras",
664
- llm_model_code: "qwen/qwen3-235b-a22b-2507",
665
- llm_api_code: "openrouter",
666
- context_window: 131_072,
667
- max_output_tokens: 131_072,
668
- cents_input: 60,
669
- cents_output: 120,
670
- default_reasoning: false,
671
- has_structured_json: true,
672
- recommended_temperature: undefined,
673
- provider_options: provider_options_openrouter({ only: "cerebras" }),
674
- },
675
- {
676
- llm_model_name: "qwen3-coder@alibaba",
677
- llm_model_code: "qwen/qwen3-coder",
663
+ llm_model_name: "qwen3.5-397b-a17b@alibaba",
664
+ llm_model_code: "qwen/qwen3.5-397b-a17b",
678
665
  llm_api_code: "openrouter",
679
666
  context_window: 262_144,
680
667
  max_output_tokens: 65_536,
681
- cents_input: 150, // for input tokens <= 128K
682
- cents_output: 750, // for input tokens <= 128K
683
- default_reasoning: false,
684
- has_structured_json: true,
685
- recommended_temperature: undefined,
686
- provider_options: provider_options_openrouter({ only: "alibaba/opensource" }),
687
- },
688
- {
689
- llm_model_name: "qwen-plus@alibaba",
690
- llm_model_code: "qwen/qwen-plus-2025-07-28",
691
- llm_api_code: "openrouter",
692
- context_window: 1_000_000,
693
- max_output_tokens: 32_768,
694
- cents_input: 40, // for input tokens <= 256K
695
- cents_output: 120, // for input tokens <= 256K
668
+ cents_input: 60,
669
+ cents_output: 360,
696
670
  default_reasoning: false,
697
671
  has_structured_json: true,
698
672
  recommended_temperature: undefined,
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@johnowennixon/diffdash",
3
- "version": "1.14.0",
3
+ "version": "1.15.0",
4
4
  "description": "A command-line tool to generate Git commit messages using AI",
5
5
  "license": "0BSD",
6
6
  "author": "John Owen Nixon",