claudish 7.0.3 → 7.1.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,115 +0,0 @@
1
- ; Query from: https://raw.githubusercontent.com/nvim-treesitter/nvim-treesitter/99ddf573531c4dbe53f743ecbc1595af5eb1d32f/queries/markdown_inline/highlights.scm
2
- ; From MDeiml/tree-sitter-markdown
3
- (code_span) @markup.raw @nospell
4
-
5
- (emphasis) @markup.italic
6
-
7
- (strong_emphasis) @markup.strong
8
-
9
- (strikethrough) @markup.strikethrough
10
-
11
- (shortcut_link
12
- (link_text) @nospell)
13
-
14
- [
15
- (backslash_escape)
16
- (hard_line_break)
17
- ] @string.escape
18
-
19
- ; Conceal codeblock and text style markers
20
- ([
21
- (code_span_delimiter)
22
- (emphasis_delimiter)
23
- ] @conceal
24
- (#set! conceal ""))
25
-
26
- ; Inline links - style all parts
27
- (inline_link
28
- ["[" "(" ")"] @markup.link)
29
-
30
- (inline_link
31
- "]" @markup.link.bracket.close)
32
-
33
- ; Conceal opening bracket
34
- ((inline_link
35
- "[" @conceal)
36
- (#set! conceal ""))
37
-
38
- ; Conceal closing bracket with space replacement
39
- ((inline_link
40
- "]" @conceal)
41
- (#set! conceal " "))
42
-
43
- ; Conceal image links
44
- (image
45
- [
46
- "!"
47
- "["
48
- "]"
49
- "("
50
- (link_destination)
51
- ")"
52
- ] @markup.link
53
- (#set! conceal ""))
54
-
55
- ; Conceal full reference links
56
- (full_reference_link
57
- [
58
- "["
59
- "]"
60
- (link_label)
61
- ] @markup.link
62
- (#set! conceal ""))
63
-
64
- ; Conceal collapsed reference links
65
- (collapsed_reference_link
66
- [
67
- "["
68
- "]"
69
- ] @markup.link
70
- (#set! conceal ""))
71
-
72
- ; Conceal shortcut links
73
- (shortcut_link
74
- [
75
- "["
76
- "]"
77
- ] @markup.link
78
- (#set! conceal ""))
79
-
80
- [
81
- (link_destination)
82
- (uri_autolink)
83
- ] @markup.link.url @nospell
84
-
85
- [
86
- (link_label)
87
- (link_text)
88
- (link_title)
89
- (image_description)
90
- ] @markup.link.label
91
-
92
- ; Replace common HTML entities.
93
- ((entity_reference) @character.special
94
- (#eq? @character.special " ")
95
- (#set! conceal ""))
96
-
97
- ((entity_reference) @character.special
98
- (#eq? @character.special "<")
99
- (#set! conceal "<"))
100
-
101
- ((entity_reference) @character.special
102
- (#eq? @character.special "&gt;")
103
- (#set! conceal ">"))
104
-
105
- ((entity_reference) @character.special
106
- (#eq? @character.special "&amp;")
107
- (#set! conceal "&"))
108
-
109
- ((entity_reference) @character.special
110
- (#eq? @character.special "&quot;")
111
- (#set! conceal "\""))
112
-
113
- ((entity_reference) @character.special
114
- (#any-of? @character.special "&ensp;" "&emsp;")
115
- (#set! conceal " "))
@@ -1,27 +0,0 @@
1
- ; Query from: https://raw.githubusercontent.com/nvim-treesitter/nvim-treesitter/refs/heads/master/queries/markdown/injections.scm
2
- (fenced_code_block
3
- (info_string
4
- (language) @_lang)
5
- (code_fence_content) @injection.content
6
- (#set-lang-from-info-string! @_lang))
7
-
8
- ((html_block) @injection.content
9
- (#set! injection.language "html")
10
- (#set! injection.combined)
11
- (#set! injection.include-children))
12
-
13
- ((minus_metadata) @injection.content
14
- (#set! injection.language "yaml")
15
- (#offset! @injection.content 1 0 -1 0)
16
- (#set! injection.include-children))
17
-
18
- ((plus_metadata) @injection.content
19
- (#set! injection.language "toml")
20
- (#offset! @injection.content 1 0 -1 0)
21
- (#set! injection.include-children))
22
-
23
- ([
24
- (inline)
25
- (pipe_table_cell)
26
- ] @injection.content
27
- (#set! injection.language "markdown_inline"))
Binary file
@@ -1,139 +0,0 @@
1
- {
2
- "version": "1.2.0",
3
- "lastUpdated": "2026-04-21",
4
- "source": "https://openrouter.ai/models?categories=programming&fmt=cards&order=top-weekly",
5
- "models": [
6
- {
7
- "id": "minimax-m2.5",
8
- "openrouterId": "minimax/minimax-m2.5",
9
- "name": "MiniMax: MiniMax M2.5",
10
- "description": "MiniMax-M2.5 is a SOTA large language model designed for real-world productivity. Trained in a diverse range of complex real-world digital working environments, M2.5 builds upon the coding expertise of M2.1...",
11
- "provider": "Minimax",
12
- "category": "programming",
13
- "priority": 1,
14
- "pricing": {
15
- "input": "$0.15/1M",
16
- "output": "$1.20/1M",
17
- "average": "$0.68/1M"
18
- },
19
- "context": "196K",
20
- "maxOutputTokens": 65536,
21
- "modality": "text->text",
22
- "supportsTools": true,
23
- "supportsReasoning": true,
24
- "supportsVision": false,
25
- "isModerated": false,
26
- "recommended": true
27
- },
28
- {
29
- "id": "kimi-k2.5",
30
- "openrouterId": "moonshotai/kimi-k2.5",
31
- "name": "MoonshotAI: Kimi K2.5",
32
- "description": "Kimi K2.5 is Moonshot AI's native multimodal model, delivering state-of-the-art visual coding capability and a self-directed agent swarm paradigm. Built on Kimi K2 with continued pretraining over approximately 15T mixed...",
33
- "provider": "Moonshotai",
34
- "category": "vision",
35
- "priority": 2,
36
- "pricing": {
37
- "input": "$0.44/1M",
38
- "output": "$2.00/1M",
39
- "average": "$1.22/1M"
40
- },
41
- "context": "262K",
42
- "maxOutputTokens": 65535,
43
- "modality": "text+image->text",
44
- "supportsTools": true,
45
- "supportsReasoning": true,
46
- "supportsVision": true,
47
- "isModerated": false,
48
- "recommended": true
49
- },
50
- {
51
- "id": "glm-5",
52
- "openrouterId": "z-ai/glm-5",
53
- "name": "Z.ai: GLM 5",
54
- "description": "GLM-5 is Z.ai’s flagship open-source foundation model engineered for complex systems design and long-horizon agent workflows. Built for expert developers, it delivers production-grade performance on large-scale programming tasks, rivaling leading...",
55
- "provider": "Z-ai",
56
- "category": "programming",
57
- "priority": 3,
58
- "pricing": {
59
- "input": "$0.72/1M",
60
- "output": "$2.30/1M",
61
- "average": "$1.51/1M"
62
- },
63
- "context": "80K",
64
- "maxOutputTokens": 131072,
65
- "modality": "text->text",
66
- "supportsTools": true,
67
- "supportsReasoning": true,
68
- "supportsVision": false,
69
- "isModerated": false,
70
- "recommended": true
71
- },
72
- {
73
- "id": "gemini-3.1-pro-preview",
74
- "openrouterId": "google/gemini-3.1-pro-preview",
75
- "name": "Google: Gemini 3.1 Pro Preview",
76
- "description": "Gemini 3.1 Pro Preview is Google’s frontier reasoning model, delivering enhanced software engineering performance, improved agentic reliability, and more efficient token usage across complex workflows. Building on the multimodal foundation...",
77
- "provider": "Google",
78
- "category": "vision",
79
- "priority": 4,
80
- "pricing": {
81
- "input": "$2.00/1M",
82
- "output": "$12.00/1M",
83
- "average": "$7.00/1M"
84
- },
85
- "context": "1048K",
86
- "maxOutputTokens": 65536,
87
- "modality": "text+image+file+audio+video->text",
88
- "supportsTools": true,
89
- "supportsReasoning": true,
90
- "supportsVision": true,
91
- "isModerated": false,
92
- "recommended": true
93
- },
94
- {
95
- "id": "gpt-5.2",
96
- "openrouterId": "openai/gpt-5.2",
97
- "name": "OpenAI: GPT-5.2",
98
- "description": "GPT-5.2 is the latest frontier-grade model in the GPT-5 series, offering stronger agentic and long context perfomance compared to GPT-5.1. It uses adaptive reasoning to allocate computation dynamically, responding quickly...",
99
- "provider": "Openai",
100
- "category": "reasoning",
101
- "priority": 5,
102
- "pricing": {
103
- "input": "$1.75/1M",
104
- "output": "$14.00/1M",
105
- "average": "$7.88/1M"
106
- },
107
- "context": "400K",
108
- "maxOutputTokens": 128000,
109
- "modality": "text+image+file->text",
110
- "supportsTools": true,
111
- "supportsReasoning": true,
112
- "supportsVision": true,
113
- "isModerated": false,
114
- "recommended": true
115
- },
116
- {
117
- "id": "qwen3.5-plus-02-15",
118
- "openrouterId": "qwen/qwen3.5-plus-02-15",
119
- "name": "Qwen: Qwen3.5 Plus 2026-02-15",
120
- "description": "The Qwen3.5 native vision-language series Plus models are built on a hybrid architecture that integrates linear attention mechanisms with sparse mixture-of-experts models, achieving higher inference efficiency. In a variety of...",
121
- "provider": "Qwen",
122
- "category": "vision",
123
- "priority": 6,
124
- "pricing": {
125
- "input": "$0.26/1M",
126
- "output": "$1.56/1M",
127
- "average": "$0.91/1M"
128
- },
129
- "context": "1000K",
130
- "maxOutputTokens": 65536,
131
- "modality": "text+image+video->text",
132
- "supportsTools": true,
133
- "supportsReasoning": true,
134
- "supportsVision": true,
135
- "isModerated": false,
136
- "recommended": true
137
- }
138
- ]
139
- }