@wcj/github-rank 24.5.7 → 24.5.8
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -1
- package/dist/repos.json +23112 -23112
- package/dist/trending-daily.json +181 -181
- package/dist/trending-monthly.json +165 -165
- package/dist/trending-weekly.json +168 -168
- package/dist/users.china.json +1314 -1279
- package/dist/users.json +3844 -3844
- package/dist/users.org.json +409 -409
- package/package.json +1 -1
- package/web/data/repos.json +23112 -23112
- package/web/data/trending-daily.json +181 -181
- package/web/data/trending-monthly.json +165 -165
- package/web/data/trending-weekly.json +168 -168
- package/web/data/users.china.json +1314 -1279
- package/web/data/users.json +3844 -3844
- package/web/data/users.org.json +409 -409
- package/web/index.html +3943 -3936
- package/web/org.html +327 -327
- package/web/repos.html +2903 -2903
- package/web/trending-monthly.html +157 -157
- package/web/trending-weekly.html +158 -158
- package/web/trending.html +169 -173
- package/web/users.china.html +1342 -1292
|
@@ -1,24 +1,24 @@
|
|
|
1
1
|
[
|
|
2
|
+
{
|
|
3
|
+
"full_name": "nashsu/FreeAskInternet",
|
|
4
|
+
"language": "Python",
|
|
5
|
+
"color": "#3572A5",
|
|
6
|
+
"description": "FreeAskInternet is a completely free, PRIVATE and LOCALLY running search aggregator & answer generate using MULTI LLMs, without GPU needed. The user can ask a question and the system will make a multi engine search and combine the search result to LLM and generate the answer based on search results. It's all FREE to use.",
|
|
7
|
+
"forked": "825",
|
|
8
|
+
"stargazers_count": 7636,
|
|
9
|
+
"todayStar": "6504 stars this month",
|
|
10
|
+
"html_url": "https://github.com/nashsu/FreeAskInternet",
|
|
11
|
+
"rank": 1
|
|
12
|
+
},
|
|
2
13
|
{
|
|
3
14
|
"full_name": "jwasham/coding-interview-university",
|
|
4
15
|
"language": "",
|
|
5
16
|
"color": "",
|
|
6
17
|
"description": "A complete computer science study plan to become a software engineer.",
|
|
7
|
-
"forked": "
|
|
8
|
-
"stargazers_count":
|
|
9
|
-
"todayStar": "
|
|
18
|
+
"forked": "73829",
|
|
19
|
+
"stargazers_count": 291431,
|
|
20
|
+
"todayStar": "10410 stars this month",
|
|
10
21
|
"html_url": "https://github.com/jwasham/coding-interview-university",
|
|
11
|
-
"rank": 1
|
|
12
|
-
},
|
|
13
|
-
{
|
|
14
|
-
"full_name": "langgenius/dify",
|
|
15
|
-
"language": "TypeScript",
|
|
16
|
-
"color": "#3178c6",
|
|
17
|
-
"description": "Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.",
|
|
18
|
-
"forked": "3508",
|
|
19
|
-
"stargazers_count": 27153,
|
|
20
|
-
"todayStar": "8113 stars this month",
|
|
21
|
-
"html_url": "https://github.com/langgenius/dify",
|
|
22
22
|
"rank": 2
|
|
23
23
|
},
|
|
24
24
|
{
|
|
@@ -26,21 +26,21 @@
|
|
|
26
26
|
"language": "Python",
|
|
27
27
|
"color": "#3572A5",
|
|
28
28
|
"description": "Instant voice cloning by MyShell.",
|
|
29
|
-
"forked": "
|
|
30
|
-
"stargazers_count":
|
|
31
|
-
"todayStar": "
|
|
29
|
+
"forked": "2252",
|
|
30
|
+
"stargazers_count": 24333,
|
|
31
|
+
"todayStar": "7522 stars this month",
|
|
32
32
|
"html_url": "https://github.com/myshell-ai/OpenVoice",
|
|
33
33
|
"rank": 3
|
|
34
34
|
},
|
|
35
35
|
{
|
|
36
|
-
"full_name": "
|
|
37
|
-
"language": "
|
|
38
|
-
"color": "#
|
|
39
|
-
"description": "
|
|
40
|
-
"forked": "
|
|
41
|
-
"stargazers_count":
|
|
42
|
-
"todayStar": "
|
|
43
|
-
"html_url": "https://github.com/
|
|
36
|
+
"full_name": "langgenius/dify",
|
|
37
|
+
"language": "TypeScript",
|
|
38
|
+
"color": "#3178c6",
|
|
39
|
+
"description": "Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.",
|
|
40
|
+
"forked": "3545",
|
|
41
|
+
"stargazers_count": 27455,
|
|
42
|
+
"todayStar": "8369 stars this month",
|
|
43
|
+
"html_url": "https://github.com/langgenius/dify",
|
|
44
44
|
"rank": 4
|
|
45
45
|
},
|
|
46
46
|
{
|
|
@@ -48,54 +48,54 @@
|
|
|
48
48
|
"language": "Vue",
|
|
49
49
|
"color": "#41b883",
|
|
50
50
|
"description": "Vue3 + Pinia 仿抖音,Vue 在移动端的最佳实践 . Imitate TikTok ,Vue Best practices on Mobile",
|
|
51
|
-
"forked": "
|
|
52
|
-
"stargazers_count":
|
|
53
|
-
"todayStar": "
|
|
51
|
+
"forked": "1715",
|
|
52
|
+
"stargazers_count": 7067,
|
|
53
|
+
"todayStar": "4628 stars this month",
|
|
54
54
|
"html_url": "https://github.com/zyronon/douyin",
|
|
55
55
|
"rank": 5
|
|
56
56
|
},
|
|
57
57
|
{
|
|
58
|
-
"full_name": "
|
|
58
|
+
"full_name": "infiniflow/ragflow",
|
|
59
59
|
"language": "Python",
|
|
60
60
|
"color": "#3572A5",
|
|
61
|
-
"description": "
|
|
62
|
-
"forked": "
|
|
63
|
-
"stargazers_count":
|
|
64
|
-
"todayStar": "
|
|
65
|
-
"html_url": "https://github.com/
|
|
61
|
+
"description": "RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.",
|
|
62
|
+
"forked": "562",
|
|
63
|
+
"stargazers_count": 6645,
|
|
64
|
+
"todayStar": "3887 stars this month",
|
|
65
|
+
"html_url": "https://github.com/infiniflow/ragflow",
|
|
66
66
|
"rank": 6
|
|
67
67
|
},
|
|
68
68
|
{
|
|
69
|
-
"full_name": "
|
|
70
|
-
"language": "
|
|
71
|
-
"color": "#
|
|
72
|
-
"description": "
|
|
73
|
-
"forked": "
|
|
74
|
-
"stargazers_count":
|
|
75
|
-
"todayStar": "
|
|
76
|
-
"html_url": "https://github.com/
|
|
69
|
+
"full_name": "FoundationVision/VAR",
|
|
70
|
+
"language": "Python",
|
|
71
|
+
"color": "#3572A5",
|
|
72
|
+
"description": "[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of \"Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction\". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!",
|
|
73
|
+
"forked": "272",
|
|
74
|
+
"stargazers_count": 3408,
|
|
75
|
+
"todayStar": "2592 stars this month",
|
|
76
|
+
"html_url": "https://github.com/FoundationVision/VAR",
|
|
77
77
|
"rank": 7
|
|
78
78
|
},
|
|
79
79
|
{
|
|
80
|
-
"full_name": "
|
|
80
|
+
"full_name": "LlamaFamily/Llama-Chinese",
|
|
81
81
|
"language": "Python",
|
|
82
82
|
"color": "#3572A5",
|
|
83
|
-
"description": "
|
|
84
|
-
"forked": "
|
|
85
|
-
"stargazers_count":
|
|
86
|
-
"todayStar": "
|
|
87
|
-
"html_url": "https://github.com/
|
|
83
|
+
"description": "Llama中文社区,Llama3在线体验和微调模型已开放,实时汇总最新Llama3学习资料,已将所有代码更新适配Llama3,构建最好的中文Llama大模型,完全开源可商用",
|
|
84
|
+
"forked": "1084",
|
|
85
|
+
"stargazers_count": 11641,
|
|
86
|
+
"todayStar": "2578 stars this month",
|
|
87
|
+
"html_url": "https://github.com/LlamaFamily/Llama-Chinese",
|
|
88
88
|
"rank": 8
|
|
89
89
|
},
|
|
90
90
|
{
|
|
91
|
-
"full_name": "
|
|
92
|
-
"language": "
|
|
93
|
-
"color": "#
|
|
94
|
-
"description": "
|
|
95
|
-
"forked": "
|
|
96
|
-
"stargazers_count":
|
|
97
|
-
"todayStar": "
|
|
98
|
-
"html_url": "https://github.com/
|
|
91
|
+
"full_name": "phidatahq/phidata",
|
|
92
|
+
"language": "Python",
|
|
93
|
+
"color": "#3572A5",
|
|
94
|
+
"description": "Memory, knowledge and tools for LLMs",
|
|
95
|
+
"forked": "745",
|
|
96
|
+
"stargazers_count": 5518,
|
|
97
|
+
"todayStar": "2586 stars this month",
|
|
98
|
+
"html_url": "https://github.com/phidatahq/phidata",
|
|
99
99
|
"rank": 9
|
|
100
100
|
},
|
|
101
101
|
{
|
|
@@ -103,175 +103,175 @@
|
|
|
103
103
|
"language": "Python",
|
|
104
104
|
"color": "#3572A5",
|
|
105
105
|
"description": "30 days of Python programming challenge is a step-by-step guide to learn the Python programming language in 30 days. This challenge may take more than100 days, follow your own pace. These videos may help too: https://www.youtube.com/channel/UC7PNRuno1rzYPb1xLa4yktw",
|
|
106
|
-
"forked": "
|
|
107
|
-
"stargazers_count":
|
|
108
|
-
"todayStar": "
|
|
106
|
+
"forked": "6514",
|
|
107
|
+
"stargazers_count": 32061,
|
|
108
|
+
"todayStar": "2029 stars this month",
|
|
109
109
|
"html_url": "https://github.com/Asabeneh/30-Days-Of-Python",
|
|
110
110
|
"rank": 10
|
|
111
111
|
},
|
|
112
112
|
{
|
|
113
|
-
"full_name": "
|
|
113
|
+
"full_name": "hiyouga/LLaMA-Factory",
|
|
114
|
+
"language": "Python",
|
|
115
|
+
"color": "#3572A5",
|
|
116
|
+
"description": "Unify Efficient Fine-Tuning of 100+ LLMs",
|
|
117
|
+
"forked": "2559",
|
|
118
|
+
"stargazers_count": 21270,
|
|
119
|
+
"todayStar": "5629 stars this month",
|
|
120
|
+
"html_url": "https://github.com/hiyouga/LLaMA-Factory",
|
|
121
|
+
"rank": 11
|
|
122
|
+
},
|
|
123
|
+
{
|
|
124
|
+
"full_name": "ollama/ollama",
|
|
114
125
|
"language": "Go",
|
|
115
126
|
"color": "#00ADD8",
|
|
116
|
-
"description": "
|
|
117
|
-
"forked": "
|
|
118
|
-
"stargazers_count":
|
|
119
|
-
"todayStar": "
|
|
120
|
-
"html_url": "https://github.com/
|
|
121
|
-
"rank":
|
|
127
|
+
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
|
128
|
+
"forked": "4657",
|
|
129
|
+
"stargazers_count": 64631,
|
|
130
|
+
"todayStar": "11525 stars this month",
|
|
131
|
+
"html_url": "https://github.com/ollama/ollama",
|
|
132
|
+
"rank": 12
|
|
133
|
+
},
|
|
134
|
+
{
|
|
135
|
+
"full_name": "NaiboWang/EasySpider",
|
|
136
|
+
"language": "JavaScript",
|
|
137
|
+
"color": "#f1e05a",
|
|
138
|
+
"description": "A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:ServiceWrapper面向Web应用的智能化服务封装系统。",
|
|
139
|
+
"forked": "2690",
|
|
140
|
+
"stargazers_count": 23096,
|
|
141
|
+
"todayStar": "2720 stars this month",
|
|
142
|
+
"html_url": "https://github.com/NaiboWang/EasySpider",
|
|
143
|
+
"rank": 13
|
|
144
|
+
},
|
|
145
|
+
{
|
|
146
|
+
"full_name": "unslothai/unsloth",
|
|
147
|
+
"language": "Python",
|
|
148
|
+
"color": "#3572A5",
|
|
149
|
+
"description": "Finetune Llama 3, Mistral & Gemma LLMs 2-5x faster with 80% less memory",
|
|
150
|
+
"forked": "553",
|
|
151
|
+
"stargazers_count": 8811,
|
|
152
|
+
"todayStar": "3433 stars this month",
|
|
153
|
+
"html_url": "https://github.com/unslothai/unsloth",
|
|
154
|
+
"rank": 14
|
|
122
155
|
},
|
|
123
156
|
{
|
|
124
157
|
"full_name": "ordinals/ord",
|
|
125
158
|
"language": "Rust",
|
|
126
159
|
"color": "#dea584",
|
|
127
160
|
"description": "👁🗨 Rare and exotic sats",
|
|
128
|
-
"forked": "
|
|
129
|
-
"stargazers_count":
|
|
161
|
+
"forked": "1250",
|
|
162
|
+
"stargazers_count": 3671,
|
|
130
163
|
"todayStar": "487 stars this month",
|
|
131
164
|
"html_url": "https://github.com/ordinals/ord",
|
|
132
|
-
"rank":
|
|
165
|
+
"rank": 15
|
|
133
166
|
},
|
|
134
167
|
{
|
|
135
|
-
"full_name": "
|
|
136
|
-
"language": "
|
|
137
|
-
"color": "#
|
|
138
|
-
"description": "
|
|
139
|
-
"forked": "
|
|
140
|
-
"stargazers_count":
|
|
141
|
-
"todayStar": "
|
|
142
|
-
"html_url": "https://github.com/
|
|
143
|
-
"rank":
|
|
168
|
+
"full_name": "wandb/openui",
|
|
169
|
+
"language": "TypeScript",
|
|
170
|
+
"color": "#3178c6",
|
|
171
|
+
"description": "OpenUI let's you describe UI using your imagination, then see it rendered live.",
|
|
172
|
+
"forked": "866",
|
|
173
|
+
"stargazers_count": 10689,
|
|
174
|
+
"todayStar": "5760 stars this month",
|
|
175
|
+
"html_url": "https://github.com/wandb/openui",
|
|
176
|
+
"rank": 16
|
|
144
177
|
},
|
|
145
178
|
{
|
|
146
179
|
"full_name": "meta-llama/llama-recipes",
|
|
147
180
|
"language": "JupyterNotebook",
|
|
148
181
|
"color": "#DA5B0B",
|
|
149
182
|
"description": "Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama3 for WhatsApp & Messenger.",
|
|
150
|
-
"forked": "
|
|
151
|
-
"stargazers_count":
|
|
152
|
-
"todayStar": "
|
|
183
|
+
"forked": "1338",
|
|
184
|
+
"stargazers_count": 9500,
|
|
185
|
+
"todayStar": "1381 stars this month",
|
|
153
186
|
"html_url": "https://github.com/meta-llama/llama-recipes",
|
|
154
|
-
"rank":
|
|
187
|
+
"rank": 17
|
|
155
188
|
},
|
|
156
189
|
{
|
|
157
190
|
"full_name": "pytorch/torchtune",
|
|
158
191
|
"language": "Python",
|
|
159
192
|
"color": "#3572A5",
|
|
160
193
|
"description": "A Native-PyTorch Library for LLM Fine-tuning",
|
|
161
|
-
"forked": "
|
|
162
|
-
"stargazers_count":
|
|
163
|
-
"todayStar": "
|
|
194
|
+
"forked": "200",
|
|
195
|
+
"stargazers_count": 3077,
|
|
196
|
+
"todayStar": "2344 stars this month",
|
|
164
197
|
"html_url": "https://github.com/pytorch/torchtune",
|
|
165
|
-
"rank":
|
|
198
|
+
"rank": 18
|
|
166
199
|
},
|
|
167
200
|
{
|
|
168
|
-
"full_name": "
|
|
201
|
+
"full_name": "open-webui/open-webui",
|
|
202
|
+
"language": "Svelte",
|
|
203
|
+
"color": "#ff3e00",
|
|
204
|
+
"description": "User-friendly WebUI for LLMs (Formerly Ollama WebUI)",
|
|
205
|
+
"forked": "1939",
|
|
206
|
+
"stargazers_count": 19190,
|
|
207
|
+
"todayStar": "7019 stars this month",
|
|
208
|
+
"html_url": "https://github.com/open-webui/open-webui",
|
|
209
|
+
"rank": 19
|
|
210
|
+
},
|
|
211
|
+
{
|
|
212
|
+
"full_name": "paul-gauthier/aider",
|
|
169
213
|
"language": "Python",
|
|
170
214
|
"color": "#3572A5",
|
|
171
|
-
"description": "
|
|
172
|
-
"forked": "
|
|
173
|
-
"stargazers_count":
|
|
174
|
-
"todayStar": "
|
|
175
|
-
"html_url": "https://github.com/
|
|
176
|
-
"rank":
|
|
215
|
+
"description": "aider is AI pair programming in your terminal",
|
|
216
|
+
"forked": "975",
|
|
217
|
+
"stargazers_count": 9855,
|
|
218
|
+
"todayStar": "2501 stars this month",
|
|
219
|
+
"html_url": "https://github.com/paul-gauthier/aider",
|
|
220
|
+
"rank": 20
|
|
177
221
|
},
|
|
178
222
|
{
|
|
179
223
|
"full_name": "CorentinTh/it-tools",
|
|
180
224
|
"language": "Vue",
|
|
181
225
|
"color": "#41b883",
|
|
182
226
|
"description": "Collection of handy online tools for developers, with great UX.",
|
|
183
|
-
"forked": "
|
|
184
|
-
"stargazers_count":
|
|
185
|
-
"todayStar": "
|
|
227
|
+
"forked": "1464",
|
|
228
|
+
"stargazers_count": 11757,
|
|
229
|
+
"todayStar": "2344 stars this month",
|
|
186
230
|
"html_url": "https://github.com/CorentinTh/it-tools",
|
|
187
|
-
"rank":
|
|
188
|
-
},
|
|
189
|
-
{
|
|
190
|
-
"full_name": "paul-gauthier/aider",
|
|
191
|
-
"language": "Python",
|
|
192
|
-
"color": "#3572A5",
|
|
193
|
-
"description": "aider is AI pair programming in your terminal",
|
|
194
|
-
"forked": "974",
|
|
195
|
-
"stargazers_count": 9819,
|
|
196
|
-
"todayStar": "2507 stars this month",
|
|
197
|
-
"html_url": "https://github.com/paul-gauthier/aider",
|
|
198
|
-
"rank": 18
|
|
231
|
+
"rank": 21
|
|
199
232
|
},
|
|
200
233
|
{
|
|
201
234
|
"full_name": "karpathy/llama2.c",
|
|
202
235
|
"language": "C",
|
|
203
236
|
"color": "#555555",
|
|
204
237
|
"description": "Inference Llama 2 in one file of pure C",
|
|
205
|
-
"forked": "
|
|
206
|
-
"stargazers_count":
|
|
207
|
-
"todayStar": "
|
|
238
|
+
"forked": "1841",
|
|
239
|
+
"stargazers_count": 16113,
|
|
240
|
+
"todayStar": "1332 stars this month",
|
|
208
241
|
"html_url": "https://github.com/karpathy/llama2.c",
|
|
209
|
-
"rank":
|
|
210
|
-
},
|
|
211
|
-
{
|
|
212
|
-
"full_name": "open-webui/open-webui",
|
|
213
|
-
"language": "Svelte",
|
|
214
|
-
"color": "#ff3e00",
|
|
215
|
-
"description": "User-friendly WebUI for LLMs (Formerly Ollama WebUI)",
|
|
216
|
-
"forked": "1915",
|
|
217
|
-
"stargazers_count": 18954,
|
|
218
|
-
"todayStar": "6836 stars this month",
|
|
219
|
-
"html_url": "https://github.com/open-webui/open-webui",
|
|
220
|
-
"rank": 20
|
|
221
|
-
},
|
|
222
|
-
{
|
|
223
|
-
"full_name": "Mozilla-Ocho/llamafile",
|
|
224
|
-
"language": "C++",
|
|
225
|
-
"color": "#f34b7d",
|
|
226
|
-
"description": "Distribute and run LLMs with a single file.",
|
|
227
|
-
"forked": "746",
|
|
228
|
-
"stargazers_count": 15067,
|
|
229
|
-
"todayStar": "2939 stars this month",
|
|
230
|
-
"html_url": "https://github.com/Mozilla-Ocho/llamafile",
|
|
231
|
-
"rank": 21
|
|
242
|
+
"rank": 22
|
|
232
243
|
},
|
|
233
244
|
{
|
|
234
|
-
"full_name": "
|
|
245
|
+
"full_name": "LLM-Red-Team/kimi-free-api",
|
|
235
246
|
"language": "TypeScript",
|
|
236
247
|
"color": "#3178c6",
|
|
237
|
-
"description": "
|
|
238
|
-
"forked": "
|
|
239
|
-
"stargazers_count":
|
|
240
|
-
"todayStar": "
|
|
241
|
-
"html_url": "https://github.com/
|
|
242
|
-
"rank": 22
|
|
243
|
-
},
|
|
244
|
-
{
|
|
245
|
-
"full_name": "ggerganov/whisper.cpp",
|
|
246
|
-
"language": "C",
|
|
247
|
-
"color": "#555555",
|
|
248
|
-
"description": "Port of OpenAI's Whisper model in C/C++",
|
|
249
|
-
"forked": "3118",
|
|
250
|
-
"stargazers_count": 31508,
|
|
251
|
-
"todayStar": "2242 stars this month",
|
|
252
|
-
"html_url": "https://github.com/ggerganov/whisper.cpp",
|
|
248
|
+
"description": "🚀 KIMI AI 长文本大模型逆向API白嫖测试【特长:长文本解读整理】,支持高速流式输出、智能体对话、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。",
|
|
249
|
+
"forked": "379",
|
|
250
|
+
"stargazers_count": 2654,
|
|
251
|
+
"todayStar": "1429 stars this month",
|
|
252
|
+
"html_url": "https://github.com/LLM-Red-Team/kimi-free-api",
|
|
253
253
|
"rank": 23
|
|
254
254
|
},
|
|
255
255
|
{
|
|
256
|
-
"full_name": "
|
|
256
|
+
"full_name": "quilljs/quill",
|
|
257
257
|
"language": "TypeScript",
|
|
258
258
|
"color": "#3178c6",
|
|
259
|
-
"description": "
|
|
260
|
-
"forked": "
|
|
261
|
-
"stargazers_count":
|
|
262
|
-
"todayStar": "
|
|
263
|
-
"html_url": "https://github.com/
|
|
259
|
+
"description": "Quill is a modern WYSIWYG editor built for compatibility and extensibility.",
|
|
260
|
+
"forked": "3256",
|
|
261
|
+
"stargazers_count": 41432,
|
|
262
|
+
"todayStar": "2666 stars this month",
|
|
263
|
+
"html_url": "https://github.com/quilljs/quill",
|
|
264
264
|
"rank": 24
|
|
265
265
|
},
|
|
266
266
|
{
|
|
267
|
-
"full_name": "
|
|
268
|
-
"language": "
|
|
269
|
-
"color": "#
|
|
270
|
-
"description": "
|
|
271
|
-
"forked": "
|
|
272
|
-
"stargazers_count":
|
|
273
|
-
"todayStar": "
|
|
274
|
-
"html_url": "https://github.com/
|
|
267
|
+
"full_name": "Mozilla-Ocho/llamafile",
|
|
268
|
+
"language": "C++",
|
|
269
|
+
"color": "#f34b7d",
|
|
270
|
+
"description": "Distribute and run LLMs with a single file.",
|
|
271
|
+
"forked": "754",
|
|
272
|
+
"stargazers_count": 15098,
|
|
273
|
+
"todayStar": "2830 stars this month",
|
|
274
|
+
"html_url": "https://github.com/Mozilla-Ocho/llamafile",
|
|
275
275
|
"rank": 25
|
|
276
276
|
}
|
|
277
277
|
]
|