@wcj/github-rank 24.5.6 → 24.5.8
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -1
- package/dist/repos.json +29356 -29356
- package/dist/trending-daily.json +190 -190
- package/dist/trending-monthly.json +169 -169
- package/dist/trending-weekly.json +183 -183
- package/dist/users.china.json +1900 -1795
- package/dist/users.json +5532 -5532
- package/dist/users.org.json +483 -483
- package/package.json +1 -1
- package/web/data/repos.json +29356 -29356
- package/web/data/trending-daily.json +190 -190
- package/web/data/trending-monthly.json +169 -169
- package/web/data/trending-weekly.json +183 -183
- package/web/data/users.china.json +1900 -1795
- package/web/data/users.json +5532 -5532
- package/web/data/users.org.json +483 -483
- package/web/index.html +5894 -5887
- package/web/org.html +373 -373
- package/web/repos.html +3711 -3711
- package/web/trending-monthly.html +163 -163
- package/web/trending-weekly.html +169 -165
- package/web/trending.html +169 -165
- package/web/users.china.html +1915 -1759
|
@@ -1,24 +1,24 @@
|
|
|
1
1
|
[
|
|
2
|
+
{
|
|
3
|
+
"full_name": "nashsu/FreeAskInternet",
|
|
4
|
+
"language": "Python",
|
|
5
|
+
"color": "#3572A5",
|
|
6
|
+
"description": "FreeAskInternet is a completely free, PRIVATE and LOCALLY running search aggregator & answer generate using MULTI LLMs, without GPU needed. The user can ask a question and the system will make a multi engine search and combine the search result to LLM and generate the answer based on search results. It's all FREE to use.",
|
|
7
|
+
"forked": "825",
|
|
8
|
+
"stargazers_count": 7636,
|
|
9
|
+
"todayStar": "6504 stars this month",
|
|
10
|
+
"html_url": "https://github.com/nashsu/FreeAskInternet",
|
|
11
|
+
"rank": 1
|
|
12
|
+
},
|
|
2
13
|
{
|
|
3
14
|
"full_name": "jwasham/coding-interview-university",
|
|
4
15
|
"language": "",
|
|
5
16
|
"color": "",
|
|
6
17
|
"description": "A complete computer science study plan to become a software engineer.",
|
|
7
|
-
"forked": "
|
|
8
|
-
"stargazers_count":
|
|
9
|
-
"todayStar": "
|
|
18
|
+
"forked": "73829",
|
|
19
|
+
"stargazers_count": 291431,
|
|
20
|
+
"todayStar": "10410 stars this month",
|
|
10
21
|
"html_url": "https://github.com/jwasham/coding-interview-university",
|
|
11
|
-
"rank": 1
|
|
12
|
-
},
|
|
13
|
-
{
|
|
14
|
-
"full_name": "langgenius/dify",
|
|
15
|
-
"language": "TypeScript",
|
|
16
|
-
"color": "#3178c6",
|
|
17
|
-
"description": "Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.",
|
|
18
|
-
"forked": "3477",
|
|
19
|
-
"stargazers_count": 26780,
|
|
20
|
-
"todayStar": "7636 stars this month",
|
|
21
|
-
"html_url": "https://github.com/langgenius/dify",
|
|
22
22
|
"rank": 2
|
|
23
23
|
},
|
|
24
24
|
{
|
|
@@ -26,252 +26,252 @@
|
|
|
26
26
|
"language": "Python",
|
|
27
27
|
"color": "#3572A5",
|
|
28
28
|
"description": "Instant voice cloning by MyShell.",
|
|
29
|
-
"forked": "
|
|
30
|
-
"stargazers_count":
|
|
31
|
-
"todayStar": "
|
|
29
|
+
"forked": "2252",
|
|
30
|
+
"stargazers_count": 24333,
|
|
31
|
+
"todayStar": "7522 stars this month",
|
|
32
32
|
"html_url": "https://github.com/myshell-ai/OpenVoice",
|
|
33
33
|
"rank": 3
|
|
34
34
|
},
|
|
35
35
|
{
|
|
36
|
-
"full_name": "
|
|
37
|
-
"language": "
|
|
38
|
-
"color": "#
|
|
39
|
-
"description": "
|
|
40
|
-
"forked": "
|
|
41
|
-
"stargazers_count":
|
|
42
|
-
"todayStar": "
|
|
43
|
-
"html_url": "https://github.com/
|
|
36
|
+
"full_name": "langgenius/dify",
|
|
37
|
+
"language": "TypeScript",
|
|
38
|
+
"color": "#3178c6",
|
|
39
|
+
"description": "Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.",
|
|
40
|
+
"forked": "3545",
|
|
41
|
+
"stargazers_count": 27455,
|
|
42
|
+
"todayStar": "8369 stars this month",
|
|
43
|
+
"html_url": "https://github.com/langgenius/dify",
|
|
44
44
|
"rank": 4
|
|
45
45
|
},
|
|
46
|
+
{
|
|
47
|
+
"full_name": "zyronon/douyin",
|
|
48
|
+
"language": "Vue",
|
|
49
|
+
"color": "#41b883",
|
|
50
|
+
"description": "Vue3 + Pinia 仿抖音,Vue 在移动端的最佳实践 . Imitate TikTok ,Vue Best practices on Mobile",
|
|
51
|
+
"forked": "1715",
|
|
52
|
+
"stargazers_count": 7067,
|
|
53
|
+
"todayStar": "4628 stars this month",
|
|
54
|
+
"html_url": "https://github.com/zyronon/douyin",
|
|
55
|
+
"rank": 5
|
|
56
|
+
},
|
|
46
57
|
{
|
|
47
58
|
"full_name": "infiniflow/ragflow",
|
|
48
59
|
"language": "Python",
|
|
49
60
|
"color": "#3572A5",
|
|
50
61
|
"description": "RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.",
|
|
51
|
-
"forked": "
|
|
52
|
-
"stargazers_count":
|
|
53
|
-
"todayStar": "
|
|
62
|
+
"forked": "562",
|
|
63
|
+
"stargazers_count": 6645,
|
|
64
|
+
"todayStar": "3887 stars this month",
|
|
54
65
|
"html_url": "https://github.com/infiniflow/ragflow",
|
|
55
|
-
"rank":
|
|
66
|
+
"rank": 6
|
|
56
67
|
},
|
|
57
68
|
{
|
|
58
|
-
"full_name": "
|
|
59
|
-
"language": "
|
|
60
|
-
"color": "#
|
|
61
|
-
"description": "
|
|
62
|
-
"forked": "
|
|
63
|
-
"stargazers_count":
|
|
64
|
-
"todayStar": "
|
|
65
|
-
"html_url": "https://github.com/
|
|
66
|
-
"rank":
|
|
69
|
+
"full_name": "FoundationVision/VAR",
|
|
70
|
+
"language": "Python",
|
|
71
|
+
"color": "#3572A5",
|
|
72
|
+
"description": "[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of \"Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction\". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!",
|
|
73
|
+
"forked": "272",
|
|
74
|
+
"stargazers_count": 3408,
|
|
75
|
+
"todayStar": "2592 stars this month",
|
|
76
|
+
"html_url": "https://github.com/FoundationVision/VAR",
|
|
77
|
+
"rank": 7
|
|
67
78
|
},
|
|
68
79
|
{
|
|
69
80
|
"full_name": "LlamaFamily/Llama-Chinese",
|
|
70
81
|
"language": "Python",
|
|
71
82
|
"color": "#3572A5",
|
|
72
83
|
"description": "Llama中文社区,Llama3在线体验和微调模型已开放,实时汇总最新Llama3学习资料,已将所有代码更新适配Llama3,构建最好的中文Llama大模型,完全开源可商用",
|
|
73
|
-
"forked": "
|
|
74
|
-
"stargazers_count":
|
|
75
|
-
"todayStar": "
|
|
84
|
+
"forked": "1084",
|
|
85
|
+
"stargazers_count": 11641,
|
|
86
|
+
"todayStar": "2578 stars this month",
|
|
76
87
|
"html_url": "https://github.com/LlamaFamily/Llama-Chinese",
|
|
77
|
-
"rank":
|
|
88
|
+
"rank": 8
|
|
89
|
+
},
|
|
90
|
+
{
|
|
91
|
+
"full_name": "phidatahq/phidata",
|
|
92
|
+
"language": "Python",
|
|
93
|
+
"color": "#3572A5",
|
|
94
|
+
"description": "Memory, knowledge and tools for LLMs",
|
|
95
|
+
"forked": "745",
|
|
96
|
+
"stargazers_count": 5518,
|
|
97
|
+
"todayStar": "2586 stars this month",
|
|
98
|
+
"html_url": "https://github.com/phidatahq/phidata",
|
|
99
|
+
"rank": 9
|
|
100
|
+
},
|
|
101
|
+
{
|
|
102
|
+
"full_name": "Asabeneh/30-Days-Of-Python",
|
|
103
|
+
"language": "Python",
|
|
104
|
+
"color": "#3572A5",
|
|
105
|
+
"description": "30 days of Python programming challenge is a step-by-step guide to learn the Python programming language in 30 days. This challenge may take more than100 days, follow your own pace. These videos may help too: https://www.youtube.com/channel/UC7PNRuno1rzYPb1xLa4yktw",
|
|
106
|
+
"forked": "6514",
|
|
107
|
+
"stargazers_count": 32061,
|
|
108
|
+
"todayStar": "2029 stars this month",
|
|
109
|
+
"html_url": "https://github.com/Asabeneh/30-Days-Of-Python",
|
|
110
|
+
"rank": 10
|
|
78
111
|
},
|
|
79
112
|
{
|
|
80
113
|
"full_name": "hiyouga/LLaMA-Factory",
|
|
81
114
|
"language": "Python",
|
|
82
115
|
"color": "#3572A5",
|
|
83
116
|
"description": "Unify Efficient Fine-Tuning of 100+ LLMs",
|
|
84
|
-
"forked": "
|
|
85
|
-
"stargazers_count":
|
|
86
|
-
"todayStar": "
|
|
117
|
+
"forked": "2559",
|
|
118
|
+
"stargazers_count": 21270,
|
|
119
|
+
"todayStar": "5629 stars this month",
|
|
87
120
|
"html_url": "https://github.com/hiyouga/LLaMA-Factory",
|
|
88
|
-
"rank":
|
|
121
|
+
"rank": 11
|
|
89
122
|
},
|
|
90
123
|
{
|
|
91
124
|
"full_name": "ollama/ollama",
|
|
92
125
|
"language": "Go",
|
|
93
126
|
"color": "#00ADD8",
|
|
94
127
|
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
|
95
|
-
"forked": "
|
|
96
|
-
"stargazers_count":
|
|
97
|
-
"todayStar": "
|
|
128
|
+
"forked": "4657",
|
|
129
|
+
"stargazers_count": 64631,
|
|
130
|
+
"todayStar": "11525 stars this month",
|
|
98
131
|
"html_url": "https://github.com/ollama/ollama",
|
|
99
|
-
"rank":
|
|
132
|
+
"rank": 12
|
|
100
133
|
},
|
|
101
134
|
{
|
|
102
|
-
"full_name": "
|
|
135
|
+
"full_name": "NaiboWang/EasySpider",
|
|
136
|
+
"language": "JavaScript",
|
|
137
|
+
"color": "#f1e05a",
|
|
138
|
+
"description": "A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:ServiceWrapper面向Web应用的智能化服务封装系统。",
|
|
139
|
+
"forked": "2690",
|
|
140
|
+
"stargazers_count": 23096,
|
|
141
|
+
"todayStar": "2720 stars this month",
|
|
142
|
+
"html_url": "https://github.com/NaiboWang/EasySpider",
|
|
143
|
+
"rank": 13
|
|
144
|
+
},
|
|
145
|
+
{
|
|
146
|
+
"full_name": "unslothai/unsloth",
|
|
103
147
|
"language": "Python",
|
|
104
148
|
"color": "#3572A5",
|
|
105
|
-
"description": "
|
|
106
|
-
"forked": "
|
|
107
|
-
"stargazers_count":
|
|
108
|
-
"todayStar": "
|
|
109
|
-
"html_url": "https://github.com/
|
|
110
|
-
"rank":
|
|
149
|
+
"description": "Finetune Llama 3, Mistral & Gemma LLMs 2-5x faster with 80% less memory",
|
|
150
|
+
"forked": "553",
|
|
151
|
+
"stargazers_count": 8811,
|
|
152
|
+
"todayStar": "3433 stars this month",
|
|
153
|
+
"html_url": "https://github.com/unslothai/unsloth",
|
|
154
|
+
"rank": 14
|
|
111
155
|
},
|
|
112
156
|
{
|
|
113
157
|
"full_name": "ordinals/ord",
|
|
114
158
|
"language": "Rust",
|
|
115
159
|
"color": "#dea584",
|
|
116
160
|
"description": "👁🗨 Rare and exotic sats",
|
|
117
|
-
"forked": "
|
|
118
|
-
"stargazers_count":
|
|
119
|
-
"todayStar": "
|
|
161
|
+
"forked": "1250",
|
|
162
|
+
"stargazers_count": 3671,
|
|
163
|
+
"todayStar": "487 stars this month",
|
|
120
164
|
"html_url": "https://github.com/ordinals/ord",
|
|
121
|
-
"rank":
|
|
165
|
+
"rank": 15
|
|
122
166
|
},
|
|
123
167
|
{
|
|
124
|
-
"full_name": "
|
|
125
|
-
"language": "
|
|
126
|
-
"color": "#
|
|
127
|
-
"description": "
|
|
128
|
-
"forked": "
|
|
129
|
-
"stargazers_count":
|
|
130
|
-
"todayStar": "
|
|
131
|
-
"html_url": "https://github.com/
|
|
132
|
-
"rank":
|
|
168
|
+
"full_name": "wandb/openui",
|
|
169
|
+
"language": "TypeScript",
|
|
170
|
+
"color": "#3178c6",
|
|
171
|
+
"description": "OpenUI let's you describe UI using your imagination, then see it rendered live.",
|
|
172
|
+
"forked": "866",
|
|
173
|
+
"stargazers_count": 10689,
|
|
174
|
+
"todayStar": "5760 stars this month",
|
|
175
|
+
"html_url": "https://github.com/wandb/openui",
|
|
176
|
+
"rank": 16
|
|
133
177
|
},
|
|
134
178
|
{
|
|
135
179
|
"full_name": "meta-llama/llama-recipes",
|
|
136
180
|
"language": "JupyterNotebook",
|
|
137
181
|
"color": "#DA5B0B",
|
|
138
182
|
"description": "Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama3 for WhatsApp & Messenger.",
|
|
139
|
-
"forked": "
|
|
140
|
-
"stargazers_count":
|
|
141
|
-
"todayStar": "
|
|
183
|
+
"forked": "1338",
|
|
184
|
+
"stargazers_count": 9500,
|
|
185
|
+
"todayStar": "1381 stars this month",
|
|
142
186
|
"html_url": "https://github.com/meta-llama/llama-recipes",
|
|
143
|
-
"rank":
|
|
144
|
-
},
|
|
145
|
-
{
|
|
146
|
-
"full_name": "karpathy/llama2.c",
|
|
147
|
-
"language": "C",
|
|
148
|
-
"color": "#555555",
|
|
149
|
-
"description": "Inference Llama 2 in one file of pure C",
|
|
150
|
-
"forked": "1830",
|
|
151
|
-
"stargazers_count": 16074,
|
|
152
|
-
"todayStar": "1292 stars this month",
|
|
153
|
-
"html_url": "https://github.com/karpathy/llama2.c",
|
|
154
|
-
"rank": 14
|
|
155
|
-
},
|
|
156
|
-
{
|
|
157
|
-
"full_name": "Mozilla-Ocho/llamafile",
|
|
158
|
-
"language": "C++",
|
|
159
|
-
"color": "#f34b7d",
|
|
160
|
-
"description": "Distribute and run LLMs with a single file.",
|
|
161
|
-
"forked": "743",
|
|
162
|
-
"stargazers_count": 15035,
|
|
163
|
-
"todayStar": "2976 stars this month",
|
|
164
|
-
"html_url": "https://github.com/Mozilla-Ocho/llamafile",
|
|
165
|
-
"rank": 15
|
|
187
|
+
"rank": 17
|
|
166
188
|
},
|
|
167
189
|
{
|
|
168
190
|
"full_name": "pytorch/torchtune",
|
|
169
191
|
"language": "Python",
|
|
170
192
|
"color": "#3572A5",
|
|
171
193
|
"description": "A Native-PyTorch Library for LLM Fine-tuning",
|
|
172
|
-
"forked": "
|
|
173
|
-
"stargazers_count":
|
|
174
|
-
"todayStar": "
|
|
194
|
+
"forked": "200",
|
|
195
|
+
"stargazers_count": 3077,
|
|
196
|
+
"todayStar": "2344 stars this month",
|
|
175
197
|
"html_url": "https://github.com/pytorch/torchtune",
|
|
176
|
-
"rank":
|
|
198
|
+
"rank": 18
|
|
177
199
|
},
|
|
178
200
|
{
|
|
179
201
|
"full_name": "open-webui/open-webui",
|
|
180
202
|
"language": "Svelte",
|
|
181
203
|
"color": "#ff3e00",
|
|
182
204
|
"description": "User-friendly WebUI for LLMs (Formerly Ollama WebUI)",
|
|
183
|
-
"forked": "
|
|
184
|
-
"stargazers_count":
|
|
185
|
-
"todayStar": "
|
|
205
|
+
"forked": "1939",
|
|
206
|
+
"stargazers_count": 19190,
|
|
207
|
+
"todayStar": "7019 stars this month",
|
|
186
208
|
"html_url": "https://github.com/open-webui/open-webui",
|
|
187
|
-
"rank": 17
|
|
188
|
-
},
|
|
189
|
-
{
|
|
190
|
-
"full_name": "CorentinTh/it-tools",
|
|
191
|
-
"language": "Vue",
|
|
192
|
-
"color": "#41b883",
|
|
193
|
-
"description": "Collection of handy online tools for developers, with great UX.",
|
|
194
|
-
"forked": "1445",
|
|
195
|
-
"stargazers_count": 11606,
|
|
196
|
-
"todayStar": "2302 stars this month",
|
|
197
|
-
"html_url": "https://github.com/CorentinTh/it-tools",
|
|
198
|
-
"rank": 18
|
|
199
|
-
},
|
|
200
|
-
{
|
|
201
|
-
"full_name": "LLM-Red-Team/kimi-free-api",
|
|
202
|
-
"language": "TypeScript",
|
|
203
|
-
"color": "#3178c6",
|
|
204
|
-
"description": "🚀 KIMI AI 长文本大模型逆向API白嫖测试【特长:长文本解读整理】,支持高速流式输出、智能体对话、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。",
|
|
205
|
-
"forked": "366",
|
|
206
|
-
"stargazers_count": 2568,
|
|
207
|
-
"todayStar": "1504 stars this month",
|
|
208
|
-
"html_url": "https://github.com/LLM-Red-Team/kimi-free-api",
|
|
209
209
|
"rank": 19
|
|
210
210
|
},
|
|
211
|
-
{
|
|
212
|
-
"full_name": "phidatahq/phidata",
|
|
213
|
-
"language": "Python",
|
|
214
|
-
"color": "#3572A5",
|
|
215
|
-
"description": "Add memory, knowledge and tools to LLMs",
|
|
216
|
-
"forked": "701",
|
|
217
|
-
"stargazers_count": 5154,
|
|
218
|
-
"todayStar": "2211 stars this month",
|
|
219
|
-
"html_url": "https://github.com/phidatahq/phidata",
|
|
220
|
-
"rank": 20
|
|
221
|
-
},
|
|
222
211
|
{
|
|
223
212
|
"full_name": "paul-gauthier/aider",
|
|
224
213
|
"language": "Python",
|
|
225
214
|
"color": "#3572A5",
|
|
226
215
|
"description": "aider is AI pair programming in your terminal",
|
|
227
|
-
"forked": "
|
|
228
|
-
"stargazers_count":
|
|
229
|
-
"todayStar": "
|
|
216
|
+
"forked": "975",
|
|
217
|
+
"stargazers_count": 9855,
|
|
218
|
+
"todayStar": "2501 stars this month",
|
|
230
219
|
"html_url": "https://github.com/paul-gauthier/aider",
|
|
220
|
+
"rank": 20
|
|
221
|
+
},
|
|
222
|
+
{
|
|
223
|
+
"full_name": "CorentinTh/it-tools",
|
|
224
|
+
"language": "Vue",
|
|
225
|
+
"color": "#41b883",
|
|
226
|
+
"description": "Collection of handy online tools for developers, with great UX.",
|
|
227
|
+
"forked": "1464",
|
|
228
|
+
"stargazers_count": 11757,
|
|
229
|
+
"todayStar": "2344 stars this month",
|
|
230
|
+
"html_url": "https://github.com/CorentinTh/it-tools",
|
|
231
231
|
"rank": 21
|
|
232
232
|
},
|
|
233
233
|
{
|
|
234
|
-
"full_name": "
|
|
235
|
-
"language": "
|
|
236
|
-
"color": "#
|
|
237
|
-
"description": "
|
|
238
|
-
"forked": "
|
|
239
|
-
"stargazers_count":
|
|
240
|
-
"todayStar": "
|
|
241
|
-
"html_url": "https://github.com/
|
|
234
|
+
"full_name": "karpathy/llama2.c",
|
|
235
|
+
"language": "C",
|
|
236
|
+
"color": "#555555",
|
|
237
|
+
"description": "Inference Llama 2 in one file of pure C",
|
|
238
|
+
"forked": "1841",
|
|
239
|
+
"stargazers_count": 16113,
|
|
240
|
+
"todayStar": "1332 stars this month",
|
|
241
|
+
"html_url": "https://github.com/karpathy/llama2.c",
|
|
242
242
|
"rank": 22
|
|
243
243
|
},
|
|
244
244
|
{
|
|
245
|
-
"full_name": "
|
|
246
|
-
"language": "
|
|
247
|
-
"color": "#
|
|
248
|
-
"description": "
|
|
249
|
-
"forked": "
|
|
250
|
-
"stargazers_count":
|
|
251
|
-
"todayStar": "
|
|
252
|
-
"html_url": "https://github.com/
|
|
245
|
+
"full_name": "LLM-Red-Team/kimi-free-api",
|
|
246
|
+
"language": "TypeScript",
|
|
247
|
+
"color": "#3178c6",
|
|
248
|
+
"description": "🚀 KIMI AI 长文本大模型逆向API白嫖测试【特长:长文本解读整理】,支持高速流式输出、智能体对话、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。",
|
|
249
|
+
"forked": "379",
|
|
250
|
+
"stargazers_count": 2654,
|
|
251
|
+
"todayStar": "1429 stars this month",
|
|
252
|
+
"html_url": "https://github.com/LLM-Red-Team/kimi-free-api",
|
|
253
253
|
"rank": 23
|
|
254
254
|
},
|
|
255
|
-
{
|
|
256
|
-
"full_name": "missuo/FreeGPT35",
|
|
257
|
-
"language": "JavaScript",
|
|
258
|
-
"color": "#f1e05a",
|
|
259
|
-
"description": "Utilize the unlimited free GPT-3.5-Turbo API service provided by the login-free ChatGPT Web.",
|
|
260
|
-
"forked": "918",
|
|
261
|
-
"stargazers_count": 3072,
|
|
262
|
-
"todayStar": "1879 stars this month",
|
|
263
|
-
"html_url": "https://github.com/missuo/FreeGPT35",
|
|
264
|
-
"rank": 24
|
|
265
|
-
},
|
|
266
255
|
{
|
|
267
256
|
"full_name": "quilljs/quill",
|
|
268
257
|
"language": "TypeScript",
|
|
269
258
|
"color": "#3178c6",
|
|
270
259
|
"description": "Quill is a modern WYSIWYG editor built for compatibility and extensibility.",
|
|
271
|
-
"forked": "
|
|
272
|
-
"stargazers_count":
|
|
273
|
-
"todayStar": "
|
|
260
|
+
"forked": "3256",
|
|
261
|
+
"stargazers_count": 41432,
|
|
262
|
+
"todayStar": "2666 stars this month",
|
|
274
263
|
"html_url": "https://github.com/quilljs/quill",
|
|
264
|
+
"rank": 24
|
|
265
|
+
},
|
|
266
|
+
{
|
|
267
|
+
"full_name": "Mozilla-Ocho/llamafile",
|
|
268
|
+
"language": "C++",
|
|
269
|
+
"color": "#f34b7d",
|
|
270
|
+
"description": "Distribute and run LLMs with a single file.",
|
|
271
|
+
"forked": "754",
|
|
272
|
+
"stargazers_count": 15098,
|
|
273
|
+
"todayStar": "2830 stars this month",
|
|
274
|
+
"html_url": "https://github.com/Mozilla-Ocho/llamafile",
|
|
275
275
|
"rank": 25
|
|
276
276
|
}
|
|
277
277
|
]
|