webscout 8.2.8__py3-none-any.whl → 8.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of webscout might be problematic. Click here for more details.
- webscout/AIauto.py +34 -16
- webscout/AIbase.py +96 -37
- webscout/AIutel.py +491 -87
- webscout/Bard.py +441 -323
- webscout/Extra/GitToolkit/__init__.py +10 -10
- webscout/Extra/YTToolkit/ytapi/video.py +232 -232
- webscout/Litlogger/README.md +10 -0
- webscout/Litlogger/__init__.py +7 -59
- webscout/Litlogger/formats.py +4 -0
- webscout/Litlogger/handlers.py +103 -0
- webscout/Litlogger/levels.py +13 -0
- webscout/Litlogger/logger.py +92 -0
- webscout/Provider/AISEARCH/Perplexity.py +332 -358
- webscout/Provider/AISEARCH/felo_search.py +9 -35
- webscout/Provider/AISEARCH/genspark_search.py +30 -56
- webscout/Provider/AISEARCH/hika_search.py +4 -16
- webscout/Provider/AISEARCH/iask_search.py +410 -436
- webscout/Provider/AISEARCH/monica_search.py +4 -30
- webscout/Provider/AISEARCH/scira_search.py +6 -32
- webscout/Provider/AISEARCH/webpilotai_search.py +38 -64
- webscout/Provider/Blackboxai.py +155 -35
- webscout/Provider/ChatSandbox.py +2 -1
- webscout/Provider/Deepinfra.py +339 -339
- webscout/Provider/ExaChat.py +358 -358
- webscout/Provider/Gemini.py +169 -169
- webscout/Provider/GithubChat.py +1 -2
- webscout/Provider/Glider.py +3 -3
- webscout/Provider/HeckAI.py +172 -82
- webscout/Provider/LambdaChat.py +1 -0
- webscout/Provider/MCPCore.py +7 -3
- webscout/Provider/OPENAI/BLACKBOXAI.py +421 -139
- webscout/Provider/OPENAI/Cloudflare.py +38 -21
- webscout/Provider/OPENAI/FalconH1.py +457 -0
- webscout/Provider/OPENAI/FreeGemini.py +35 -18
- webscout/Provider/OPENAI/NEMOTRON.py +34 -34
- webscout/Provider/OPENAI/PI.py +427 -0
- webscout/Provider/OPENAI/Qwen3.py +304 -0
- webscout/Provider/OPENAI/README.md +952 -1253
- webscout/Provider/OPENAI/TwoAI.py +374 -0
- webscout/Provider/OPENAI/__init__.py +7 -1
- webscout/Provider/OPENAI/ai4chat.py +73 -63
- webscout/Provider/OPENAI/api.py +869 -644
- webscout/Provider/OPENAI/base.py +2 -0
- webscout/Provider/OPENAI/c4ai.py +34 -13
- webscout/Provider/OPENAI/chatgpt.py +575 -556
- webscout/Provider/OPENAI/chatgptclone.py +512 -487
- webscout/Provider/OPENAI/chatsandbox.py +11 -6
- webscout/Provider/OPENAI/copilot.py +258 -0
- webscout/Provider/OPENAI/deepinfra.py +327 -318
- webscout/Provider/OPENAI/e2b.py +140 -104
- webscout/Provider/OPENAI/exaai.py +420 -411
- webscout/Provider/OPENAI/exachat.py +448 -443
- webscout/Provider/OPENAI/flowith.py +7 -3
- webscout/Provider/OPENAI/freeaichat.py +12 -8
- webscout/Provider/OPENAI/glider.py +15 -8
- webscout/Provider/OPENAI/groq.py +5 -2
- webscout/Provider/OPENAI/heckai.py +311 -307
- webscout/Provider/OPENAI/llmchatco.py +9 -7
- webscout/Provider/OPENAI/mcpcore.py +18 -9
- webscout/Provider/OPENAI/multichat.py +7 -5
- webscout/Provider/OPENAI/netwrck.py +16 -11
- webscout/Provider/OPENAI/oivscode.py +290 -0
- webscout/Provider/OPENAI/opkfc.py +507 -496
- webscout/Provider/OPENAI/pydantic_imports.py +172 -0
- webscout/Provider/OPENAI/scirachat.py +29 -17
- webscout/Provider/OPENAI/sonus.py +308 -303
- webscout/Provider/OPENAI/standardinput.py +442 -433
- webscout/Provider/OPENAI/textpollinations.py +18 -11
- webscout/Provider/OPENAI/toolbaz.py +419 -413
- webscout/Provider/OPENAI/typefully.py +17 -10
- webscout/Provider/OPENAI/typegpt.py +21 -11
- webscout/Provider/OPENAI/uncovrAI.py +477 -462
- webscout/Provider/OPENAI/utils.py +90 -79
- webscout/Provider/OPENAI/venice.py +435 -425
- webscout/Provider/OPENAI/wisecat.py +387 -381
- webscout/Provider/OPENAI/writecream.py +166 -163
- webscout/Provider/OPENAI/x0gpt.py +26 -37
- webscout/Provider/OPENAI/yep.py +384 -356
- webscout/Provider/PI.py +2 -1
- webscout/Provider/TTI/README.md +55 -101
- webscout/Provider/TTI/__init__.py +4 -9
- webscout/Provider/TTI/aiarta.py +365 -0
- webscout/Provider/TTI/artbit.py +0 -0
- webscout/Provider/TTI/base.py +64 -0
- webscout/Provider/TTI/fastflux.py +200 -0
- webscout/Provider/TTI/magicstudio.py +201 -0
- webscout/Provider/TTI/piclumen.py +203 -0
- webscout/Provider/TTI/pixelmuse.py +225 -0
- webscout/Provider/TTI/pollinations.py +221 -0
- webscout/Provider/TTI/utils.py +11 -0
- webscout/Provider/TTS/__init__.py +2 -1
- webscout/Provider/TTS/base.py +159 -159
- webscout/Provider/TTS/openai_fm.py +129 -0
- webscout/Provider/TextPollinationsAI.py +308 -308
- webscout/Provider/TwoAI.py +239 -44
- webscout/Provider/UNFINISHED/Youchat.py +330 -330
- webscout/Provider/UNFINISHED/puterjs.py +635 -0
- webscout/Provider/UNFINISHED/test_lmarena.py +119 -119
- webscout/Provider/Writecream.py +246 -246
- webscout/Provider/__init__.py +2 -2
- webscout/Provider/ai4chat.py +33 -8
- webscout/Provider/granite.py +41 -6
- webscout/Provider/koala.py +169 -169
- webscout/Provider/oivscode.py +309 -0
- webscout/Provider/samurai.py +3 -2
- webscout/Provider/scnet.py +1 -0
- webscout/Provider/typegpt.py +3 -3
- webscout/Provider/uncovr.py +368 -368
- webscout/client.py +70 -0
- webscout/litprinter/__init__.py +58 -58
- webscout/optimizers.py +419 -419
- webscout/scout/README.md +3 -1
- webscout/scout/core/crawler.py +134 -64
- webscout/scout/core/scout.py +148 -109
- webscout/scout/element.py +106 -88
- webscout/swiftcli/Readme.md +323 -323
- webscout/swiftcli/plugins/manager.py +9 -2
- webscout/version.py +1 -1
- webscout/zeroart/__init__.py +134 -134
- webscout/zeroart/effects.py +100 -100
- webscout/zeroart/fonts.py +1238 -1238
- {webscout-8.2.8.dist-info → webscout-8.3.dist-info}/METADATA +160 -35
- webscout-8.3.dist-info/RECORD +290 -0
- {webscout-8.2.8.dist-info → webscout-8.3.dist-info}/WHEEL +1 -1
- {webscout-8.2.8.dist-info → webscout-8.3.dist-info}/entry_points.txt +1 -0
- webscout/Litlogger/Readme.md +0 -175
- webscout/Litlogger/core/__init__.py +0 -6
- webscout/Litlogger/core/level.py +0 -23
- webscout/Litlogger/core/logger.py +0 -165
- webscout/Litlogger/handlers/__init__.py +0 -12
- webscout/Litlogger/handlers/console.py +0 -33
- webscout/Litlogger/handlers/file.py +0 -143
- webscout/Litlogger/handlers/network.py +0 -173
- webscout/Litlogger/styles/__init__.py +0 -7
- webscout/Litlogger/styles/colors.py +0 -249
- webscout/Litlogger/styles/formats.py +0 -458
- webscout/Litlogger/styles/text.py +0 -87
- webscout/Litlogger/utils/__init__.py +0 -6
- webscout/Litlogger/utils/detectors.py +0 -153
- webscout/Litlogger/utils/formatters.py +0 -200
- webscout/Provider/ChatGPTGratis.py +0 -194
- webscout/Provider/TTI/AiForce/README.md +0 -159
- webscout/Provider/TTI/AiForce/__init__.py +0 -22
- webscout/Provider/TTI/AiForce/async_aiforce.py +0 -224
- webscout/Provider/TTI/AiForce/sync_aiforce.py +0 -245
- webscout/Provider/TTI/FreeAIPlayground/README.md +0 -99
- webscout/Provider/TTI/FreeAIPlayground/__init__.py +0 -9
- webscout/Provider/TTI/FreeAIPlayground/async_freeaiplayground.py +0 -181
- webscout/Provider/TTI/FreeAIPlayground/sync_freeaiplayground.py +0 -180
- webscout/Provider/TTI/ImgSys/README.md +0 -174
- webscout/Provider/TTI/ImgSys/__init__.py +0 -23
- webscout/Provider/TTI/ImgSys/async_imgsys.py +0 -202
- webscout/Provider/TTI/ImgSys/sync_imgsys.py +0 -195
- webscout/Provider/TTI/MagicStudio/README.md +0 -101
- webscout/Provider/TTI/MagicStudio/__init__.py +0 -2
- webscout/Provider/TTI/MagicStudio/async_magicstudio.py +0 -111
- webscout/Provider/TTI/MagicStudio/sync_magicstudio.py +0 -109
- webscout/Provider/TTI/Nexra/README.md +0 -155
- webscout/Provider/TTI/Nexra/__init__.py +0 -22
- webscout/Provider/TTI/Nexra/async_nexra.py +0 -286
- webscout/Provider/TTI/Nexra/sync_nexra.py +0 -258
- webscout/Provider/TTI/PollinationsAI/README.md +0 -146
- webscout/Provider/TTI/PollinationsAI/__init__.py +0 -23
- webscout/Provider/TTI/PollinationsAI/async_pollinations.py +0 -311
- webscout/Provider/TTI/PollinationsAI/sync_pollinations.py +0 -265
- webscout/Provider/TTI/aiarta/README.md +0 -134
- webscout/Provider/TTI/aiarta/__init__.py +0 -2
- webscout/Provider/TTI/aiarta/async_aiarta.py +0 -482
- webscout/Provider/TTI/aiarta/sync_aiarta.py +0 -440
- webscout/Provider/TTI/artbit/README.md +0 -100
- webscout/Provider/TTI/artbit/__init__.py +0 -22
- webscout/Provider/TTI/artbit/async_artbit.py +0 -155
- webscout/Provider/TTI/artbit/sync_artbit.py +0 -148
- webscout/Provider/TTI/fastflux/README.md +0 -129
- webscout/Provider/TTI/fastflux/__init__.py +0 -22
- webscout/Provider/TTI/fastflux/async_fastflux.py +0 -261
- webscout/Provider/TTI/fastflux/sync_fastflux.py +0 -252
- webscout/Provider/TTI/huggingface/README.md +0 -114
- webscout/Provider/TTI/huggingface/__init__.py +0 -22
- webscout/Provider/TTI/huggingface/async_huggingface.py +0 -199
- webscout/Provider/TTI/huggingface/sync_huggingface.py +0 -195
- webscout/Provider/TTI/piclumen/README.md +0 -161
- webscout/Provider/TTI/piclumen/__init__.py +0 -23
- webscout/Provider/TTI/piclumen/async_piclumen.py +0 -268
- webscout/Provider/TTI/piclumen/sync_piclumen.py +0 -233
- webscout/Provider/TTI/pixelmuse/README.md +0 -79
- webscout/Provider/TTI/pixelmuse/__init__.py +0 -4
- webscout/Provider/TTI/pixelmuse/async_pixelmuse.py +0 -249
- webscout/Provider/TTI/pixelmuse/sync_pixelmuse.py +0 -182
- webscout/Provider/TTI/talkai/README.md +0 -139
- webscout/Provider/TTI/talkai/__init__.py +0 -4
- webscout/Provider/TTI/talkai/async_talkai.py +0 -229
- webscout/Provider/TTI/talkai/sync_talkai.py +0 -207
- webscout/Provider/UNFINISHED/oivscode.py +0 -351
- webscout-8.2.8.dist-info/RECORD +0 -334
- {webscout-8.2.8.dist-info → webscout-8.3.dist-info}/licenses/LICENSE.md +0 -0
- {webscout-8.2.8.dist-info → webscout-8.3.dist-info}/top_level.txt +0 -0
|
@@ -1,10 +1,10 @@
|
|
|
1
|
-
from .gitapi import *
|
|
2
|
-
|
|
3
|
-
__all__ = [
|
|
4
|
-
'Repository',
|
|
5
|
-
'User',
|
|
6
|
-
'GitError',
|
|
7
|
-
'RateLimitError',
|
|
8
|
-
'NotFoundError',
|
|
9
|
-
'RequestError'
|
|
10
|
-
]
|
|
1
|
+
from .gitapi import *
|
|
2
|
+
|
|
3
|
+
__all__ = [
|
|
4
|
+
'Repository',
|
|
5
|
+
'User',
|
|
6
|
+
'GitError',
|
|
7
|
+
'RateLimitError',
|
|
8
|
+
'NotFoundError',
|
|
9
|
+
'RequestError'
|
|
10
|
+
]
|
|
@@ -1,232 +1,232 @@
|
|
|
1
|
-
import re
|
|
2
|
-
import json
|
|
3
|
-
from typing import Dict, Any
|
|
4
|
-
from .https import video_data
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
class Video:
|
|
8
|
-
|
|
9
|
-
_HEAD = 'https://www.youtube.com/watch?v='
|
|
10
|
-
|
|
11
|
-
def __init__(self, video_id: str):
|
|
12
|
-
"""
|
|
13
|
-
Represents a YouTube video
|
|
14
|
-
|
|
15
|
-
Parameters
|
|
16
|
-
----------
|
|
17
|
-
video_id : str
|
|
18
|
-
The id or url of the video
|
|
19
|
-
"""
|
|
20
|
-
pattern = re.compile('.be/(.*?)$|=(.*?)$|^(\w{11})$') # noqa
|
|
21
|
-
match = pattern.search(video_id)
|
|
22
|
-
|
|
23
|
-
if not match:
|
|
24
|
-
raise ValueError('Invalid YouTube video ID or URL')
|
|
25
|
-
|
|
26
|
-
self._matched_id = (
|
|
27
|
-
match.group(1)
|
|
28
|
-
or match.group(2)
|
|
29
|
-
or match.group(3)
|
|
30
|
-
)
|
|
31
|
-
|
|
32
|
-
if self._matched_id:
|
|
33
|
-
self._url = self._HEAD + self._matched_id
|
|
34
|
-
self._video_data = video_data(self._matched_id)
|
|
35
|
-
# Extract basic info for fallback
|
|
36
|
-
title_match = re.search('<title>(.*?) - YouTube</title>', self._video_data)
|
|
37
|
-
self.title = title_match.group(1) if title_match else None
|
|
38
|
-
self.id = self._matched_id
|
|
39
|
-
else:
|
|
40
|
-
raise ValueError('Invalid YouTube video ID or URL')
|
|
41
|
-
|
|
42
|
-
def __repr__(self):
|
|
43
|
-
return f'<Video {self._url}>'
|
|
44
|
-
|
|
45
|
-
@property
|
|
46
|
-
def metadata(self) -> Dict[str, Any]:
|
|
47
|
-
"""
|
|
48
|
-
Fetches video metadata in a dict format
|
|
49
|
-
|
|
50
|
-
Returns
|
|
51
|
-
-------
|
|
52
|
-
Dict
|
|
53
|
-
Video metadata in a dict format containing keys: title, id, views, duration, author_id,
|
|
54
|
-
upload_date, url, thumbnails, tags, description, likes, genre, etc.
|
|
55
|
-
"""
|
|
56
|
-
# Multiple patterns to try for video details extraction for robustness
|
|
57
|
-
details_patterns = [
|
|
58
|
-
re.compile('videoDetails\":(.*?)\"isLiveContent\":.*?}'),
|
|
59
|
-
re.compile('videoDetails\":(.*?),\"playerConfig'),
|
|
60
|
-
re.compile('videoDetails\":(.*?),\"playabilityStatus')
|
|
61
|
-
]
|
|
62
|
-
|
|
63
|
-
# Other metadata patterns
|
|
64
|
-
upload_date_pattern = re.compile("<meta itemprop=\"uploadDate\" content=\"(.*?)\">")
|
|
65
|
-
genre_pattern = re.compile("<meta itemprop=\"genre\" content=\"(.*?)\">")
|
|
66
|
-
like_count_patterns = [
|
|
67
|
-
re.compile("iconType\":\"LIKE\"},\"defaultText\":(.*?)}"),
|
|
68
|
-
re.compile('\"likeCount\":\"(\\d+)\"')
|
|
69
|
-
]
|
|
70
|
-
channel_name_pattern = re.compile('"ownerChannelName":"(.*?)"')
|
|
71
|
-
|
|
72
|
-
# Try each pattern for video details
|
|
73
|
-
raw_details_match = None
|
|
74
|
-
for pattern in details_patterns:
|
|
75
|
-
match = pattern.search(self._video_data)
|
|
76
|
-
if match:
|
|
77
|
-
raw_details_match = match
|
|
78
|
-
break
|
|
79
|
-
|
|
80
|
-
if not raw_details_match:
|
|
81
|
-
# Fallback metadata for search results or incomplete video data
|
|
82
|
-
return {
|
|
83
|
-
'title': getattr(self, 'title', None),
|
|
84
|
-
'id': getattr(self, 'id', None),
|
|
85
|
-
'views': getattr(self, 'views', None),
|
|
86
|
-
'streamed': False,
|
|
87
|
-
'duration': None,
|
|
88
|
-
'author_id': None,
|
|
89
|
-
'author_name': None,
|
|
90
|
-
'upload_date': None,
|
|
91
|
-
'url': f"https://www.youtube.com/watch?v={getattr(self, 'id', '')}" if hasattr(self, 'id') else None,
|
|
92
|
-
'thumbnails': None,
|
|
93
|
-
'tags': None,
|
|
94
|
-
'description': None,
|
|
95
|
-
'likes': None,
|
|
96
|
-
'genre': None,
|
|
97
|
-
'is_age_restricted': 'age-restricted' in self._video_data.lower(),
|
|
98
|
-
'is_unlisted': 'unlisted' in self._video_data.lower()
|
|
99
|
-
}
|
|
100
|
-
|
|
101
|
-
raw_details = raw_details_match.group(0)
|
|
102
|
-
|
|
103
|
-
# Extract upload date
|
|
104
|
-
upload_date_match = upload_date_pattern.search(self._video_data)
|
|
105
|
-
upload_date = upload_date_match.group(1) if upload_date_match else None
|
|
106
|
-
|
|
107
|
-
# Extract channel name
|
|
108
|
-
channel_name_match = channel_name_pattern.search(self._video_data)
|
|
109
|
-
channel_name = channel_name_match.group(1) if channel_name_match else None
|
|
110
|
-
|
|
111
|
-
# Parse video details
|
|
112
|
-
try:
|
|
113
|
-
# Clean up the JSON string for parsing
|
|
114
|
-
clean_json = raw_details.replace('videoDetails\":', '')
|
|
115
|
-
# Handle potential JSON parsing issues
|
|
116
|
-
if clean_json.endswith(','):
|
|
117
|
-
clean_json = clean_json[:-1]
|
|
118
|
-
metadata = json.loads(clean_json)
|
|
119
|
-
|
|
120
|
-
data = {
|
|
121
|
-
'title': metadata.get('title'),
|
|
122
|
-
'id': metadata.get('videoId', self._matched_id),
|
|
123
|
-
'views': metadata.get('viewCount'),
|
|
124
|
-
'streamed': metadata.get('isLiveContent', False),
|
|
125
|
-
'duration': metadata.get('lengthSeconds'),
|
|
126
|
-
'author_id': metadata.get('channelId'),
|
|
127
|
-
'author_name': channel_name or metadata.get('author'),
|
|
128
|
-
'upload_date': upload_date,
|
|
129
|
-
'url': f"https://www.youtube.com/watch?v={metadata.get('videoId', self._matched_id)}",
|
|
130
|
-
'thumbnails': metadata.get('thumbnail', {}).get('thumbnails'),
|
|
131
|
-
'tags': metadata.get('keywords'),
|
|
132
|
-
'description': metadata.get('shortDescription'),
|
|
133
|
-
'is_age_restricted': metadata.get('isAgeRestricted', False) or 'age-restricted' in self._video_data.lower(),
|
|
134
|
-
'is_unlisted': 'unlisted' in self._video_data.lower(),
|
|
135
|
-
'is_family_safe': metadata.get('isFamilySafe', True),
|
|
136
|
-
'is_private': metadata.get('isPrivate', False),
|
|
137
|
-
'is_live_content': metadata.get('isLiveContent', False),
|
|
138
|
-
'is_crawlable': metadata.get('isCrawlable', True),
|
|
139
|
-
'allow_ratings': metadata.get('allowRatings', True)
|
|
140
|
-
}
|
|
141
|
-
except (json.JSONDecodeError, KeyError, TypeError) as e:
|
|
142
|
-
# Fallback to basic metadata if JSON parsing fails
|
|
143
|
-
return {
|
|
144
|
-
'title': getattr(self, 'title', None),
|
|
145
|
-
'id': self._matched_id,
|
|
146
|
-
'url': self._url,
|
|
147
|
-
'error': f"Failed to parse video details: {str(e)}"
|
|
148
|
-
}
|
|
149
|
-
|
|
150
|
-
# Try to extract likes count
|
|
151
|
-
likes = None
|
|
152
|
-
for pattern in like_count_patterns:
|
|
153
|
-
try:
|
|
154
|
-
likes_match = pattern.search(self._video_data)
|
|
155
|
-
if likes_match:
|
|
156
|
-
likes_text = likes_match.group(1)
|
|
157
|
-
# Handle different formats of like count
|
|
158
|
-
if '{' in likes_text:
|
|
159
|
-
likes = json.loads(likes_text + '}}}')['accessibility']['accessibilityData']['label'].split(' ')[0].replace(',', '')
|
|
160
|
-
else:
|
|
161
|
-
likes = likes_text
|
|
162
|
-
break
|
|
163
|
-
except (AttributeError, KeyError, json.decoder.JSONDecodeError):
|
|
164
|
-
continue
|
|
165
|
-
|
|
166
|
-
data['likes'] = likes
|
|
167
|
-
|
|
168
|
-
# Try to extract genre
|
|
169
|
-
try:
|
|
170
|
-
genre_match = genre_pattern.search(self._video_data)
|
|
171
|
-
data['genre'] = genre_match.group(1) if genre_match else None
|
|
172
|
-
except AttributeError:
|
|
173
|
-
data['genre'] = None
|
|
174
|
-
|
|
175
|
-
return data
|
|
176
|
-
|
|
177
|
-
|
|
178
|
-
|
|
179
|
-
@property
|
|
180
|
-
def embed_html(self) -> str:
|
|
181
|
-
"""
|
|
182
|
-
Get the embed HTML code for this video
|
|
183
|
-
|
|
184
|
-
Returns:
|
|
185
|
-
HTML iframe code for embedding the video
|
|
186
|
-
"""
|
|
187
|
-
return f'<iframe width="560" height="315" src="https://www.youtube.com/embed/{self._matched_id}" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture" allowfullscreen></iframe>'
|
|
188
|
-
|
|
189
|
-
@property
|
|
190
|
-
def embed_url(self) -> str:
|
|
191
|
-
"""
|
|
192
|
-
Get the embed URL for this video
|
|
193
|
-
|
|
194
|
-
Returns:
|
|
195
|
-
URL for embedding the video
|
|
196
|
-
"""
|
|
197
|
-
return f'https://www.youtube.com/embed/{self._matched_id}'
|
|
198
|
-
|
|
199
|
-
@property
|
|
200
|
-
def thumbnail_url(self) -> str:
|
|
201
|
-
"""
|
|
202
|
-
Get the thumbnail URL for this video
|
|
203
|
-
|
|
204
|
-
Returns:
|
|
205
|
-
URL of the video thumbnail (high quality)
|
|
206
|
-
"""
|
|
207
|
-
return f'https://i.ytimg.com/vi/{self._matched_id}/hqdefault.jpg'
|
|
208
|
-
|
|
209
|
-
@property
|
|
210
|
-
def thumbnail_urls(self) -> Dict[str, str]:
|
|
211
|
-
"""
|
|
212
|
-
Get all thumbnail URLs for this video in different qualities
|
|
213
|
-
|
|
214
|
-
Returns:
|
|
215
|
-
Dictionary of thumbnail URLs with quality labels
|
|
216
|
-
"""
|
|
217
|
-
return {
|
|
218
|
-
'default': f'https://i.ytimg.com/vi/{self._matched_id}/default.jpg',
|
|
219
|
-
'medium': f'https://i.ytimg.com/vi/{self._matched_id}/mqdefault.jpg',
|
|
220
|
-
'high': f'https://i.ytimg.com/vi/{self._matched_id}/hqdefault.jpg',
|
|
221
|
-
'standard': f'https://i.ytimg.com/vi/{self._matched_id}/sddefault.jpg',
|
|
222
|
-
'maxres': f'https://i.ytimg.com/vi/{self._matched_id}/maxresdefault.jpg'
|
|
223
|
-
}
|
|
224
|
-
|
|
225
|
-
if __name__ == '__main__':
|
|
226
|
-
video = Video('https://www.youtube.com/watch?v=9bZkp7q19f0')
|
|
227
|
-
print(video.metadata)
|
|
228
|
-
|
|
229
|
-
# Example of getting comments
|
|
230
|
-
print("\nFirst 3 comments:")
|
|
231
|
-
for i, comment in enumerate(video.stream_comments(3), 1):
|
|
232
|
-
print(f"{i}. {comment['author']}: {comment['text'][:50]}...")
|
|
1
|
+
import re
|
|
2
|
+
import json
|
|
3
|
+
from typing import Dict, Any
|
|
4
|
+
from .https import video_data
|
|
5
|
+
|
|
6
|
+
|
|
7
|
+
class Video:
|
|
8
|
+
|
|
9
|
+
_HEAD = 'https://www.youtube.com/watch?v='
|
|
10
|
+
|
|
11
|
+
def __init__(self, video_id: str):
|
|
12
|
+
"""
|
|
13
|
+
Represents a YouTube video
|
|
14
|
+
|
|
15
|
+
Parameters
|
|
16
|
+
----------
|
|
17
|
+
video_id : str
|
|
18
|
+
The id or url of the video
|
|
19
|
+
"""
|
|
20
|
+
pattern = re.compile('.be/(.*?)$|=(.*?)$|^(\w{11})$') # noqa
|
|
21
|
+
match = pattern.search(video_id)
|
|
22
|
+
|
|
23
|
+
if not match:
|
|
24
|
+
raise ValueError('Invalid YouTube video ID or URL')
|
|
25
|
+
|
|
26
|
+
self._matched_id = (
|
|
27
|
+
match.group(1)
|
|
28
|
+
or match.group(2)
|
|
29
|
+
or match.group(3)
|
|
30
|
+
)
|
|
31
|
+
|
|
32
|
+
if self._matched_id:
|
|
33
|
+
self._url = self._HEAD + self._matched_id
|
|
34
|
+
self._video_data = video_data(self._matched_id)
|
|
35
|
+
# Extract basic info for fallback
|
|
36
|
+
title_match = re.search('<title>(.*?) - YouTube</title>', self._video_data)
|
|
37
|
+
self.title = title_match.group(1) if title_match else None
|
|
38
|
+
self.id = self._matched_id
|
|
39
|
+
else:
|
|
40
|
+
raise ValueError('Invalid YouTube video ID or URL')
|
|
41
|
+
|
|
42
|
+
def __repr__(self):
|
|
43
|
+
return f'<Video {self._url}>'
|
|
44
|
+
|
|
45
|
+
@property
|
|
46
|
+
def metadata(self) -> Dict[str, Any]:
|
|
47
|
+
"""
|
|
48
|
+
Fetches video metadata in a dict format
|
|
49
|
+
|
|
50
|
+
Returns
|
|
51
|
+
-------
|
|
52
|
+
Dict
|
|
53
|
+
Video metadata in a dict format containing keys: title, id, views, duration, author_id,
|
|
54
|
+
upload_date, url, thumbnails, tags, description, likes, genre, etc.
|
|
55
|
+
"""
|
|
56
|
+
# Multiple patterns to try for video details extraction for robustness
|
|
57
|
+
details_patterns = [
|
|
58
|
+
re.compile('videoDetails\":(.*?)\"isLiveContent\":.*?}'),
|
|
59
|
+
re.compile('videoDetails\":(.*?),\"playerConfig'),
|
|
60
|
+
re.compile('videoDetails\":(.*?),\"playabilityStatus')
|
|
61
|
+
]
|
|
62
|
+
|
|
63
|
+
# Other metadata patterns
|
|
64
|
+
upload_date_pattern = re.compile("<meta itemprop=\"uploadDate\" content=\"(.*?)\">")
|
|
65
|
+
genre_pattern = re.compile("<meta itemprop=\"genre\" content=\"(.*?)\">")
|
|
66
|
+
like_count_patterns = [
|
|
67
|
+
re.compile("iconType\":\"LIKE\"},\"defaultText\":(.*?)}"),
|
|
68
|
+
re.compile('\"likeCount\":\"(\\d+)\"')
|
|
69
|
+
]
|
|
70
|
+
channel_name_pattern = re.compile('"ownerChannelName":"(.*?)"')
|
|
71
|
+
|
|
72
|
+
# Try each pattern for video details
|
|
73
|
+
raw_details_match = None
|
|
74
|
+
for pattern in details_patterns:
|
|
75
|
+
match = pattern.search(self._video_data)
|
|
76
|
+
if match:
|
|
77
|
+
raw_details_match = match
|
|
78
|
+
break
|
|
79
|
+
|
|
80
|
+
if not raw_details_match:
|
|
81
|
+
# Fallback metadata for search results or incomplete video data
|
|
82
|
+
return {
|
|
83
|
+
'title': getattr(self, 'title', None),
|
|
84
|
+
'id': getattr(self, 'id', None),
|
|
85
|
+
'views': getattr(self, 'views', None),
|
|
86
|
+
'streamed': False,
|
|
87
|
+
'duration': None,
|
|
88
|
+
'author_id': None,
|
|
89
|
+
'author_name': None,
|
|
90
|
+
'upload_date': None,
|
|
91
|
+
'url': f"https://www.youtube.com/watch?v={getattr(self, 'id', '')}" if hasattr(self, 'id') else None,
|
|
92
|
+
'thumbnails': None,
|
|
93
|
+
'tags': None,
|
|
94
|
+
'description': None,
|
|
95
|
+
'likes': None,
|
|
96
|
+
'genre': None,
|
|
97
|
+
'is_age_restricted': 'age-restricted' in self._video_data.lower(),
|
|
98
|
+
'is_unlisted': 'unlisted' in self._video_data.lower()
|
|
99
|
+
}
|
|
100
|
+
|
|
101
|
+
raw_details = raw_details_match.group(0)
|
|
102
|
+
|
|
103
|
+
# Extract upload date
|
|
104
|
+
upload_date_match = upload_date_pattern.search(self._video_data)
|
|
105
|
+
upload_date = upload_date_match.group(1) if upload_date_match else None
|
|
106
|
+
|
|
107
|
+
# Extract channel name
|
|
108
|
+
channel_name_match = channel_name_pattern.search(self._video_data)
|
|
109
|
+
channel_name = channel_name_match.group(1) if channel_name_match else None
|
|
110
|
+
|
|
111
|
+
# Parse video details
|
|
112
|
+
try:
|
|
113
|
+
# Clean up the JSON string for parsing
|
|
114
|
+
clean_json = raw_details.replace('videoDetails\":', '')
|
|
115
|
+
# Handle potential JSON parsing issues
|
|
116
|
+
if clean_json.endswith(','):
|
|
117
|
+
clean_json = clean_json[:-1]
|
|
118
|
+
metadata = json.loads(clean_json)
|
|
119
|
+
|
|
120
|
+
data = {
|
|
121
|
+
'title': metadata.get('title'),
|
|
122
|
+
'id': metadata.get('videoId', self._matched_id),
|
|
123
|
+
'views': metadata.get('viewCount'),
|
|
124
|
+
'streamed': metadata.get('isLiveContent', False),
|
|
125
|
+
'duration': metadata.get('lengthSeconds'),
|
|
126
|
+
'author_id': metadata.get('channelId'),
|
|
127
|
+
'author_name': channel_name or metadata.get('author'),
|
|
128
|
+
'upload_date': upload_date,
|
|
129
|
+
'url': f"https://www.youtube.com/watch?v={metadata.get('videoId', self._matched_id)}",
|
|
130
|
+
'thumbnails': metadata.get('thumbnail', {}).get('thumbnails'),
|
|
131
|
+
'tags': metadata.get('keywords'),
|
|
132
|
+
'description': metadata.get('shortDescription'),
|
|
133
|
+
'is_age_restricted': metadata.get('isAgeRestricted', False) or 'age-restricted' in self._video_data.lower(),
|
|
134
|
+
'is_unlisted': 'unlisted' in self._video_data.lower(),
|
|
135
|
+
'is_family_safe': metadata.get('isFamilySafe', True),
|
|
136
|
+
'is_private': metadata.get('isPrivate', False),
|
|
137
|
+
'is_live_content': metadata.get('isLiveContent', False),
|
|
138
|
+
'is_crawlable': metadata.get('isCrawlable', True),
|
|
139
|
+
'allow_ratings': metadata.get('allowRatings', True)
|
|
140
|
+
}
|
|
141
|
+
except (json.JSONDecodeError, KeyError, TypeError) as e:
|
|
142
|
+
# Fallback to basic metadata if JSON parsing fails
|
|
143
|
+
return {
|
|
144
|
+
'title': getattr(self, 'title', None),
|
|
145
|
+
'id': self._matched_id,
|
|
146
|
+
'url': self._url,
|
|
147
|
+
'error': f"Failed to parse video details: {str(e)}"
|
|
148
|
+
}
|
|
149
|
+
|
|
150
|
+
# Try to extract likes count
|
|
151
|
+
likes = None
|
|
152
|
+
for pattern in like_count_patterns:
|
|
153
|
+
try:
|
|
154
|
+
likes_match = pattern.search(self._video_data)
|
|
155
|
+
if likes_match:
|
|
156
|
+
likes_text = likes_match.group(1)
|
|
157
|
+
# Handle different formats of like count
|
|
158
|
+
if '{' in likes_text:
|
|
159
|
+
likes = json.loads(likes_text + '}}}')['accessibility']['accessibilityData']['label'].split(' ')[0].replace(',', '')
|
|
160
|
+
else:
|
|
161
|
+
likes = likes_text
|
|
162
|
+
break
|
|
163
|
+
except (AttributeError, KeyError, json.decoder.JSONDecodeError):
|
|
164
|
+
continue
|
|
165
|
+
|
|
166
|
+
data['likes'] = likes
|
|
167
|
+
|
|
168
|
+
# Try to extract genre
|
|
169
|
+
try:
|
|
170
|
+
genre_match = genre_pattern.search(self._video_data)
|
|
171
|
+
data['genre'] = genre_match.group(1) if genre_match else None
|
|
172
|
+
except AttributeError:
|
|
173
|
+
data['genre'] = None
|
|
174
|
+
|
|
175
|
+
return data
|
|
176
|
+
|
|
177
|
+
|
|
178
|
+
|
|
179
|
+
@property
|
|
180
|
+
def embed_html(self) -> str:
|
|
181
|
+
"""
|
|
182
|
+
Get the embed HTML code for this video
|
|
183
|
+
|
|
184
|
+
Returns:
|
|
185
|
+
HTML iframe code for embedding the video
|
|
186
|
+
"""
|
|
187
|
+
return f'<iframe width="560" height="315" src="https://www.youtube.com/embed/{self._matched_id}" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture" allowfullscreen></iframe>'
|
|
188
|
+
|
|
189
|
+
@property
|
|
190
|
+
def embed_url(self) -> str:
|
|
191
|
+
"""
|
|
192
|
+
Get the embed URL for this video
|
|
193
|
+
|
|
194
|
+
Returns:
|
|
195
|
+
URL for embedding the video
|
|
196
|
+
"""
|
|
197
|
+
return f'https://www.youtube.com/embed/{self._matched_id}'
|
|
198
|
+
|
|
199
|
+
@property
|
|
200
|
+
def thumbnail_url(self) -> str:
|
|
201
|
+
"""
|
|
202
|
+
Get the thumbnail URL for this video
|
|
203
|
+
|
|
204
|
+
Returns:
|
|
205
|
+
URL of the video thumbnail (high quality)
|
|
206
|
+
"""
|
|
207
|
+
return f'https://i.ytimg.com/vi/{self._matched_id}/hqdefault.jpg'
|
|
208
|
+
|
|
209
|
+
@property
|
|
210
|
+
def thumbnail_urls(self) -> Dict[str, str]:
|
|
211
|
+
"""
|
|
212
|
+
Get all thumbnail URLs for this video in different qualities
|
|
213
|
+
|
|
214
|
+
Returns:
|
|
215
|
+
Dictionary of thumbnail URLs with quality labels
|
|
216
|
+
"""
|
|
217
|
+
return {
|
|
218
|
+
'default': f'https://i.ytimg.com/vi/{self._matched_id}/default.jpg',
|
|
219
|
+
'medium': f'https://i.ytimg.com/vi/{self._matched_id}/mqdefault.jpg',
|
|
220
|
+
'high': f'https://i.ytimg.com/vi/{self._matched_id}/hqdefault.jpg',
|
|
221
|
+
'standard': f'https://i.ytimg.com/vi/{self._matched_id}/sddefault.jpg',
|
|
222
|
+
'maxres': f'https://i.ytimg.com/vi/{self._matched_id}/maxresdefault.jpg'
|
|
223
|
+
}
|
|
224
|
+
|
|
225
|
+
if __name__ == '__main__':
|
|
226
|
+
video = Video('https://www.youtube.com/watch?v=9bZkp7q19f0')
|
|
227
|
+
print(video.metadata)
|
|
228
|
+
|
|
229
|
+
# Example of getting comments
|
|
230
|
+
print("\nFirst 3 comments:")
|
|
231
|
+
for i, comment in enumerate(video.stream_comments(3), 1):
|
|
232
|
+
print(f"{i}. {comment['author']}: {comment['text'][:50]}...")
|
|
@@ -0,0 +1,10 @@
|
|
|
1
|
+
# LitLogger
|
|
2
|
+
|
|
3
|
+
A minimal yet flexible logging library built from scratch without external dependencies. It provides colored console output, file logging with rotation, simple network logging, and optional asynchronous support.
|
|
4
|
+
|
|
5
|
+
```python
|
|
6
|
+
from webscout.litlogger import Logger, LogLevel, FileHandler
|
|
7
|
+
|
|
8
|
+
logger = Logger(name="demo", level=LogLevel.DEBUG, handlers=[FileHandler("app.log")])
|
|
9
|
+
logger.info("hello world")
|
|
10
|
+
```
|
webscout/Litlogger/__init__.py
CHANGED
|
@@ -1,67 +1,15 @@
|
|
|
1
|
-
"""
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
- Multiple output formats including JSON
|
|
7
|
-
- File logging with rotation
|
|
8
|
-
- Network logging (HTTP/HTTPS/TCP)
|
|
9
|
-
- Async logging support
|
|
10
|
-
- Intelligent log level detection
|
|
11
|
-
- Context managers
|
|
12
|
-
- Performance metrics
|
|
13
|
-
- Log aggregation
|
|
14
|
-
"""
|
|
15
|
-
|
|
16
|
-
from .core.logger import Logger
|
|
17
|
-
from .core.level import LogLevel
|
|
18
|
-
from .styles.colors import LogColors
|
|
19
|
-
from .styles.formats import LogFormat
|
|
20
|
-
from .styles.text import TextStyle
|
|
21
|
-
from .handlers.console import ConsoleHandler, ErrorConsoleHandler
|
|
22
|
-
from .handlers.file import FileHandler
|
|
23
|
-
from .handlers.network import NetworkHandler
|
|
24
|
-
from .utils.detectors import LevelDetector
|
|
25
|
-
from .utils.formatters import MessageFormatter
|
|
26
|
-
|
|
27
|
-
# Create a default logger instance
|
|
28
|
-
default_logger = Logger(
|
|
29
|
-
name="LitLogger",
|
|
30
|
-
handlers=[ConsoleHandler()]
|
|
31
|
-
)
|
|
32
|
-
|
|
33
|
-
# Expose common logging methods at package level
|
|
34
|
-
debug = default_logger.debug
|
|
35
|
-
info = default_logger.info
|
|
36
|
-
warning = default_logger.warning
|
|
37
|
-
error = default_logger.error
|
|
38
|
-
critical = default_logger.critical
|
|
1
|
+
"""Lightweight logger built from scratch."""
|
|
2
|
+
from .levels import LogLevel
|
|
3
|
+
from .handlers import ConsoleHandler, FileHandler, NetworkHandler, TCPHandler
|
|
4
|
+
from .logger import Logger
|
|
5
|
+
from .formats import LogFormat
|
|
39
6
|
|
|
40
7
|
__all__ = [
|
|
41
|
-
# Core
|
|
42
8
|
"Logger",
|
|
43
9
|
"LogLevel",
|
|
44
|
-
|
|
45
|
-
# Styles
|
|
46
|
-
"LogColors",
|
|
47
|
-
"LogFormat",
|
|
48
|
-
"TextStyle",
|
|
49
|
-
|
|
50
|
-
# Handlers
|
|
51
10
|
"ConsoleHandler",
|
|
52
|
-
"ErrorConsoleHandler",
|
|
53
11
|
"FileHandler",
|
|
54
12
|
"NetworkHandler",
|
|
55
|
-
|
|
56
|
-
|
|
57
|
-
"LevelDetector",
|
|
58
|
-
"MessageFormatter",
|
|
59
|
-
|
|
60
|
-
# Package-level logging functions
|
|
61
|
-
"debug",
|
|
62
|
-
"info",
|
|
63
|
-
"warning",
|
|
64
|
-
"error",
|
|
65
|
-
"critical",
|
|
66
|
-
|
|
13
|
+
"TCPHandler",
|
|
14
|
+
"LogFormat",
|
|
67
15
|
]
|