KekikStream 1.8.1__py3-none-any.whl → 1.9.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- KekikStream/Core/Extractor/ExtractorBase.py +2 -14
- KekikStream/Core/Media/MediaHandler.py +30 -8
- KekikStream/Core/Plugin/PluginBase.py +2 -15
- KekikStream/Extractors/YTDLP.py +109 -0
- KekikStream/Plugins/DiziBox.py +2 -7
- KekikStream/Plugins/DiziPal.py +1 -1
- KekikStream/Plugins/DiziYou.py +41 -16
- KekikStream/Plugins/Dizilla.py +38 -27
- KekikStream/Plugins/FilmBip.py +1 -1
- KekikStream/Plugins/FilmMakinesi.py +1 -5
- KekikStream/Plugins/FullHDFilmizlesene.py +1 -5
- KekikStream/Plugins/HDFilmCehennemi.py +11 -18
- KekikStream/Plugins/JetFilmizle.py +1 -5
- KekikStream/Plugins/RecTV.py +1 -5
- KekikStream/Plugins/RoketDizi.py +80 -42
- KekikStream/Plugins/SelcukFlix.py +160 -67
- KekikStream/Plugins/SezonlukDizi.py +1 -5
- KekikStream/Plugins/SineWix.py +1 -5
- KekikStream/Plugins/Sinefy.py +72 -51
- KekikStream/Plugins/Sinezy.py +74 -42
- KekikStream/Plugins/UgurFilm.py +1 -5
- KekikStream/requirements.txt +2 -3
- kekikstream-1.9.0.dist-info/METADATA +290 -0
- {kekikstream-1.8.1.dist-info → kekikstream-1.9.0.dist-info}/RECORD +28 -27
- kekikstream-1.8.1.dist-info/METADATA +0 -110
- {kekikstream-1.8.1.dist-info → kekikstream-1.9.0.dist-info}/WHEEL +0 -0
- {kekikstream-1.8.1.dist-info → kekikstream-1.9.0.dist-info}/entry_points.txt +0 -0
- {kekikstream-1.8.1.dist-info → kekikstream-1.9.0.dist-info}/licenses/LICENSE +0 -0
- {kekikstream-1.8.1.dist-info → kekikstream-1.9.0.dist-info}/top_level.txt +0 -0
|
@@ -1,7 +1,6 @@
|
|
|
1
1
|
# Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
|
|
2
2
|
|
|
3
3
|
from abc import ABC, abstractmethod
|
|
4
|
-
from curl_cffi import AsyncSession
|
|
5
4
|
from cloudscraper import CloudScraper
|
|
6
5
|
from httpx import AsyncClient
|
|
7
6
|
from typing import Optional
|
|
@@ -12,7 +11,6 @@ class ExtractorBase(ABC):
|
|
|
12
11
|
# Çıkarıcının temel özellikleri
|
|
13
12
|
name = "Extractor"
|
|
14
13
|
main_url = ""
|
|
15
|
-
requires_cffi = False
|
|
16
14
|
|
|
17
15
|
def __init__(self):
|
|
18
16
|
# cloudscraper - for bypassing Cloudflare
|
|
@@ -21,19 +19,11 @@ class ExtractorBase(ABC):
|
|
|
21
19
|
# httpx - lightweight and safe for most HTTP requests
|
|
22
20
|
self.httpx = AsyncClient(
|
|
23
21
|
timeout = 3,
|
|
24
|
-
follow_redirects = True
|
|
22
|
+
follow_redirects = True
|
|
25
23
|
)
|
|
26
24
|
self.httpx.headers.update(self.cloudscraper.headers)
|
|
27
25
|
self.httpx.cookies.update(self.cloudscraper.cookies)
|
|
28
26
|
|
|
29
|
-
# curl_cffi - only initialize if needed for anti-bot bypass
|
|
30
|
-
self.cffi = None
|
|
31
|
-
|
|
32
|
-
if self.requires_cffi:
|
|
33
|
-
self.cffi = AsyncSession(impersonate="firefox135")
|
|
34
|
-
self.cffi.cookies.update(self.cloudscraper.cookies)
|
|
35
|
-
self.cffi.headers.update({"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 15.7; rv:135.0) Gecko/20100101 Firefox/135.0"})
|
|
36
|
-
|
|
37
27
|
def can_handle_url(self, url: str) -> bool:
|
|
38
28
|
# URL'nin bu çıkarıcı tarafından işlenip işlenemeyeceğini kontrol et
|
|
39
29
|
return self.main_url in url
|
|
@@ -44,10 +34,8 @@ class ExtractorBase(ABC):
|
|
|
44
34
|
pass
|
|
45
35
|
|
|
46
36
|
async def close(self):
|
|
47
|
-
"""Close
|
|
37
|
+
"""Close HTTP client."""
|
|
48
38
|
await self.httpx.aclose()
|
|
49
|
-
if self.cffi:
|
|
50
|
-
await self.cffi.close()
|
|
51
39
|
|
|
52
40
|
def fix_url(self, url: str) -> str:
|
|
53
41
|
# Eksik URL'leri düzelt ve tam URL formatına çevir
|
|
@@ -18,19 +18,34 @@ class MediaHandler:
|
|
|
18
18
|
if extract_data.referer:
|
|
19
19
|
self.headers["referer"] = extract_data.referer
|
|
20
20
|
|
|
21
|
-
#
|
|
21
|
+
# Özel Durumlar (RecTV vs. Googleusercontent)
|
|
22
22
|
if user_agent in ["googleusercontent", "Mozilla/5.0 (X11; Linux x86_64; rv:101.0) Gecko/20100101 Firefox/101.0"]:
|
|
23
23
|
return self.play_with_ytdlp(extract_data)
|
|
24
24
|
|
|
25
|
-
# İşletim sistemine göre oynatıcı seç
|
|
25
|
+
# İşletim sistemine göre oynatıcı seç (Android durumu)
|
|
26
26
|
if subprocess.check_output(['uname', '-o']).strip() == b'Android':
|
|
27
27
|
return self.play_with_android_mxplayer(extract_data)
|
|
28
28
|
|
|
29
|
-
#
|
|
30
|
-
|
|
31
|
-
|
|
29
|
+
# Oynatıcı öncelik sırası (fallback zincirleme)
|
|
30
|
+
players = [
|
|
31
|
+
("MPV", self.play_with_mpv),
|
|
32
|
+
("VLC", self.play_with_vlc),
|
|
33
|
+
("yt-dlp", self.play_with_ytdlp)
|
|
34
|
+
]
|
|
35
|
+
|
|
36
|
+
# Fallback zincirleme
|
|
37
|
+
for player_name, player_func in players:
|
|
38
|
+
try:
|
|
39
|
+
result = player_func(extract_data)
|
|
40
|
+
if result or result is None: # None = MPV (exception yok)
|
|
41
|
+
konsol.log(f"[green][✓] {player_name} ile başarılı[/green]")
|
|
42
|
+
return True
|
|
43
|
+
except Exception as e:
|
|
44
|
+
konsol.log(f"[yellow][⚠] {player_name} hatası: {e}[/yellow]")
|
|
45
|
+
continue
|
|
32
46
|
|
|
33
|
-
|
|
47
|
+
konsol.print("[red][✗] Hiçbir oynatıcı çalışmadı![/red]")
|
|
48
|
+
return False
|
|
34
49
|
|
|
35
50
|
def play_with_vlc(self, extract_data: ExtractResult):
|
|
36
51
|
konsol.log(f"[yellow][»] VLC ile Oynatılıyor : {extract_data.url}")
|
|
@@ -88,12 +103,15 @@ class MediaHandler:
|
|
|
88
103
|
with open(os.devnull, "w") as devnull:
|
|
89
104
|
subprocess.run(mpv_command, stdout=devnull, stderr=devnull, check=True)
|
|
90
105
|
|
|
106
|
+
return True
|
|
91
107
|
except subprocess.CalledProcessError as hata:
|
|
92
108
|
konsol.print(f"[red]mpv oynatma hatası: {hata}[/red]")
|
|
93
109
|
konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
|
|
110
|
+
return False
|
|
94
111
|
except FileNotFoundError:
|
|
95
112
|
konsol.print("[red]mpv bulunamadı! mpv kurulu olduğundan emin olun.[/red]")
|
|
96
113
|
konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
|
|
114
|
+
return False
|
|
97
115
|
|
|
98
116
|
def play_with_ytdlp(self, extract_data: ExtractResult):
|
|
99
117
|
konsol.log(f"[yellow][»] yt-dlp ile Oynatılıyor : {extract_data.url}")
|
|
@@ -121,12 +139,15 @@ class MediaHandler:
|
|
|
121
139
|
with subprocess.Popen(ytdlp_command, stdout=subprocess.PIPE) as ytdlp_proc:
|
|
122
140
|
subprocess.run(mpv_command, stdin=ytdlp_proc.stdout, check=True)
|
|
123
141
|
|
|
142
|
+
return True
|
|
124
143
|
except subprocess.CalledProcessError as hata:
|
|
125
144
|
konsol.print(f"[red]Oynatma hatası: {hata}[/red]")
|
|
126
145
|
konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
|
|
146
|
+
return False
|
|
127
147
|
except FileNotFoundError:
|
|
128
148
|
konsol.print("[red]yt-dlp veya mpv bulunamadı! Kurulumlarından emin olun.[/red]")
|
|
129
149
|
konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
|
|
150
|
+
return False
|
|
130
151
|
|
|
131
152
|
def play_with_android_mxplayer(self, extract_data: ExtractResult):
|
|
132
153
|
konsol.log(f"[yellow][»] MxPlayer ile Oynatılıyor : {extract_data.url}")
|
|
@@ -151,11 +172,12 @@ class MediaHandler:
|
|
|
151
172
|
with open(os.devnull, "w") as devnull:
|
|
152
173
|
subprocess.run(android_command, stdout=devnull, stderr=devnull, check=True)
|
|
153
174
|
|
|
154
|
-
return
|
|
155
|
-
|
|
175
|
+
return True
|
|
156
176
|
except subprocess.CalledProcessError as hata:
|
|
157
177
|
konsol.print(f"[red]{paket} oynatma hatası: {hata}[/red]")
|
|
158
178
|
konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
|
|
179
|
+
return False
|
|
159
180
|
except FileNotFoundError:
|
|
160
181
|
konsol.print(f"Paket: {paket}, Hata: MX Player kurulu değil")
|
|
161
182
|
konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
|
|
183
|
+
return False
|
|
@@ -1,7 +1,6 @@
|
|
|
1
1
|
# Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
|
|
2
2
|
|
|
3
3
|
from abc import ABC, abstractmethod
|
|
4
|
-
from curl_cffi import AsyncSession
|
|
5
4
|
from cloudscraper import CloudScraper
|
|
6
5
|
from httpx import AsyncClient
|
|
7
6
|
from .PluginModels import MainPageResult, SearchResult, MovieInfo
|
|
@@ -17,8 +16,6 @@ class PluginBase(ABC):
|
|
|
17
16
|
favicon = f"https://www.google.com/s2/favicons?domain={main_url}&sz=64"
|
|
18
17
|
description = "No description provided."
|
|
19
18
|
|
|
20
|
-
requires_cffi = False
|
|
21
|
-
|
|
22
19
|
main_page = {}
|
|
23
20
|
|
|
24
21
|
async def url_update(self, new_url: str):
|
|
@@ -33,19 +30,11 @@ class PluginBase(ABC):
|
|
|
33
30
|
# httpx - lightweight and safe for most HTTP requests
|
|
34
31
|
self.httpx = AsyncClient(
|
|
35
32
|
timeout = 3,
|
|
36
|
-
follow_redirects = True
|
|
33
|
+
follow_redirects = True
|
|
37
34
|
)
|
|
38
35
|
self.httpx.headers.update(self.cloudscraper.headers)
|
|
39
36
|
self.httpx.cookies.update(self.cloudscraper.cookies)
|
|
40
37
|
|
|
41
|
-
# curl_cffi - only initialize if needed for anti-bot bypass
|
|
42
|
-
self.cffi = None
|
|
43
|
-
|
|
44
|
-
if self.requires_cffi:
|
|
45
|
-
self.cffi = AsyncSession(impersonate="firefox135")
|
|
46
|
-
self.cffi.cookies.update(self.cloudscraper.cookies)
|
|
47
|
-
self.cffi.headers.update({"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 15.7; rv:135.0) Gecko/20100101 Firefox/135.0"})
|
|
48
|
-
|
|
49
38
|
self.media_handler = MediaHandler()
|
|
50
39
|
self.ex_manager = ExtractorManager()
|
|
51
40
|
|
|
@@ -90,10 +79,8 @@ class PluginBase(ABC):
|
|
|
90
79
|
pass
|
|
91
80
|
|
|
92
81
|
async def close(self):
|
|
93
|
-
"""Close
|
|
82
|
+
"""Close HTTP client."""
|
|
94
83
|
await self.httpx.aclose()
|
|
95
|
-
if self.cffi:
|
|
96
|
-
await self.cffi.close()
|
|
97
84
|
|
|
98
85
|
def fix_url(self, url: str) -> str:
|
|
99
86
|
if not url:
|
|
@@ -0,0 +1,109 @@
|
|
|
1
|
+
# Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
|
|
2
|
+
|
|
3
|
+
from KekikStream.Core import ExtractorBase, ExtractResult, Subtitle
|
|
4
|
+
import yt_dlp, sys, os
|
|
5
|
+
|
|
6
|
+
class YTDLP(ExtractorBase):
|
|
7
|
+
name = "yt-dlp"
|
|
8
|
+
main_url = "" # Universal - tüm siteleri destekler
|
|
9
|
+
|
|
10
|
+
def __init__(self):
|
|
11
|
+
pass
|
|
12
|
+
|
|
13
|
+
def can_handle_url(self, url: str) -> bool:
|
|
14
|
+
"""
|
|
15
|
+
yt-dlp'nin bu URL'yi işleyip işleyemeyeceğini kontrol et
|
|
16
|
+
"""
|
|
17
|
+
try:
|
|
18
|
+
# stderr'ı geçici olarak kapat (hata mesajlarını gizle)
|
|
19
|
+
old_stderr = sys.stderr
|
|
20
|
+
sys.stderr = open(os.devnull, "w")
|
|
21
|
+
|
|
22
|
+
try:
|
|
23
|
+
ydl_opts = {
|
|
24
|
+
"simulate" : True, # Download yok, sadece tespit
|
|
25
|
+
"quiet" : True, # Log kirliliği yok
|
|
26
|
+
"no_warnings" : True, # Uyarı mesajları yok
|
|
27
|
+
"extract_flat" : True, # Minimal işlem
|
|
28
|
+
"no_check_certificates" : True,
|
|
29
|
+
"ignoreerrors" : True # Hataları yoksay
|
|
30
|
+
}
|
|
31
|
+
|
|
32
|
+
with yt_dlp.YoutubeDL(ydl_opts) as ydl:
|
|
33
|
+
# URL'yi işleyebiliyor mu kontrol et
|
|
34
|
+
info = ydl.extract_info(url, download=False, process=False)
|
|
35
|
+
|
|
36
|
+
# Generic extractor ise atla
|
|
37
|
+
if info and info.get("extractor_key") != "Generic":
|
|
38
|
+
return True
|
|
39
|
+
|
|
40
|
+
return False
|
|
41
|
+
finally:
|
|
42
|
+
# stderr'ı geri yükle
|
|
43
|
+
sys.stderr.close()
|
|
44
|
+
sys.stderr = old_stderr
|
|
45
|
+
|
|
46
|
+
except Exception:
|
|
47
|
+
# yt-dlp işleyemezse False döndür
|
|
48
|
+
return False
|
|
49
|
+
|
|
50
|
+
async def extract(self, url: str, referer: str | None = None) -> ExtractResult:
|
|
51
|
+
ydl_opts = {
|
|
52
|
+
"quiet" : True,
|
|
53
|
+
"no_warnings" : True,
|
|
54
|
+
"extract_flat" : False, # Tam bilgi al
|
|
55
|
+
"format" : "best", # En iyi kalite
|
|
56
|
+
"no_check_certificates" : True
|
|
57
|
+
}
|
|
58
|
+
|
|
59
|
+
# Referer varsa header olarak ekle
|
|
60
|
+
if referer:
|
|
61
|
+
ydl_opts["http_headers"] = {"Referer": referer}
|
|
62
|
+
|
|
63
|
+
with yt_dlp.YoutubeDL(ydl_opts) as ydl:
|
|
64
|
+
info = ydl.extract_info(url, download=False)
|
|
65
|
+
|
|
66
|
+
if not info:
|
|
67
|
+
raise ValueError("yt-dlp video bilgisi döndürmedi")
|
|
68
|
+
|
|
69
|
+
# Video URL'sini al
|
|
70
|
+
video_url = info.get("url")
|
|
71
|
+
if not video_url:
|
|
72
|
+
# Bazen formatlar listesinde olabilir
|
|
73
|
+
formats = info.get("formats", [])
|
|
74
|
+
if formats:
|
|
75
|
+
video_url = formats[-1].get("url") # Son format (genellikle en iyi)
|
|
76
|
+
|
|
77
|
+
if not video_url:
|
|
78
|
+
raise ValueError("Video URL bulunamadı")
|
|
79
|
+
|
|
80
|
+
# Altyazıları çıkar
|
|
81
|
+
subtitles = []
|
|
82
|
+
if subtitle_data := info.get("subtitles"):
|
|
83
|
+
for lang, subs in subtitle_data.items():
|
|
84
|
+
for sub in subs:
|
|
85
|
+
if sub_url := sub.get("url"):
|
|
86
|
+
subtitles.append(
|
|
87
|
+
Subtitle(
|
|
88
|
+
name=f"{lang} ({sub.get('ext', 'unknown')})",
|
|
89
|
+
url=sub_url
|
|
90
|
+
)
|
|
91
|
+
)
|
|
92
|
+
|
|
93
|
+
# User-Agent al
|
|
94
|
+
user_agent = None
|
|
95
|
+
http_headers = info.get("http_headers", {})
|
|
96
|
+
if http_headers:
|
|
97
|
+
user_agent = http_headers.get("User-Agent")
|
|
98
|
+
|
|
99
|
+
return ExtractResult(
|
|
100
|
+
name = self.name,
|
|
101
|
+
url = video_url,
|
|
102
|
+
referer = referer or info.get("webpage_url"),
|
|
103
|
+
user_agent = user_agent,
|
|
104
|
+
subtitles = subtitles
|
|
105
|
+
)
|
|
106
|
+
|
|
107
|
+
async def close(self):
|
|
108
|
+
"""yt-dlp için cleanup gerekmez"""
|
|
109
|
+
pass
|
KekikStream/Plugins/DiziBox.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
|
|
2
2
|
|
|
3
|
-
from KekikStream.Core import
|
|
3
|
+
from KekikStream.Core import PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode
|
|
4
4
|
from Kekik.Sifreleme import CryptoJS
|
|
5
5
|
from parsel import Selector
|
|
6
6
|
import re, urllib.parse, base64, contextlib, asyncio, time
|
|
@@ -40,7 +40,6 @@ class DiziBox(PluginBase):
|
|
|
40
40
|
f"{main_url}/dizi-arsivi/page/SAYFA/?tur[0]=yarisma&yil&imdb" : "Yarışma"
|
|
41
41
|
}
|
|
42
42
|
|
|
43
|
-
#@kekik_cache(ttl=60*60)
|
|
44
43
|
async def get_main_page(self, page: int, url: str, category: str) -> list[MainPageResult]:
|
|
45
44
|
self.httpx.cookies.update({
|
|
46
45
|
"isTrustedUser" : "true",
|
|
@@ -62,7 +61,6 @@ class DiziBox(PluginBase):
|
|
|
62
61
|
for veri in secici.css("article.detailed-article")
|
|
63
62
|
]
|
|
64
63
|
|
|
65
|
-
#@kekik_cache(ttl=60*60)
|
|
66
64
|
async def search(self, query: str) -> list[SearchResult]:
|
|
67
65
|
self.httpx.cookies.update({
|
|
68
66
|
"isTrustedUser" : "true",
|
|
@@ -80,7 +78,6 @@ class DiziBox(PluginBase):
|
|
|
80
78
|
for item in secici.css("article.detailed-article")
|
|
81
79
|
]
|
|
82
80
|
|
|
83
|
-
#@kekik_cache(ttl=60*60)
|
|
84
81
|
async def load_item(self, url: str) -> SeriesInfo:
|
|
85
82
|
istek = await self.httpx.get(url)
|
|
86
83
|
secici = Selector(istek.text)
|
|
@@ -127,7 +124,6 @@ class DiziBox(PluginBase):
|
|
|
127
124
|
actors = actors,
|
|
128
125
|
)
|
|
129
126
|
|
|
130
|
-
#@kekik_cache(ttl=60*60)
|
|
131
127
|
async def _iframe_decode(self, name:str, iframe_link:str, referer:str) -> list[str]:
|
|
132
128
|
results = []
|
|
133
129
|
|
|
@@ -178,7 +174,6 @@ class DiziBox(PluginBase):
|
|
|
178
174
|
|
|
179
175
|
return results
|
|
180
176
|
|
|
181
|
-
#@kekik_cache(ttl=15*60)
|
|
182
177
|
async def load_links(self, url: str) -> list[dict]:
|
|
183
178
|
istek = await self.httpx.get(url)
|
|
184
179
|
secici = Selector(istek.text)
|
|
@@ -214,4 +209,4 @@ class DiziBox(PluginBase):
|
|
|
214
209
|
"name" : f"{extractor.name if extractor else alt_name}"
|
|
215
210
|
})
|
|
216
211
|
|
|
217
|
-
return results
|
|
212
|
+
return results
|
KekikStream/Plugins/DiziPal.py
CHANGED
|
@@ -9,7 +9,7 @@ class DiziPal(PluginBase):
|
|
|
9
9
|
language = "tr"
|
|
10
10
|
main_url = "https://dizipal1223.com"
|
|
11
11
|
favicon = f"https://www.google.com/s2/favicons?domain={main_url}&sz=64"
|
|
12
|
-
description = "
|
|
12
|
+
description = "dizipal güncel, dizipal yeni ve gerçek adresi. dizipal en yeni dizi ve filmleri güvenli ve hızlı şekilde sunar."
|
|
13
13
|
|
|
14
14
|
main_page = {
|
|
15
15
|
f"{main_url}/diziler/son-bolumler" : "Son Bölümler",
|
KekikStream/Plugins/DiziYou.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
|
|
2
2
|
|
|
3
|
-
from KekikStream.Core import
|
|
3
|
+
from KekikStream.Core import PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode, Subtitle, ExtractResult
|
|
4
4
|
from parsel import Selector
|
|
5
5
|
import re
|
|
6
6
|
|
|
@@ -29,7 +29,6 @@ class DiziYou(PluginBase):
|
|
|
29
29
|
f"{main_url}/dizi-arsivi/page/SAYFA/?tur=Vah%C5%9Fi+Bat%C4%B1" : "Vahşi Batı"
|
|
30
30
|
}
|
|
31
31
|
|
|
32
|
-
#@kekik_cache(ttl=60*60)
|
|
33
32
|
async def get_main_page(self, page: int, url: str, category: str) -> list[MainPageResult]:
|
|
34
33
|
istek = await self.httpx.get(f"{url.replace('SAYFA', str(page))}")
|
|
35
34
|
secici = Selector(istek.text)
|
|
@@ -44,7 +43,6 @@ class DiziYou(PluginBase):
|
|
|
44
43
|
for veri in secici.css("div.single-item")
|
|
45
44
|
]
|
|
46
45
|
|
|
47
|
-
#@kekik_cache(ttl=60*60)
|
|
48
46
|
async def search(self, query: str) -> list[SearchResult]:
|
|
49
47
|
istek = await self.httpx.get(f"{self.main_url}/?s={query}")
|
|
50
48
|
secici = Selector(istek.text)
|
|
@@ -53,18 +51,28 @@ class DiziYou(PluginBase):
|
|
|
53
51
|
SearchResult(
|
|
54
52
|
title = afis.css("div#categorytitle a::text").get().strip(),
|
|
55
53
|
url = self.fix_url(afis.css("div#categorytitle a::attr(href)").get()),
|
|
56
|
-
poster = self.fix_url(afis.css("img::attr(src)").get())
|
|
54
|
+
poster = self.fix_url(afis.css("img::attr(src)").get() or afis.css("img::attr(data-src)").get())
|
|
57
55
|
)
|
|
58
56
|
for afis in secici.css("div.incontent div#list-series")
|
|
59
57
|
]
|
|
60
58
|
|
|
61
|
-
#@kekik_cache(ttl=60*60)
|
|
62
59
|
async def load_item(self, url: str) -> SeriesInfo:
|
|
63
60
|
istek = await self.httpx.get(url)
|
|
64
61
|
secici = Selector(istek.text)
|
|
65
62
|
|
|
66
|
-
title
|
|
67
|
-
|
|
63
|
+
# Title - div.title h1 içinde
|
|
64
|
+
title_raw = secici.css("div.title h1::text").get()
|
|
65
|
+
title = title_raw.strip() if title_raw else ""
|
|
66
|
+
|
|
67
|
+
# Fallback: Eğer title boşsa URL'den çıkar (telif kısıtlaması olan sayfalar için)
|
|
68
|
+
if not title:
|
|
69
|
+
# URL'den slug'ı al: https://www.diziyou.one/jasmine/ -> jasmine -> Jasmine
|
|
70
|
+
slug = url.rstrip('/').split('/')[-1]
|
|
71
|
+
title = slug.replace('-', ' ').title()
|
|
72
|
+
|
|
73
|
+
# Poster
|
|
74
|
+
poster_raw = secici.css("div.category_image img::attr(src)").get()
|
|
75
|
+
poster = self.fix_url(poster_raw) if poster_raw else ""
|
|
68
76
|
year = secici.xpath("//span[contains(., 'Yapım Yılı')]/following-sibling::text()[1]").get()
|
|
69
77
|
description = secici.css("div.diziyou_desc::text").get()
|
|
70
78
|
if description:
|
|
@@ -75,13 +83,21 @@ class DiziYou(PluginBase):
|
|
|
75
83
|
actors = [actor.strip() for actor in _actors.split(",")] if _actors else []
|
|
76
84
|
|
|
77
85
|
episodes = []
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
86
|
+
# Episodes - bolumust her bölüm için bir <a> içinde
|
|
87
|
+
# :has() parsel'de çalışmıyor, XPath kullanıyoruz
|
|
88
|
+
for link in secici.xpath('//a[div[@class="bolumust"]]'):
|
|
89
|
+
ep_name_raw = link.css("div.baslik::text").get()
|
|
90
|
+
if not ep_name_raw:
|
|
91
|
+
continue
|
|
92
|
+
ep_name = ep_name_raw.strip()
|
|
93
|
+
|
|
94
|
+
ep_href = self.fix_url(link.css("::attr(href)").get())
|
|
95
|
+
if not ep_href:
|
|
82
96
|
continue
|
|
83
97
|
|
|
84
|
-
|
|
98
|
+
# Bölüm ismi varsa al
|
|
99
|
+
ep_name_raw_clean = link.css("div.bolumismi::text").get()
|
|
100
|
+
ep_name_clean = ep_name_raw_clean.strip().replace("(", "").replace(")", "").strip() if ep_name_raw_clean else ep_name
|
|
85
101
|
|
|
86
102
|
ep_episode = re.search(r"(\d+)\. Bölüm", ep_name)[1]
|
|
87
103
|
ep_season = re.search(r"(\d+)\. Sezon", ep_name)[1]
|
|
@@ -107,14 +123,23 @@ class DiziYou(PluginBase):
|
|
|
107
123
|
actors = actors
|
|
108
124
|
)
|
|
109
125
|
|
|
110
|
-
#@kekik_cache(ttl=15*60)
|
|
111
126
|
async def load_links(self, url: str) -> list[dict]:
|
|
112
127
|
istek = await self.httpx.get(url)
|
|
113
128
|
secici = Selector(istek.text)
|
|
114
129
|
|
|
115
|
-
|
|
116
|
-
|
|
117
|
-
|
|
130
|
+
# Title ve episode name - None kontrolü ekle
|
|
131
|
+
item_title_raw = secici.css("div.title h1::text").get()
|
|
132
|
+
item_title = item_title_raw.strip() if item_title_raw else ""
|
|
133
|
+
|
|
134
|
+
ep_name_raw = secici.css("div#bolum-ismi::text").get()
|
|
135
|
+
ep_name = ep_name_raw.strip() if ep_name_raw else ""
|
|
136
|
+
|
|
137
|
+
# Player src'den item_id çıkar
|
|
138
|
+
player_src = secici.css("iframe#diziyouPlayer::attr(src)").get()
|
|
139
|
+
if not player_src:
|
|
140
|
+
return [] # Player bulunamadıysa boş liste döndür
|
|
141
|
+
|
|
142
|
+
item_id = player_src.split("/")[-1].replace(".html", "")
|
|
118
143
|
|
|
119
144
|
subtitles = []
|
|
120
145
|
stream_urls = []
|
KekikStream/Plugins/Dizilla.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
|
|
2
2
|
|
|
3
|
-
from KekikStream.Core import
|
|
3
|
+
from KekikStream.Core import PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode
|
|
4
4
|
from parsel import Selector
|
|
5
5
|
from json import loads
|
|
6
6
|
from urllib.parse import urlparse, urlunparse
|
|
@@ -12,36 +12,47 @@ class Dizilla(PluginBase):
|
|
|
12
12
|
language = "tr"
|
|
13
13
|
main_url = "https://dizilla40.com"
|
|
14
14
|
favicon = f"https://www.google.com/s2/favicons?domain={main_url}&sz=64"
|
|
15
|
-
description = "
|
|
15
|
+
description = "1080p yabancı dizi izle. Türkçe altyazılı veya dublaj seçenekleriyle 1080p çözünürlükte yabancı dizilere anında ulaş. Popüler dizileri kesintisiz izle."
|
|
16
16
|
|
|
17
17
|
main_page = {
|
|
18
|
-
f"{main_url}/tum-bolumler"
|
|
19
|
-
f"{main_url}/
|
|
20
|
-
f"{main_url}/
|
|
21
|
-
f"{main_url}/
|
|
22
|
-
f"{main_url}/
|
|
23
|
-
f"{main_url}/
|
|
24
|
-
f"{main_url}/
|
|
18
|
+
f"{main_url}/tum-bolumler" : "Altyazılı Bölümler",
|
|
19
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=15&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Aile",
|
|
20
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=9&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Aksiyon",
|
|
21
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=17&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Animasyon",
|
|
22
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=5&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Bilim Kurgu",
|
|
23
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=2&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Dram",
|
|
24
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=12&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Fantastik",
|
|
25
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=18&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Gerilim",
|
|
26
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=3&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Gizem",
|
|
27
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=4&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Komedi",
|
|
28
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=8&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Korku",
|
|
29
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=24&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Macera",
|
|
30
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=7&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Romantik",
|
|
31
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=26&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Savaş",
|
|
32
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=1&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Suç",
|
|
33
|
+
f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=11&countryIdsComma=&orderType=date_desc&languageId=-1¤tPage=SAYFA¤tPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Western",
|
|
25
34
|
}
|
|
26
35
|
|
|
27
|
-
#@kekik_cache(ttl=60*60)
|
|
28
36
|
async def get_main_page(self, page: int, url: str, category: str) -> list[MainPageResult]:
|
|
29
|
-
istek = await self.httpx.get(url)
|
|
30
|
-
secici = Selector(istek.text)
|
|
31
|
-
|
|
32
37
|
ana_sayfa = []
|
|
33
38
|
|
|
34
|
-
if "
|
|
39
|
+
if "api/bg" in url:
|
|
40
|
+
istek = await self.httpx.post(url.replace("SAYFA", str(page)))
|
|
41
|
+
decrypted = await self.decrypt_response(istek.json().get("response"))
|
|
42
|
+
veriler = decrypted.get("result", [])
|
|
35
43
|
ana_sayfa.extend([
|
|
36
44
|
MainPageResult(
|
|
37
45
|
category = category,
|
|
38
|
-
title = veri.
|
|
39
|
-
url = self.fix_url(veri.
|
|
40
|
-
poster = self.fix_url(veri.
|
|
46
|
+
title = veri.get("original_title"),
|
|
47
|
+
url = self.fix_url(f"{self.main_url}/{veri.get('used_slug')}"),
|
|
48
|
+
poster = self.fix_url(veri.get("object_poster_url")),
|
|
41
49
|
)
|
|
42
|
-
for veri in
|
|
50
|
+
for veri in veriler
|
|
43
51
|
])
|
|
44
52
|
else:
|
|
53
|
+
istek = await self.httpx.get(url.replace("SAYFA", str(page)))
|
|
54
|
+
secici = Selector(istek.text)
|
|
55
|
+
|
|
45
56
|
for veri in secici.css("div.tab-content > div.grid a"):
|
|
46
57
|
name = veri.css("h2::text").get()
|
|
47
58
|
ep_name = veri.xpath("normalize-space(//div[contains(@class, 'opacity-80')])").get()
|
|
@@ -88,7 +99,6 @@ class Dizilla(PluginBase):
|
|
|
88
99
|
# JSON decode
|
|
89
100
|
return loads(decrypted.decode("utf-8"))
|
|
90
101
|
|
|
91
|
-
#@kekik_cache(ttl=60*60)
|
|
92
102
|
async def search(self, query: str) -> list[SearchResult]:
|
|
93
103
|
arama_istek = await self.httpx.post(f"{self.main_url}/api/bg/searchcontent?searchterm={query}")
|
|
94
104
|
decrypted = await self.decrypt_response(arama_istek.json().get("response"))
|
|
@@ -103,7 +113,6 @@ class Dizilla(PluginBase):
|
|
|
103
113
|
for veri in arama_veri
|
|
104
114
|
]
|
|
105
115
|
|
|
106
|
-
#@kekik_cache(ttl=60*60)
|
|
107
116
|
async def url_base_degis(self, eski_url:str, yeni_base:str) -> str:
|
|
108
117
|
parsed_url = urlparse(eski_url)
|
|
109
118
|
parsed_yeni_base = urlparse(yeni_base)
|
|
@@ -114,22 +123,25 @@ class Dizilla(PluginBase):
|
|
|
114
123
|
|
|
115
124
|
return urlunparse(yeni_url)
|
|
116
125
|
|
|
117
|
-
#@kekik_cache(ttl=60*60)
|
|
118
126
|
async def load_item(self, url: str) -> SeriesInfo:
|
|
119
127
|
istek = await self.httpx.get(url)
|
|
120
128
|
secici = Selector(istek.text)
|
|
121
129
|
veri = loads(secici.xpath("//script[@type='application/ld+json']/text()").getall()[-1])
|
|
122
130
|
|
|
123
|
-
title
|
|
131
|
+
title = veri.get("name")
|
|
124
132
|
if alt_title := veri.get("alternateName"):
|
|
125
133
|
title += f" - ({alt_title})"
|
|
126
134
|
|
|
127
135
|
poster = self.fix_url(veri.get("image"))
|
|
128
136
|
description = veri.get("description")
|
|
129
137
|
year = veri.get("datePublished").split("-")[0]
|
|
130
|
-
|
|
131
|
-
|
|
132
|
-
|
|
138
|
+
|
|
139
|
+
# Tags extraction from page content (h3 tag)
|
|
140
|
+
tags_raw = secici.css("h3.text-white.opacity-60::text").get()
|
|
141
|
+
tags = [t.strip() for t in tags_raw.split(",")] if tags_raw else []
|
|
142
|
+
|
|
143
|
+
rating = veri.get("aggregateRating", {}).get("ratingValue")
|
|
144
|
+
actors = [actor.get("name") for actor in veri.get("actor", []) if actor.get("name")]
|
|
133
145
|
|
|
134
146
|
bolumler = []
|
|
135
147
|
sezonlar = veri.get("containsSeason") if isinstance(veri.get("containsSeason"), list) else [veri.get("containsSeason")]
|
|
@@ -158,7 +170,6 @@ class Dizilla(PluginBase):
|
|
|
158
170
|
actors = actors
|
|
159
171
|
)
|
|
160
172
|
|
|
161
|
-
#@kekik_cache(ttl=15*60)
|
|
162
173
|
async def load_links(self, url: str) -> list[dict]:
|
|
163
174
|
istek = await self.httpx.get(url)
|
|
164
175
|
secici = Selector(istek.text)
|
|
@@ -181,4 +192,4 @@ class Dizilla(PluginBase):
|
|
|
181
192
|
"name" : f"{extractor.name if extractor else 'Main Player'} | {result.get('language_name')}",
|
|
182
193
|
})
|
|
183
194
|
|
|
184
|
-
return links
|
|
195
|
+
return links
|
KekikStream/Plugins/FilmBip.py
CHANGED
|
@@ -8,7 +8,7 @@ class FilmBip(PluginBase):
|
|
|
8
8
|
language = "tr"
|
|
9
9
|
main_url = "https://filmbip.com"
|
|
10
10
|
favicon = f"https://www.google.com/s2/favicons?domain={main_url}&sz=64"
|
|
11
|
-
description = "
|
|
11
|
+
description = "FilmBip adlı film sitemizde Full HD film izle. Yerli ve yabancı filmleri Türkçe dublaj veya altyazılı şekilde 1080p yüksek kalite film izle"
|
|
12
12
|
|
|
13
13
|
main_page = {
|
|
14
14
|
f"{main_url}/filmler/SAYFA" : "Yeni Filmler",
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
|
|
2
2
|
|
|
3
|
-
from KekikStream.Core import
|
|
3
|
+
from KekikStream.Core import PluginBase, MainPageResult, SearchResult, MovieInfo
|
|
4
4
|
from parsel import Selector
|
|
5
5
|
|
|
6
6
|
class FilmMakinesi(PluginBase):
|
|
@@ -34,7 +34,6 @@ class FilmMakinesi(PluginBase):
|
|
|
34
34
|
f"{main_url}/tur/western-fm1/film/" : "Western"
|
|
35
35
|
}
|
|
36
36
|
|
|
37
|
-
#@kekik_cache(ttl=60*60)
|
|
38
37
|
async def get_main_page(self, page: int, url: str, category: str) -> list[MainPageResult]:
|
|
39
38
|
istek = self.cloudscraper.get(f"{url}{'' if page == 1 else f'page/{page}/'}")
|
|
40
39
|
secici = Selector(istek.text)
|
|
@@ -51,7 +50,6 @@ class FilmMakinesi(PluginBase):
|
|
|
51
50
|
for veri in veriler
|
|
52
51
|
]
|
|
53
52
|
|
|
54
|
-
#@kekik_cache(ttl=60*60)
|
|
55
53
|
async def search(self, query: str) -> list[SearchResult]:
|
|
56
54
|
istek = await self.httpx.get(f"{self.main_url}/arama/?s={query}")
|
|
57
55
|
secici = Selector(istek.text)
|
|
@@ -73,7 +71,6 @@ class FilmMakinesi(PluginBase):
|
|
|
73
71
|
|
|
74
72
|
return results
|
|
75
73
|
|
|
76
|
-
#@kekik_cache(ttl=60*60)
|
|
77
74
|
async def load_item(self, url: str) -> MovieInfo:
|
|
78
75
|
istek = await self.httpx.get(url)
|
|
79
76
|
secici = Selector(istek.text)
|
|
@@ -101,7 +98,6 @@ class FilmMakinesi(PluginBase):
|
|
|
101
98
|
duration = duration
|
|
102
99
|
)
|
|
103
100
|
|
|
104
|
-
#@kekik_cache(ttl=15*60)
|
|
105
101
|
async def load_links(self, url: str) -> list[dict]:
|
|
106
102
|
istek = await self.httpx.get(url)
|
|
107
103
|
secici = Selector(istek.text)
|