KekikStream 1.8.1__py3-none-any.whl → 1.9.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,7 +1,6 @@
1
1
  # Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
2
2
 
3
3
  from abc import ABC, abstractmethod
4
- from curl_cffi import AsyncSession
5
4
  from cloudscraper import CloudScraper
6
5
  from httpx import AsyncClient
7
6
  from typing import Optional
@@ -12,7 +11,6 @@ class ExtractorBase(ABC):
12
11
  # Çıkarıcının temel özellikleri
13
12
  name = "Extractor"
14
13
  main_url = ""
15
- requires_cffi = False
16
14
 
17
15
  def __init__(self):
18
16
  # cloudscraper - for bypassing Cloudflare
@@ -21,19 +19,11 @@ class ExtractorBase(ABC):
21
19
  # httpx - lightweight and safe for most HTTP requests
22
20
  self.httpx = AsyncClient(
23
21
  timeout = 3,
24
- follow_redirects = True,
22
+ follow_redirects = True
25
23
  )
26
24
  self.httpx.headers.update(self.cloudscraper.headers)
27
25
  self.httpx.cookies.update(self.cloudscraper.cookies)
28
26
 
29
- # curl_cffi - only initialize if needed for anti-bot bypass
30
- self.cffi = None
31
-
32
- if self.requires_cffi:
33
- self.cffi = AsyncSession(impersonate="firefox135")
34
- self.cffi.cookies.update(self.cloudscraper.cookies)
35
- self.cffi.headers.update({"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 15.7; rv:135.0) Gecko/20100101 Firefox/135.0"})
36
-
37
27
  def can_handle_url(self, url: str) -> bool:
38
28
  # URL'nin bu çıkarıcı tarafından işlenip işlenemeyeceğini kontrol et
39
29
  return self.main_url in url
@@ -44,10 +34,8 @@ class ExtractorBase(ABC):
44
34
  pass
45
35
 
46
36
  async def close(self):
47
- """Close both HTTP clients if they exist."""
37
+ """Close HTTP client."""
48
38
  await self.httpx.aclose()
49
- if self.cffi:
50
- await self.cffi.close()
51
39
 
52
40
  def fix_url(self, url: str) -> str:
53
41
  # Eksik URL'leri düzelt ve tam URL formatına çevir
@@ -18,19 +18,34 @@ class MediaHandler:
18
18
  if extract_data.referer:
19
19
  self.headers["referer"] = extract_data.referer
20
20
 
21
- # Google Drive gibi özel durumlar için yt-dlp kullan
21
+ # Özel Durumlar (RecTV vs. Googleusercontent)
22
22
  if user_agent in ["googleusercontent", "Mozilla/5.0 (X11; Linux x86_64; rv:101.0) Gecko/20100101 Firefox/101.0"]:
23
23
  return self.play_with_ytdlp(extract_data)
24
24
 
25
- # İşletim sistemine göre oynatıcı seç
25
+ # İşletim sistemine göre oynatıcı seç (Android durumu)
26
26
  if subprocess.check_output(['uname', '-o']).strip() == b'Android':
27
27
  return self.play_with_android_mxplayer(extract_data)
28
28
 
29
- # Alt yazılar varsa mpv kullan
30
- if extract_data.subtitles:
31
- return self.play_with_mpv(extract_data)
29
+ # Oynatıcı öncelik sırası (fallback zincirleme)
30
+ players = [
31
+ ("MPV", self.play_with_mpv),
32
+ ("VLC", self.play_with_vlc),
33
+ ("yt-dlp", self.play_with_ytdlp)
34
+ ]
35
+
36
+ # Fallback zincirleme
37
+ for player_name, player_func in players:
38
+ try:
39
+ result = player_func(extract_data)
40
+ if result or result is None: # None = MPV (exception yok)
41
+ konsol.log(f"[green][✓] {player_name} ile başarılı[/green]")
42
+ return True
43
+ except Exception as e:
44
+ konsol.log(f"[yellow][⚠] {player_name} hatası: {e}[/yellow]")
45
+ continue
32
46
 
33
- return self.play_with_vlc(extract_data) or self.play_with_mpv(extract_data)
47
+ konsol.print("[red][✗] Hiçbir oynatıcı çalışmadı![/red]")
48
+ return False
34
49
 
35
50
  def play_with_vlc(self, extract_data: ExtractResult):
36
51
  konsol.log(f"[yellow][»] VLC ile Oynatılıyor : {extract_data.url}")
@@ -88,12 +103,15 @@ class MediaHandler:
88
103
  with open(os.devnull, "w") as devnull:
89
104
  subprocess.run(mpv_command, stdout=devnull, stderr=devnull, check=True)
90
105
 
106
+ return True
91
107
  except subprocess.CalledProcessError as hata:
92
108
  konsol.print(f"[red]mpv oynatma hatası: {hata}[/red]")
93
109
  konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
110
+ return False
94
111
  except FileNotFoundError:
95
112
  konsol.print("[red]mpv bulunamadı! mpv kurulu olduğundan emin olun.[/red]")
96
113
  konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
114
+ return False
97
115
 
98
116
  def play_with_ytdlp(self, extract_data: ExtractResult):
99
117
  konsol.log(f"[yellow][»] yt-dlp ile Oynatılıyor : {extract_data.url}")
@@ -121,12 +139,15 @@ class MediaHandler:
121
139
  with subprocess.Popen(ytdlp_command, stdout=subprocess.PIPE) as ytdlp_proc:
122
140
  subprocess.run(mpv_command, stdin=ytdlp_proc.stdout, check=True)
123
141
 
142
+ return True
124
143
  except subprocess.CalledProcessError as hata:
125
144
  konsol.print(f"[red]Oynatma hatası: {hata}[/red]")
126
145
  konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
146
+ return False
127
147
  except FileNotFoundError:
128
148
  konsol.print("[red]yt-dlp veya mpv bulunamadı! Kurulumlarından emin olun.[/red]")
129
149
  konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
150
+ return False
130
151
 
131
152
  def play_with_android_mxplayer(self, extract_data: ExtractResult):
132
153
  konsol.log(f"[yellow][»] MxPlayer ile Oynatılıyor : {extract_data.url}")
@@ -151,11 +172,12 @@ class MediaHandler:
151
172
  with open(os.devnull, "w") as devnull:
152
173
  subprocess.run(android_command, stdout=devnull, stderr=devnull, check=True)
153
174
 
154
- return
155
-
175
+ return True
156
176
  except subprocess.CalledProcessError as hata:
157
177
  konsol.print(f"[red]{paket} oynatma hatası: {hata}[/red]")
158
178
  konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
179
+ return False
159
180
  except FileNotFoundError:
160
181
  konsol.print(f"Paket: {paket}, Hata: MX Player kurulu değil")
161
182
  konsol.print({"title": self.title, "url": extract_data.url, "headers": self.headers})
183
+ return False
@@ -1,7 +1,6 @@
1
1
  # Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
2
2
 
3
3
  from abc import ABC, abstractmethod
4
- from curl_cffi import AsyncSession
5
4
  from cloudscraper import CloudScraper
6
5
  from httpx import AsyncClient
7
6
  from .PluginModels import MainPageResult, SearchResult, MovieInfo
@@ -17,8 +16,6 @@ class PluginBase(ABC):
17
16
  favicon = f"https://www.google.com/s2/favicons?domain={main_url}&sz=64"
18
17
  description = "No description provided."
19
18
 
20
- requires_cffi = False
21
-
22
19
  main_page = {}
23
20
 
24
21
  async def url_update(self, new_url: str):
@@ -33,19 +30,11 @@ class PluginBase(ABC):
33
30
  # httpx - lightweight and safe for most HTTP requests
34
31
  self.httpx = AsyncClient(
35
32
  timeout = 3,
36
- follow_redirects = True,
33
+ follow_redirects = True
37
34
  )
38
35
  self.httpx.headers.update(self.cloudscraper.headers)
39
36
  self.httpx.cookies.update(self.cloudscraper.cookies)
40
37
 
41
- # curl_cffi - only initialize if needed for anti-bot bypass
42
- self.cffi = None
43
-
44
- if self.requires_cffi:
45
- self.cffi = AsyncSession(impersonate="firefox135")
46
- self.cffi.cookies.update(self.cloudscraper.cookies)
47
- self.cffi.headers.update({"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 15.7; rv:135.0) Gecko/20100101 Firefox/135.0"})
48
-
49
38
  self.media_handler = MediaHandler()
50
39
  self.ex_manager = ExtractorManager()
51
40
 
@@ -90,10 +79,8 @@ class PluginBase(ABC):
90
79
  pass
91
80
 
92
81
  async def close(self):
93
- """Close both HTTP clients if they exist."""
82
+ """Close HTTP client."""
94
83
  await self.httpx.aclose()
95
- if self.cffi:
96
- await self.cffi.close()
97
84
 
98
85
  def fix_url(self, url: str) -> str:
99
86
  if not url:
@@ -0,0 +1,109 @@
1
+ # Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
2
+
3
+ from KekikStream.Core import ExtractorBase, ExtractResult, Subtitle
4
+ import yt_dlp, sys, os
5
+
6
+ class YTDLP(ExtractorBase):
7
+ name = "yt-dlp"
8
+ main_url = "" # Universal - tüm siteleri destekler
9
+
10
+ def __init__(self):
11
+ pass
12
+
13
+ def can_handle_url(self, url: str) -> bool:
14
+ """
15
+ yt-dlp'nin bu URL'yi işleyip işleyemeyeceğini kontrol et
16
+ """
17
+ try:
18
+ # stderr'ı geçici olarak kapat (hata mesajlarını gizle)
19
+ old_stderr = sys.stderr
20
+ sys.stderr = open(os.devnull, "w")
21
+
22
+ try:
23
+ ydl_opts = {
24
+ "simulate" : True, # Download yok, sadece tespit
25
+ "quiet" : True, # Log kirliliği yok
26
+ "no_warnings" : True, # Uyarı mesajları yok
27
+ "extract_flat" : True, # Minimal işlem
28
+ "no_check_certificates" : True,
29
+ "ignoreerrors" : True # Hataları yoksay
30
+ }
31
+
32
+ with yt_dlp.YoutubeDL(ydl_opts) as ydl:
33
+ # URL'yi işleyebiliyor mu kontrol et
34
+ info = ydl.extract_info(url, download=False, process=False)
35
+
36
+ # Generic extractor ise atla
37
+ if info and info.get("extractor_key") != "Generic":
38
+ return True
39
+
40
+ return False
41
+ finally:
42
+ # stderr'ı geri yükle
43
+ sys.stderr.close()
44
+ sys.stderr = old_stderr
45
+
46
+ except Exception:
47
+ # yt-dlp işleyemezse False döndür
48
+ return False
49
+
50
+ async def extract(self, url: str, referer: str | None = None) -> ExtractResult:
51
+ ydl_opts = {
52
+ "quiet" : True,
53
+ "no_warnings" : True,
54
+ "extract_flat" : False, # Tam bilgi al
55
+ "format" : "best", # En iyi kalite
56
+ "no_check_certificates" : True
57
+ }
58
+
59
+ # Referer varsa header olarak ekle
60
+ if referer:
61
+ ydl_opts["http_headers"] = {"Referer": referer}
62
+
63
+ with yt_dlp.YoutubeDL(ydl_opts) as ydl:
64
+ info = ydl.extract_info(url, download=False)
65
+
66
+ if not info:
67
+ raise ValueError("yt-dlp video bilgisi döndürmedi")
68
+
69
+ # Video URL'sini al
70
+ video_url = info.get("url")
71
+ if not video_url:
72
+ # Bazen formatlar listesinde olabilir
73
+ formats = info.get("formats", [])
74
+ if formats:
75
+ video_url = formats[-1].get("url") # Son format (genellikle en iyi)
76
+
77
+ if not video_url:
78
+ raise ValueError("Video URL bulunamadı")
79
+
80
+ # Altyazıları çıkar
81
+ subtitles = []
82
+ if subtitle_data := info.get("subtitles"):
83
+ for lang, subs in subtitle_data.items():
84
+ for sub in subs:
85
+ if sub_url := sub.get("url"):
86
+ subtitles.append(
87
+ Subtitle(
88
+ name=f"{lang} ({sub.get('ext', 'unknown')})",
89
+ url=sub_url
90
+ )
91
+ )
92
+
93
+ # User-Agent al
94
+ user_agent = None
95
+ http_headers = info.get("http_headers", {})
96
+ if http_headers:
97
+ user_agent = http_headers.get("User-Agent")
98
+
99
+ return ExtractResult(
100
+ name = self.name,
101
+ url = video_url,
102
+ referer = referer or info.get("webpage_url"),
103
+ user_agent = user_agent,
104
+ subtitles = subtitles
105
+ )
106
+
107
+ async def close(self):
108
+ """yt-dlp için cleanup gerekmez"""
109
+ pass
@@ -1,6 +1,6 @@
1
1
  # Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
2
2
 
3
- from KekikStream.Core import kekik_cache, PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode
3
+ from KekikStream.Core import PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode
4
4
  from Kekik.Sifreleme import CryptoJS
5
5
  from parsel import Selector
6
6
  import re, urllib.parse, base64, contextlib, asyncio, time
@@ -40,7 +40,6 @@ class DiziBox(PluginBase):
40
40
  f"{main_url}/dizi-arsivi/page/SAYFA/?tur[0]=yarisma&yil&imdb" : "Yarışma"
41
41
  }
42
42
 
43
- #@kekik_cache(ttl=60*60)
44
43
  async def get_main_page(self, page: int, url: str, category: str) -> list[MainPageResult]:
45
44
  self.httpx.cookies.update({
46
45
  "isTrustedUser" : "true",
@@ -62,7 +61,6 @@ class DiziBox(PluginBase):
62
61
  for veri in secici.css("article.detailed-article")
63
62
  ]
64
63
 
65
- #@kekik_cache(ttl=60*60)
66
64
  async def search(self, query: str) -> list[SearchResult]:
67
65
  self.httpx.cookies.update({
68
66
  "isTrustedUser" : "true",
@@ -80,7 +78,6 @@ class DiziBox(PluginBase):
80
78
  for item in secici.css("article.detailed-article")
81
79
  ]
82
80
 
83
- #@kekik_cache(ttl=60*60)
84
81
  async def load_item(self, url: str) -> SeriesInfo:
85
82
  istek = await self.httpx.get(url)
86
83
  secici = Selector(istek.text)
@@ -127,7 +124,6 @@ class DiziBox(PluginBase):
127
124
  actors = actors,
128
125
  )
129
126
 
130
- #@kekik_cache(ttl=60*60)
131
127
  async def _iframe_decode(self, name:str, iframe_link:str, referer:str) -> list[str]:
132
128
  results = []
133
129
 
@@ -178,7 +174,6 @@ class DiziBox(PluginBase):
178
174
 
179
175
  return results
180
176
 
181
- #@kekik_cache(ttl=15*60)
182
177
  async def load_links(self, url: str) -> list[dict]:
183
178
  istek = await self.httpx.get(url)
184
179
  secici = Selector(istek.text)
@@ -214,4 +209,4 @@ class DiziBox(PluginBase):
214
209
  "name" : f"{extractor.name if extractor else alt_name}"
215
210
  })
216
211
 
217
- return results
212
+ return results
@@ -9,7 +9,7 @@ class DiziPal(PluginBase):
9
9
  language = "tr"
10
10
  main_url = "https://dizipal1223.com"
11
11
  favicon = f"https://www.google.com/s2/favicons?domain={main_url}&sz=64"
12
- description = "Yabancı Dizi ve Film izle."
12
+ description = "dizipal güncel, dizipal yeni ve gerçek adresi. dizipal en yeni dizi ve filmleri güvenli ve hızlı şekilde sunar."
13
13
 
14
14
  main_page = {
15
15
  f"{main_url}/diziler/son-bolumler" : "Son Bölümler",
@@ -1,6 +1,6 @@
1
1
  # Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
2
2
 
3
- from KekikStream.Core import kekik_cache, PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode, Subtitle, ExtractResult
3
+ from KekikStream.Core import PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode, Subtitle, ExtractResult
4
4
  from parsel import Selector
5
5
  import re
6
6
 
@@ -29,7 +29,6 @@ class DiziYou(PluginBase):
29
29
  f"{main_url}/dizi-arsivi/page/SAYFA/?tur=Vah%C5%9Fi+Bat%C4%B1" : "Vahşi Batı"
30
30
  }
31
31
 
32
- #@kekik_cache(ttl=60*60)
33
32
  async def get_main_page(self, page: int, url: str, category: str) -> list[MainPageResult]:
34
33
  istek = await self.httpx.get(f"{url.replace('SAYFA', str(page))}")
35
34
  secici = Selector(istek.text)
@@ -44,7 +43,6 @@ class DiziYou(PluginBase):
44
43
  for veri in secici.css("div.single-item")
45
44
  ]
46
45
 
47
- #@kekik_cache(ttl=60*60)
48
46
  async def search(self, query: str) -> list[SearchResult]:
49
47
  istek = await self.httpx.get(f"{self.main_url}/?s={query}")
50
48
  secici = Selector(istek.text)
@@ -53,18 +51,28 @@ class DiziYou(PluginBase):
53
51
  SearchResult(
54
52
  title = afis.css("div#categorytitle a::text").get().strip(),
55
53
  url = self.fix_url(afis.css("div#categorytitle a::attr(href)").get()),
56
- poster = self.fix_url(afis.css("img::attr(src)").get()),
54
+ poster = self.fix_url(afis.css("img::attr(src)").get() or afis.css("img::attr(data-src)").get())
57
55
  )
58
56
  for afis in secici.css("div.incontent div#list-series")
59
57
  ]
60
58
 
61
- #@kekik_cache(ttl=60*60)
62
59
  async def load_item(self, url: str) -> SeriesInfo:
63
60
  istek = await self.httpx.get(url)
64
61
  secici = Selector(istek.text)
65
62
 
66
- title = secici.css("h1::text").get().strip()
67
- poster = self.fix_url(secici.css("div.category_image img::attr(src)").get().strip())
63
+ # Title - div.title h1 içinde
64
+ title_raw = secici.css("div.title h1::text").get()
65
+ title = title_raw.strip() if title_raw else ""
66
+
67
+ # Fallback: Eğer title boşsa URL'den çıkar (telif kısıtlaması olan sayfalar için)
68
+ if not title:
69
+ # URL'den slug'ı al: https://www.diziyou.one/jasmine/ -> jasmine -> Jasmine
70
+ slug = url.rstrip('/').split('/')[-1]
71
+ title = slug.replace('-', ' ').title()
72
+
73
+ # Poster
74
+ poster_raw = secici.css("div.category_image img::attr(src)").get()
75
+ poster = self.fix_url(poster_raw) if poster_raw else ""
68
76
  year = secici.xpath("//span[contains(., 'Yapım Yılı')]/following-sibling::text()[1]").get()
69
77
  description = secici.css("div.diziyou_desc::text").get()
70
78
  if description:
@@ -75,13 +83,21 @@ class DiziYou(PluginBase):
75
83
  actors = [actor.strip() for actor in _actors.split(",")] if _actors else []
76
84
 
77
85
  episodes = []
78
- for it in secici.css("div.bolumust"):
79
- ep_name = it.css("div.baslik::text").get().strip()
80
- ep_href = it.xpath("ancestor::a/@href").get()
81
- if not ep_name or not ep_href:
86
+ # Episodes - bolumust her bölüm için bir <a> içinde
87
+ # :has() parsel'de çalışmıyor, XPath kullanıyoruz
88
+ for link in secici.xpath('//a[div[@class="bolumust"]]'):
89
+ ep_name_raw = link.css("div.baslik::text").get()
90
+ if not ep_name_raw:
91
+ continue
92
+ ep_name = ep_name_raw.strip()
93
+
94
+ ep_href = self.fix_url(link.css("::attr(href)").get())
95
+ if not ep_href:
82
96
  continue
83
97
 
84
- ep_name_clean = it.css("div.bolumismi::text").get().strip().replace("(", "").replace(")", "").strip() if it.css("div.bolumismi::text").get() else ep_name
98
+ # Bölüm ismi varsa al
99
+ ep_name_raw_clean = link.css("div.bolumismi::text").get()
100
+ ep_name_clean = ep_name_raw_clean.strip().replace("(", "").replace(")", "").strip() if ep_name_raw_clean else ep_name
85
101
 
86
102
  ep_episode = re.search(r"(\d+)\. Bölüm", ep_name)[1]
87
103
  ep_season = re.search(r"(\d+)\. Sezon", ep_name)[1]
@@ -107,14 +123,23 @@ class DiziYou(PluginBase):
107
123
  actors = actors
108
124
  )
109
125
 
110
- #@kekik_cache(ttl=15*60)
111
126
  async def load_links(self, url: str) -> list[dict]:
112
127
  istek = await self.httpx.get(url)
113
128
  secici = Selector(istek.text)
114
129
 
115
- item_title = secici.css("div.title h1::text").get()
116
- ep_name = secici.css("div#bolum-ismi::text").get().strip()
117
- item_id = secici.css("iframe#diziyouPlayer::attr(src)").get().split("/")[-1].replace(".html", "")
130
+ # Title ve episode name - None kontrolü ekle
131
+ item_title_raw = secici.css("div.title h1::text").get()
132
+ item_title = item_title_raw.strip() if item_title_raw else ""
133
+
134
+ ep_name_raw = secici.css("div#bolum-ismi::text").get()
135
+ ep_name = ep_name_raw.strip() if ep_name_raw else ""
136
+
137
+ # Player src'den item_id çıkar
138
+ player_src = secici.css("iframe#diziyouPlayer::attr(src)").get()
139
+ if not player_src:
140
+ return [] # Player bulunamadıysa boş liste döndür
141
+
142
+ item_id = player_src.split("/")[-1].replace(".html", "")
118
143
 
119
144
  subtitles = []
120
145
  stream_urls = []
@@ -1,6 +1,6 @@
1
1
  # Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
2
2
 
3
- from KekikStream.Core import kekik_cache, PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode
3
+ from KekikStream.Core import PluginBase, MainPageResult, SearchResult, SeriesInfo, Episode
4
4
  from parsel import Selector
5
5
  from json import loads
6
6
  from urllib.parse import urlparse, urlunparse
@@ -12,36 +12,47 @@ class Dizilla(PluginBase):
12
12
  language = "tr"
13
13
  main_url = "https://dizilla40.com"
14
14
  favicon = f"https://www.google.com/s2/favicons?domain={main_url}&sz=64"
15
- description = "Dizilla tüm yabancı dizileri ücretsiz olarak Türkçe Dublaj ve altyazılı seçenekleri ile 1080P kalite izleyebileceğiniz yeni nesil yabancı dizi izleme siteniz."
15
+ description = "1080p yabancı dizi izle. Türkçe altyazılı veya dublaj seçenekleriyle 1080p çözünürlükte yabancı dizilere anında ulaş. Popüler dizileri kesintisiz izle."
16
16
 
17
17
  main_page = {
18
- f"{main_url}/tum-bolumler" : "Altyazılı Bölümler",
19
- f"{main_url}/dublaj-bolumler" : "Dublaj Bölümler",
20
- f"{main_url}/dizi-turu/aile" : "Aile",
21
- f"{main_url}/dizi-turu/aksiyon" : "Aksiyon",
22
- f"{main_url}/dizi-turu/bilim-kurgu" : "Bilim Kurgu",
23
- f"{main_url}/dizi-turu/romantik" : "Romantik",
24
- f"{main_url}/dizi-turu/komedi" : "Komedi"
18
+ f"{main_url}/tum-bolumler" : "Altyazılı Bölümler",
19
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=15&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Aile",
20
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=9&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Aksiyon",
21
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=17&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Animasyon",
22
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=5&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Bilim Kurgu",
23
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=2&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Dram",
24
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=12&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Fantastik",
25
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=18&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Gerilim",
26
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=3&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Gizem",
27
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=4&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Komedi",
28
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=8&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Korku",
29
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=24&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Macera",
30
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=7&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Romantik",
31
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=26&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Savaş",
32
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=1&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Suç",
33
+ f"{main_url}/api/bg/findSeries?releaseYearStart=1900&releaseYearEnd=2050&imdbPointMin=0&imdbPointMax=10&categoryIdsComma=11&countryIdsComma=&orderType=date_desc&languageId=-1&currentPage=SAYFA&currentPageCount=24&queryStr=&categorySlugsComma=&countryCodesComma=" : "Western",
25
34
  }
26
35
 
27
- #@kekik_cache(ttl=60*60)
28
36
  async def get_main_page(self, page: int, url: str, category: str) -> list[MainPageResult]:
29
- istek = await self.httpx.get(url)
30
- secici = Selector(istek.text)
31
-
32
37
  ana_sayfa = []
33
38
 
34
- if "dizi-turu" in url:
39
+ if "api/bg" in url:
40
+ istek = await self.httpx.post(url.replace("SAYFA", str(page)))
41
+ decrypted = await self.decrypt_response(istek.json().get("response"))
42
+ veriler = decrypted.get("result", [])
35
43
  ana_sayfa.extend([
36
44
  MainPageResult(
37
45
  category = category,
38
- title = veri.css("h2::text").get(),
39
- url = self.fix_url(veri.css("::attr(href)").get()),
40
- poster = self.fix_url(veri.css("img::attr(src)").get() or veri.css("img::attr(data-src)").get())
46
+ title = veri.get("original_title"),
47
+ url = self.fix_url(f"{self.main_url}/{veri.get('used_slug')}"),
48
+ poster = self.fix_url(veri.get("object_poster_url")),
41
49
  )
42
- for veri in secici.css("div.grid-cols-3 a")
50
+ for veri in veriler
43
51
  ])
44
52
  else:
53
+ istek = await self.httpx.get(url.replace("SAYFA", str(page)))
54
+ secici = Selector(istek.text)
55
+
45
56
  for veri in secici.css("div.tab-content > div.grid a"):
46
57
  name = veri.css("h2::text").get()
47
58
  ep_name = veri.xpath("normalize-space(//div[contains(@class, 'opacity-80')])").get()
@@ -88,7 +99,6 @@ class Dizilla(PluginBase):
88
99
  # JSON decode
89
100
  return loads(decrypted.decode("utf-8"))
90
101
 
91
- #@kekik_cache(ttl=60*60)
92
102
  async def search(self, query: str) -> list[SearchResult]:
93
103
  arama_istek = await self.httpx.post(f"{self.main_url}/api/bg/searchcontent?searchterm={query}")
94
104
  decrypted = await self.decrypt_response(arama_istek.json().get("response"))
@@ -103,7 +113,6 @@ class Dizilla(PluginBase):
103
113
  for veri in arama_veri
104
114
  ]
105
115
 
106
- #@kekik_cache(ttl=60*60)
107
116
  async def url_base_degis(self, eski_url:str, yeni_base:str) -> str:
108
117
  parsed_url = urlparse(eski_url)
109
118
  parsed_yeni_base = urlparse(yeni_base)
@@ -114,22 +123,25 @@ class Dizilla(PluginBase):
114
123
 
115
124
  return urlunparse(yeni_url)
116
125
 
117
- #@kekik_cache(ttl=60*60)
118
126
  async def load_item(self, url: str) -> SeriesInfo:
119
127
  istek = await self.httpx.get(url)
120
128
  secici = Selector(istek.text)
121
129
  veri = loads(secici.xpath("//script[@type='application/ld+json']/text()").getall()[-1])
122
130
 
123
- title = veri.get("name")
131
+ title = veri.get("name")
124
132
  if alt_title := veri.get("alternateName"):
125
133
  title += f" - ({alt_title})"
126
134
 
127
135
  poster = self.fix_url(veri.get("image"))
128
136
  description = veri.get("description")
129
137
  year = veri.get("datePublished").split("-")[0]
130
- tags = []
131
- rating = veri.get("aggregateRating", {}).get("ratingValue")
132
- actors = [actor.get("name") for actor in veri.get("actor", []) if actor.get("name")]
138
+
139
+ # Tags extraction from page content (h3 tag)
140
+ tags_raw = secici.css("h3.text-white.opacity-60::text").get()
141
+ tags = [t.strip() for t in tags_raw.split(",")] if tags_raw else []
142
+
143
+ rating = veri.get("aggregateRating", {}).get("ratingValue")
144
+ actors = [actor.get("name") for actor in veri.get("actor", []) if actor.get("name")]
133
145
 
134
146
  bolumler = []
135
147
  sezonlar = veri.get("containsSeason") if isinstance(veri.get("containsSeason"), list) else [veri.get("containsSeason")]
@@ -158,7 +170,6 @@ class Dizilla(PluginBase):
158
170
  actors = actors
159
171
  )
160
172
 
161
- #@kekik_cache(ttl=15*60)
162
173
  async def load_links(self, url: str) -> list[dict]:
163
174
  istek = await self.httpx.get(url)
164
175
  secici = Selector(istek.text)
@@ -181,4 +192,4 @@ class Dizilla(PluginBase):
181
192
  "name" : f"{extractor.name if extractor else 'Main Player'} | {result.get('language_name')}",
182
193
  })
183
194
 
184
- return links
195
+ return links
@@ -8,7 +8,7 @@ class FilmBip(PluginBase):
8
8
  language = "tr"
9
9
  main_url = "https://filmbip.com"
10
10
  favicon = f"https://www.google.com/s2/favicons?domain={main_url}&sz=64"
11
- description = "Film izleme sitesi."
11
+ description = "FilmBip adlı film sitemizde Full HD film izle. Yerli ve yabancı filmleri Türkçe dublaj veya altyazılı şekilde 1080p yüksek kalite film izle"
12
12
 
13
13
  main_page = {
14
14
  f"{main_url}/filmler/SAYFA" : "Yeni Filmler",
@@ -1,6 +1,6 @@
1
1
  # Bu araç @keyiflerolsun tarafından | @KekikAkademi için yazılmıştır.
2
2
 
3
- from KekikStream.Core import kekik_cache, PluginBase, MainPageResult, SearchResult, MovieInfo
3
+ from KekikStream.Core import PluginBase, MainPageResult, SearchResult, MovieInfo
4
4
  from parsel import Selector
5
5
 
6
6
  class FilmMakinesi(PluginBase):
@@ -34,7 +34,6 @@ class FilmMakinesi(PluginBase):
34
34
  f"{main_url}/tur/western-fm1/film/" : "Western"
35
35
  }
36
36
 
37
- #@kekik_cache(ttl=60*60)
38
37
  async def get_main_page(self, page: int, url: str, category: str) -> list[MainPageResult]:
39
38
  istek = self.cloudscraper.get(f"{url}{'' if page == 1 else f'page/{page}/'}")
40
39
  secici = Selector(istek.text)
@@ -51,7 +50,6 @@ class FilmMakinesi(PluginBase):
51
50
  for veri in veriler
52
51
  ]
53
52
 
54
- #@kekik_cache(ttl=60*60)
55
53
  async def search(self, query: str) -> list[SearchResult]:
56
54
  istek = await self.httpx.get(f"{self.main_url}/arama/?s={query}")
57
55
  secici = Selector(istek.text)
@@ -73,7 +71,6 @@ class FilmMakinesi(PluginBase):
73
71
 
74
72
  return results
75
73
 
76
- #@kekik_cache(ttl=60*60)
77
74
  async def load_item(self, url: str) -> MovieInfo:
78
75
  istek = await self.httpx.get(url)
79
76
  secici = Selector(istek.text)
@@ -101,7 +98,6 @@ class FilmMakinesi(PluginBase):
101
98
  duration = duration
102
99
  )
103
100
 
104
- #@kekik_cache(ttl=15*60)
105
101
  async def load_links(self, url: str) -> list[dict]:
106
102
  istek = await self.httpx.get(url)
107
103
  secici = Selector(istek.text)