hyperbrowser 0.33.0__tar.gz → 0.34.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of hyperbrowser might be problematic. Click here for more details.

Files changed (44) hide show
  1. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/PKG-INFO +2 -1
  2. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/async_client.py +8 -8
  3. {hyperbrowser-0.33.0/hyperbrowser/client/managers/async_manager/beta → hyperbrowser-0.34.0/hyperbrowser/client/managers/async_manager}/agents/__init__.py +3 -2
  4. {hyperbrowser-0.33.0/hyperbrowser/client/managers/async_manager/beta → hyperbrowser-0.34.0/hyperbrowser/client/managers/async_manager}/agents/browser_use.py +5 -3
  5. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/async_manager/crawl.py +30 -15
  6. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/async_manager/extract.py +15 -7
  7. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/async_manager/profile.py +2 -1
  8. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/async_manager/scrape.py +42 -21
  9. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/async_manager/session.py +2 -1
  10. {hyperbrowser-0.33.0/hyperbrowser/client/managers/sync_manager/beta → hyperbrowser-0.34.0/hyperbrowser/client/managers/sync_manager}/agents/__init__.py +3 -2
  11. {hyperbrowser-0.33.0/hyperbrowser/client/managers/sync_manager/beta → hyperbrowser-0.34.0/hyperbrowser/client/managers/sync_manager}/agents/browser_use.py +5 -3
  12. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/sync_manager/crawl.py +31 -16
  13. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/sync_manager/extract.py +15 -7
  14. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/sync_manager/profile.py +2 -1
  15. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/sync_manager/scrape.py +44 -23
  16. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/sync_manager/session.py +2 -1
  17. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/sync.py +8 -8
  18. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/models/__init__.py +76 -67
  19. {hyperbrowser-0.33.0/hyperbrowser/models/beta → hyperbrowser-0.34.0/hyperbrowser/models}/agents/browser_use.py +4 -2
  20. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/models/crawl.py +12 -0
  21. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/models/extract.py +12 -0
  22. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/models/scrape.py +24 -0
  23. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/tools/__init__.py +23 -0
  24. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/tools/anthropic.py +12 -1
  25. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/tools/openai.py +16 -1
  26. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/tools/schema.py +47 -0
  27. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/pyproject.toml +2 -1
  28. hyperbrowser-0.33.0/hyperbrowser/client/managers/async_manager/beta/__init__.py +0 -6
  29. hyperbrowser-0.33.0/hyperbrowser/client/managers/sync_manager/beta/__init__.py +0 -6
  30. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/LICENSE +0 -0
  31. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/README.md +0 -0
  32. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/__init__.py +0 -0
  33. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/base.py +0 -0
  34. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/async_manager/extension.py +0 -0
  35. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/client/managers/sync_manager/extension.py +0 -0
  36. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/config.py +0 -0
  37. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/exceptions.py +0 -0
  38. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/models/consts.py +0 -0
  39. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/models/extension.py +0 -0
  40. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/models/profile.py +0 -0
  41. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/models/session.py +0 -0
  42. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/transport/async_transport.py +0 -0
  43. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/transport/base.py +0 -0
  44. {hyperbrowser-0.33.0 → hyperbrowser-0.34.0}/hyperbrowser/transport/sync.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: hyperbrowser
3
- Version: 0.33.0
3
+ Version: 0.34.0
4
4
  Summary: Python SDK for hyperbrowser
5
5
  License: MIT
6
6
  Author: Nikhil Shahi
@@ -15,6 +15,7 @@ Classifier: Programming Language :: Python :: 3.11
15
15
  Classifier: Programming Language :: Python :: 3.12
16
16
  Classifier: Programming Language :: Python :: 3.13
17
17
  Requires-Dist: httpx (>=0.23.0,<1)
18
+ Requires-Dist: jsonref (>=1.1.0)
18
19
  Requires-Dist: pydantic (>=2.0,<3)
19
20
  Project-URL: Homepage, https://github.com/hyperbrowserai/python-sdk
20
21
  Project-URL: Repository, https://github.com/hyperbrowserai/python-sdk
@@ -1,15 +1,15 @@
1
1
  from typing import Optional
2
2
 
3
+ from ..config import ClientConfig
4
+ from ..transport.async_transport import AsyncTransport
5
+ from .base import HyperbrowserBase
6
+ from .managers.async_manager.agents import Agents
7
+ from .managers.async_manager.crawl import CrawlManager
8
+ from .managers.async_manager.extension import ExtensionManager
3
9
  from .managers.async_manager.extract import ExtractManager
4
10
  from .managers.async_manager.profile import ProfileManager
5
- from .managers.async_manager.session import SessionManager
6
11
  from .managers.async_manager.scrape import ScrapeManager
7
- from .managers.async_manager.crawl import CrawlManager
8
- from .managers.async_manager.extension import ExtensionManager
9
- from .managers.async_manager.beta import Beta
10
- from .base import HyperbrowserBase
11
- from ..transport.async_transport import AsyncTransport
12
- from ..config import ClientConfig
12
+ from .managers.async_manager.session import SessionManager
13
13
 
14
14
 
15
15
  class AsyncHyperbrowser(HyperbrowserBase):
@@ -30,7 +30,7 @@ class AsyncHyperbrowser(HyperbrowserBase):
30
30
  self.extract = ExtractManager(self)
31
31
  self.profiles = ProfileManager(self)
32
32
  self.extensions = ExtensionManager(self)
33
- self.beta = Beta(self)
33
+ self.agents = Agents(self)
34
34
 
35
35
  async def close(self) -> None:
36
36
  await self.transport.close()
@@ -1,5 +1,6 @@
1
+ from .browser_use import BrowserUseManager
2
+
3
+
1
4
  class Agents:
2
5
  def __init__(self, client):
3
- from .browser_use import BrowserUseManager
4
-
5
6
  self.browser_use = BrowserUseManager(client)
@@ -1,12 +1,14 @@
1
1
  import asyncio
2
+
2
3
  from hyperbrowser.exceptions import HyperbrowserError
3
- from ......models import (
4
+
5
+ from .....models import (
4
6
  POLLING_ATTEMPTS,
5
7
  BasicResponse,
8
+ BrowserUseTaskResponse,
9
+ BrowserUseTaskStatusResponse,
6
10
  StartBrowserUseTaskParams,
7
11
  StartBrowserUseTaskResponse,
8
- BrowserUseTaskStatusResponse,
9
- BrowserUseTaskResponse,
10
12
  )
11
13
 
12
14
 
@@ -3,6 +3,8 @@ import asyncio
3
3
  from hyperbrowser.models.consts import POLLING_ATTEMPTS
4
4
  from ....models.crawl import (
5
5
  CrawlJobResponse,
6
+ CrawlJobStatus,
7
+ CrawlJobStatusResponse,
6
8
  GetCrawlJobParams,
7
9
  StartCrawlJobParams,
8
10
  StartCrawlJobResponse,
@@ -21,11 +23,18 @@ class CrawlManager:
21
23
  )
22
24
  return StartCrawlJobResponse(**response.data)
23
25
 
26
+ async def get_status(self, job_id: str) -> CrawlJobStatusResponse:
27
+ response = await self._client.transport.get(
28
+ self._client._build_url(f"/crawl/{job_id}/status")
29
+ )
30
+ return CrawlJobStatusResponse(**response.data)
31
+
24
32
  async def get(
25
33
  self, job_id: str, params: GetCrawlJobParams = GetCrawlJobParams()
26
34
  ) -> CrawlJobResponse:
27
35
  response = await self._client.transport.get(
28
- self._client._build_url(f"/crawl/{job_id}"), params=params.__dict__
36
+ self._client._build_url(f"/crawl/{job_id}"),
37
+ params=params.model_dump(exclude_none=True, by_alias=True),
29
38
  )
30
39
  return CrawlJobResponse(**response.data)
31
40
 
@@ -37,18 +46,13 @@ class CrawlManager:
37
46
  if not job_id:
38
47
  raise HyperbrowserError("Failed to start crawl job")
39
48
 
40
- job_response: CrawlJobResponse
49
+ job_status: CrawlJobStatus = "pending"
41
50
  failures = 0
42
51
  while True:
43
52
  try:
44
- job_response = await self.get(
45
- job_id,
46
- params=GetCrawlJobParams(batch_size=1),
47
- )
48
- if (
49
- job_response.status == "completed"
50
- or job_response.status == "failed"
51
- ):
53
+ job_status_resp = await self.get_status(job_id)
54
+ job_status = job_status_resp.status
55
+ if job_status == "completed" or job_status == "failed":
52
56
  break
53
57
  except Exception as e:
54
58
  failures += 1
@@ -62,8 +66,7 @@ class CrawlManager:
62
66
  if not return_all_pages:
63
67
  while True:
64
68
  try:
65
- job_response = await self.get(job_id)
66
- return job_response
69
+ return await self.get(job_id)
67
70
  except Exception as e:
68
71
  failures += 1
69
72
  if failures >= POLLING_ATTEMPTS:
@@ -73,9 +76,20 @@ class CrawlManager:
73
76
  await asyncio.sleep(0.5)
74
77
 
75
78
  failures = 0
76
- job_response.current_page_batch = 0
77
- job_response.data = []
78
- while job_response.current_page_batch < job_response.total_page_batches:
79
+ job_response = CrawlJobResponse(
80
+ jobId=job_id,
81
+ status=job_status,
82
+ data=[],
83
+ currentPageBatch=0,
84
+ totalPageBatches=0,
85
+ totalCrawledPages=0,
86
+ batchSize=100,
87
+ )
88
+ first_check = True
89
+ while (
90
+ first_check
91
+ or job_response.current_page_batch < job_response.total_page_batches
92
+ ):
79
93
  try:
80
94
  tmp_job_response = await self.get(
81
95
  job_start_resp.job_id,
@@ -90,6 +104,7 @@ class CrawlManager:
90
104
  job_response.total_page_batches = tmp_job_response.total_page_batches
91
105
  job_response.batch_size = tmp_job_response.batch_size
92
106
  failures = 0
107
+ first_check = False
93
108
  except Exception as e:
94
109
  failures += 1
95
110
  if failures >= POLLING_ATTEMPTS:
@@ -3,9 +3,11 @@ from hyperbrowser.exceptions import HyperbrowserError
3
3
  from hyperbrowser.models.consts import POLLING_ATTEMPTS
4
4
  from hyperbrowser.models.extract import (
5
5
  ExtractJobResponse,
6
+ ExtractJobStatusResponse,
6
7
  StartExtractJobParams,
7
8
  StartExtractJobResponse,
8
9
  )
10
+ import jsonref
9
11
 
10
12
 
11
13
  class ExtractManager:
@@ -17,7 +19,9 @@ class ExtractManager:
17
19
  raise HyperbrowserError("Either schema or prompt must be provided")
18
20
  if params.schema_:
19
21
  if hasattr(params.schema_, "model_json_schema"):
20
- params.schema_ = params.schema_.model_json_schema()
22
+ params.schema_ = jsonref.replace_refs(
23
+ params.schema_.model_json_schema(), proxies=False, lazy_load=False
24
+ )
21
25
 
22
26
  response = await self._client.transport.post(
23
27
  self._client._build_url("/extract"),
@@ -25,6 +29,12 @@ class ExtractManager:
25
29
  )
26
30
  return StartExtractJobResponse(**response.data)
27
31
 
32
+ async def get_status(self, job_id: str) -> ExtractJobStatusResponse:
33
+ response = await self._client.transport.get(
34
+ self._client._build_url(f"/extract/{job_id}/status")
35
+ )
36
+ return ExtractJobStatusResponse(**response.data)
37
+
28
38
  async def get(self, job_id: str) -> ExtractJobResponse:
29
39
  response = await self._client.transport.get(
30
40
  self._client._build_url(f"/extract/{job_id}")
@@ -40,12 +50,10 @@ class ExtractManager:
40
50
  failures = 0
41
51
  while True:
42
52
  try:
43
- job_response = await self.get(job_id)
44
- if (
45
- job_response.status == "completed"
46
- or job_response.status == "failed"
47
- ):
48
- return job_response
53
+ job_status_resp = await self.get_status(job_id)
54
+ job_status = job_status_resp.status
55
+ if job_status == "completed" or job_status == "failed":
56
+ return await self.get(job_id)
49
57
  failures = 0
50
58
  except Exception as e:
51
59
  failures += 1
@@ -33,6 +33,7 @@ class ProfileManager:
33
33
  self, params: ProfileListParams = ProfileListParams()
34
34
  ) -> ProfileListResponse:
35
35
  response = await self._client.transport.get(
36
- self._client._build_url("/profiles"), params=params.__dict__
36
+ self._client._build_url("/profiles"),
37
+ params=params.model_dump(exclude_none=True, by_alias=True),
37
38
  )
38
39
  return ProfileListResponse(**response.data)
@@ -4,8 +4,11 @@ from typing import Optional
4
4
  from hyperbrowser.models.consts import POLLING_ATTEMPTS
5
5
  from ....models.scrape import (
6
6
  BatchScrapeJobResponse,
7
+ BatchScrapeJobStatusResponse,
7
8
  GetBatchScrapeJobParams,
8
9
  ScrapeJobResponse,
10
+ ScrapeJobStatus,
11
+ ScrapeJobStatusResponse,
9
12
  StartBatchScrapeJobParams,
10
13
  StartBatchScrapeJobResponse,
11
14
  StartScrapeJobParams,
@@ -27,11 +30,18 @@ class BatchScrapeManager:
27
30
  )
28
31
  return StartBatchScrapeJobResponse(**response.data)
29
32
 
33
+ async def get_status(self, job_id: str) -> BatchScrapeJobStatusResponse:
34
+ response = await self._client.transport.get(
35
+ self._client._build_url(f"/scrape/batch/{job_id}/status")
36
+ )
37
+ return BatchScrapeJobStatusResponse(**response.data)
38
+
30
39
  async def get(
31
40
  self, job_id: str, params: GetBatchScrapeJobParams = GetBatchScrapeJobParams()
32
41
  ) -> BatchScrapeJobResponse:
33
42
  response = await self._client.transport.get(
34
- self._client._build_url(f"/scrape/batch/{job_id}"), params=params.__dict__
43
+ self._client._build_url(f"/scrape/batch/{job_id}"),
44
+ params=params.model_dump(exclude_none=True, by_alias=True),
35
45
  )
36
46
  return BatchScrapeJobResponse(**response.data)
37
47
 
@@ -43,19 +53,14 @@ class BatchScrapeManager:
43
53
  if not job_id:
44
54
  raise HyperbrowserError("Failed to start batch scrape job")
45
55
 
46
- job_response: BatchScrapeJobResponse
56
+ job_status: ScrapeJobStatus = "pending"
47
57
  failures = 0
48
58
  while True:
49
59
  try:
50
- job_response = await self.get(
51
- job_id, params=GetBatchScrapeJobParams(batch_size=1)
52
- )
53
- if (
54
- job_response.status == "completed"
55
- or job_response.status == "failed"
56
- ):
60
+ job_status_resp = await self.get_status(job_id)
61
+ job_status = job_status_resp.status
62
+ if job_status == "completed" or job_status == "failed":
57
63
  break
58
- failures = 0
59
64
  except Exception as e:
60
65
  failures += 1
61
66
  if failures >= POLLING_ATTEMPTS:
@@ -68,8 +73,7 @@ class BatchScrapeManager:
68
73
  if not return_all_pages:
69
74
  while True:
70
75
  try:
71
- job_response = await self.get(job_id)
72
- return job_response
76
+ return await self.get(job_id)
73
77
  except Exception as e:
74
78
  failures += 1
75
79
  if failures >= POLLING_ATTEMPTS:
@@ -79,9 +83,21 @@ class BatchScrapeManager:
79
83
  await asyncio.sleep(0.5)
80
84
 
81
85
  failures = 0
82
- job_response.current_page_batch = 0
83
- job_response.data = []
84
- while job_response.current_page_batch < job_response.total_page_batches:
86
+ job_response = BatchScrapeJobResponse(
87
+ jobId=job_id,
88
+ status=job_status,
89
+ data=[],
90
+ currentPageBatch=0,
91
+ totalPageBatches=0,
92
+ totalScrapedPages=0,
93
+ batchSize=100,
94
+ )
95
+ first_check = True
96
+
97
+ while (
98
+ first_check
99
+ or job_response.current_page_batch < job_response.total_page_batches
100
+ ):
85
101
  try:
86
102
  tmp_job_response = await self.get(
87
103
  job_id,
@@ -96,6 +112,7 @@ class BatchScrapeManager:
96
112
  job_response.total_page_batches = tmp_job_response.total_page_batches
97
113
  job_response.batch_size = tmp_job_response.batch_size
98
114
  failures = 0
115
+ first_check = False
99
116
  except Exception as e:
100
117
  failures += 1
101
118
  if failures >= POLLING_ATTEMPTS:
@@ -119,6 +136,12 @@ class ScrapeManager:
119
136
  )
120
137
  return StartScrapeJobResponse(**response.data)
121
138
 
139
+ async def get_status(self, job_id: str) -> ScrapeJobStatusResponse:
140
+ response = await self._client.transport.get(
141
+ self._client._build_url(f"/scrape/{job_id}/status")
142
+ )
143
+ return ScrapeJobStatusResponse(**response.data)
144
+
122
145
  async def get(self, job_id: str) -> ScrapeJobResponse:
123
146
  response = await self._client.transport.get(
124
147
  self._client._build_url(f"/scrape/{job_id}")
@@ -134,12 +157,10 @@ class ScrapeManager:
134
157
  failures = 0
135
158
  while True:
136
159
  try:
137
- job_response = await self.get(job_id)
138
- if (
139
- job_response.status == "completed"
140
- or job_response.status == "failed"
141
- ):
142
- return job_response
160
+ job_status_resp = await self.get_status(job_id)
161
+ job_status = job_status_resp.status
162
+ if job_status == "completed" or job_status == "failed":
163
+ return await self.get(job_id)
143
164
  failures = 0
144
165
  except Exception as e:
145
166
  failures += 1
@@ -42,7 +42,8 @@ class SessionManager:
42
42
  self, params: SessionListParams = SessionListParams()
43
43
  ) -> SessionListResponse:
44
44
  response = await self._client.transport.get(
45
- self._client._build_url("/sessions"), params=params.__dict__
45
+ self._client._build_url("/sessions"),
46
+ params=params.model_dump(exclude_none=True, by_alias=True),
46
47
  )
47
48
  return SessionListResponse(**response.data)
48
49
 
@@ -1,5 +1,6 @@
1
+ from .browser_use import BrowserUseManager
2
+
3
+
1
4
  class Agents:
2
5
  def __init__(self, client):
3
- from .browser_use import BrowserUseManager
4
-
5
6
  self.browser_use = BrowserUseManager(client)
@@ -1,12 +1,14 @@
1
1
  import time
2
+
2
3
  from hyperbrowser.exceptions import HyperbrowserError
3
- from ......models import (
4
+
5
+ from .....models import (
4
6
  POLLING_ATTEMPTS,
5
7
  BasicResponse,
8
+ BrowserUseTaskResponse,
9
+ BrowserUseTaskStatusResponse,
6
10
  StartBrowserUseTaskParams,
7
11
  StartBrowserUseTaskResponse,
8
- BrowserUseTaskStatusResponse,
9
- BrowserUseTaskResponse,
10
12
  )
11
13
 
12
14
 
@@ -4,6 +4,8 @@ from typing import Optional
4
4
  from hyperbrowser.models.consts import POLLING_ATTEMPTS
5
5
  from ....models.crawl import (
6
6
  CrawlJobResponse,
7
+ CrawlJobStatus,
8
+ CrawlJobStatusResponse,
7
9
  GetCrawlJobParams,
8
10
  StartCrawlJobParams,
9
11
  StartCrawlJobResponse,
@@ -22,11 +24,18 @@ class CrawlManager:
22
24
  )
23
25
  return StartCrawlJobResponse(**response.data)
24
26
 
27
+ def get_status(self, job_id: str) -> CrawlJobStatusResponse:
28
+ response = self._client.transport.get(
29
+ self._client._build_url(f"/crawl/{job_id}/status")
30
+ )
31
+ return CrawlJobStatusResponse(**response.data)
32
+
25
33
  def get(
26
34
  self, job_id: str, params: GetCrawlJobParams = GetCrawlJobParams()
27
35
  ) -> CrawlJobResponse:
28
36
  response = self._client.transport.get(
29
- self._client._build_url(f"/crawl/{job_id}"), params=params.__dict__
37
+ self._client._build_url(f"/crawl/{job_id}"),
38
+ params=params.model_dump(exclude_none=True, by_alias=True),
30
39
  )
31
40
  return CrawlJobResponse(**response.data)
32
41
 
@@ -38,18 +47,13 @@ class CrawlManager:
38
47
  if not job_id:
39
48
  raise HyperbrowserError("Failed to start crawl job")
40
49
 
41
- job_response: CrawlJobResponse
50
+ job_status: CrawlJobStatus = "pending"
42
51
  failures = 0
43
52
  while True:
44
53
  try:
45
- job_response = self.get(
46
- job_id,
47
- params=GetCrawlJobParams(batch_size=1),
48
- )
49
- if (
50
- job_response.status == "completed"
51
- or job_response.status == "failed"
52
- ):
54
+ job_status_resp = self.get_status(job_id)
55
+ job_status = job_status_resp.status
56
+ if job_status == "completed" or job_status == "failed":
53
57
  break
54
58
  except Exception as e:
55
59
  failures += 1
@@ -63,8 +67,7 @@ class CrawlManager:
63
67
  if not return_all_pages:
64
68
  while True:
65
69
  try:
66
- job_response = self.get(job_id)
67
- return job_response
70
+ return self.get(job_id)
68
71
  except Exception as e:
69
72
  failures += 1
70
73
  if failures >= POLLING_ATTEMPTS:
@@ -74,12 +77,23 @@ class CrawlManager:
74
77
  time.sleep(0.5)
75
78
 
76
79
  failures = 0
77
- job_response.current_page_batch = 0
78
- job_response.data = []
79
- while job_response.current_page_batch < job_response.total_page_batches:
80
+ job_response = CrawlJobResponse(
81
+ jobId=job_id,
82
+ status=job_status,
83
+ data=[],
84
+ currentPageBatch=0,
85
+ totalPageBatches=0,
86
+ totalCrawledPages=0,
87
+ batchSize=100,
88
+ )
89
+ first_check = True
90
+ while (
91
+ first_check
92
+ or job_response.current_page_batch < job_response.total_page_batches
93
+ ):
80
94
  try:
81
95
  tmp_job_response = self.get(
82
- job_id,
96
+ job_start_resp.job_id,
83
97
  GetCrawlJobParams(
84
98
  page=job_response.current_page_batch + 1, batch_size=100
85
99
  ),
@@ -91,6 +105,7 @@ class CrawlManager:
91
105
  job_response.total_page_batches = tmp_job_response.total_page_batches
92
106
  job_response.batch_size = tmp_job_response.batch_size
93
107
  failures = 0
108
+ first_check = False
94
109
  except Exception as e:
95
110
  failures += 1
96
111
  if failures >= POLLING_ATTEMPTS:
@@ -3,9 +3,11 @@ from hyperbrowser.exceptions import HyperbrowserError
3
3
  from hyperbrowser.models.consts import POLLING_ATTEMPTS
4
4
  from hyperbrowser.models.extract import (
5
5
  ExtractJobResponse,
6
+ ExtractJobStatusResponse,
6
7
  StartExtractJobParams,
7
8
  StartExtractJobResponse,
8
9
  )
10
+ import jsonref
9
11
 
10
12
 
11
13
  class ExtractManager:
@@ -17,7 +19,9 @@ class ExtractManager:
17
19
  raise HyperbrowserError("Either schema or prompt must be provided")
18
20
  if params.schema_:
19
21
  if hasattr(params.schema_, "model_json_schema"):
20
- params.schema_ = params.schema_.model_json_schema()
22
+ params.schema_ = jsonref.replace_refs(
23
+ params.schema_.model_json_schema(), proxies=False, lazy_load=False
24
+ )
21
25
 
22
26
  response = self._client.transport.post(
23
27
  self._client._build_url("/extract"),
@@ -25,6 +29,12 @@ class ExtractManager:
25
29
  )
26
30
  return StartExtractJobResponse(**response.data)
27
31
 
32
+ def get_status(self, job_id: str) -> ExtractJobStatusResponse:
33
+ response = self._client.transport.get(
34
+ self._client._build_url(f"/extract/{job_id}/status")
35
+ )
36
+ return ExtractJobStatusResponse(**response.data)
37
+
28
38
  def get(self, job_id: str) -> ExtractJobResponse:
29
39
  response = self._client.transport.get(
30
40
  self._client._build_url(f"/extract/{job_id}")
@@ -40,12 +50,10 @@ class ExtractManager:
40
50
  failures = 0
41
51
  while True:
42
52
  try:
43
- job_response = self.get(job_start_resp.job_id)
44
- if (
45
- job_response.status == "completed"
46
- or job_response.status == "failed"
47
- ):
48
- return job_response
53
+ job_status_resp = self.get_status(job_id)
54
+ job_status = job_status_resp.status
55
+ if job_status == "completed" or job_status == "failed":
56
+ return self.get(job_id)
49
57
  except Exception as e:
50
58
  failures += 1
51
59
  if failures >= POLLING_ATTEMPTS:
@@ -33,6 +33,7 @@ class ProfileManager:
33
33
  self, params: ProfileListParams = ProfileListParams()
34
34
  ) -> ProfileListResponse:
35
35
  response = self._client.transport.get(
36
- self._client._build_url("/profiles"), params=params.__dict__
36
+ self._client._build_url("/profiles"),
37
+ params=params.model_dump(exclude_none=True, by_alias=True),
37
38
  )
38
39
  return ProfileListResponse(**response.data)
@@ -4,8 +4,11 @@ from typing import Optional
4
4
  from hyperbrowser.models.consts import POLLING_ATTEMPTS
5
5
  from ....models.scrape import (
6
6
  BatchScrapeJobResponse,
7
+ BatchScrapeJobStatusResponse,
7
8
  GetBatchScrapeJobParams,
8
9
  ScrapeJobResponse,
10
+ ScrapeJobStatus,
11
+ ScrapeJobStatusResponse,
9
12
  StartBatchScrapeJobParams,
10
13
  StartBatchScrapeJobResponse,
11
14
  StartScrapeJobParams,
@@ -25,11 +28,18 @@ class BatchScrapeManager:
25
28
  )
26
29
  return StartBatchScrapeJobResponse(**response.data)
27
30
 
31
+ def get_status(self, job_id: str) -> BatchScrapeJobStatusResponse:
32
+ response = self._client.transport.get(
33
+ self._client._build_url(f"/scrape/batch/{job_id}/status")
34
+ )
35
+ return BatchScrapeJobStatusResponse(**response.data)
36
+
28
37
  def get(
29
38
  self, job_id: str, params: GetBatchScrapeJobParams = GetBatchScrapeJobParams()
30
39
  ) -> BatchScrapeJobResponse:
31
40
  response = self._client.transport.get(
32
- self._client._build_url(f"/scrape/batch/{job_id}"), params=params.__dict__
41
+ self._client._build_url(f"/scrape/batch/{job_id}"),
42
+ params=params.model_dump(exclude_none=True, by_alias=True),
33
43
  )
34
44
  return BatchScrapeJobResponse(**response.data)
35
45
 
@@ -41,19 +51,14 @@ class BatchScrapeManager:
41
51
  if not job_id:
42
52
  raise HyperbrowserError("Failed to start batch scrape job")
43
53
 
44
- job_response: BatchScrapeJobResponse
54
+ job_status: ScrapeJobStatus = "pending"
45
55
  failures = 0
46
56
  while True:
47
57
  try:
48
- job_response = self.get(
49
- job_id, params=GetBatchScrapeJobParams(batch_size=1)
50
- )
51
- if (
52
- job_response.status == "completed"
53
- or job_response.status == "failed"
54
- ):
58
+ job_status_resp = self.get_status(job_id)
59
+ job_status = job_status_resp.status
60
+ if job_status == "completed" or job_status == "failed":
55
61
  break
56
- failures = 0
57
62
  except Exception as e:
58
63
  failures += 1
59
64
  if failures >= POLLING_ATTEMPTS:
@@ -66,8 +71,7 @@ class BatchScrapeManager:
66
71
  if not return_all_pages:
67
72
  while True:
68
73
  try:
69
- job_response = self.get(job_id)
70
- return job_response
74
+ return self.get(job_id)
71
75
  except Exception as e:
72
76
  failures += 1
73
77
  if failures >= POLLING_ATTEMPTS:
@@ -77,13 +81,25 @@ class BatchScrapeManager:
77
81
  time.sleep(0.5)
78
82
 
79
83
  failures = 0
80
- job_response.current_page_batch = 0
81
- job_response.data = []
82
- while job_response.current_page_batch < job_response.total_page_batches:
84
+ job_response = BatchScrapeJobResponse(
85
+ jobId=job_id,
86
+ status=job_status,
87
+ data=[],
88
+ currentPageBatch=0,
89
+ totalPageBatches=0,
90
+ totalScrapedPages=0,
91
+ batchSize=100,
92
+ )
93
+ first_check = True
94
+
95
+ while (
96
+ first_check
97
+ or job_response.current_page_batch < job_response.total_page_batches
98
+ ):
83
99
  try:
84
100
  tmp_job_response = self.get(
85
- job_start_resp.job_id,
86
- GetBatchScrapeJobParams(
101
+ job_id,
102
+ params=GetBatchScrapeJobParams(
87
103
  page=job_response.current_page_batch + 1, batch_size=100
88
104
  ),
89
105
  )
@@ -94,6 +110,7 @@ class BatchScrapeManager:
94
110
  job_response.total_page_batches = tmp_job_response.total_page_batches
95
111
  job_response.batch_size = tmp_job_response.batch_size
96
112
  failures = 0
113
+ first_check = False
97
114
  except Exception as e:
98
115
  failures += 1
99
116
  if failures >= POLLING_ATTEMPTS:
@@ -117,6 +134,12 @@ class ScrapeManager:
117
134
  )
118
135
  return StartScrapeJobResponse(**response.data)
119
136
 
137
+ def get_status(self, job_id: str) -> ScrapeJobStatusResponse:
138
+ response = self._client.transport.get(
139
+ self._client._build_url(f"/scrape/{job_id}/status")
140
+ )
141
+ return ScrapeJobStatusResponse(**response.data)
142
+
120
143
  def get(self, job_id: str) -> ScrapeJobResponse:
121
144
  response = self._client.transport.get(
122
145
  self._client._build_url(f"/scrape/{job_id}")
@@ -132,12 +155,10 @@ class ScrapeManager:
132
155
  failures = 0
133
156
  while True:
134
157
  try:
135
- job_response = self.get(job_id)
136
- if (
137
- job_response.status == "completed"
138
- or job_response.status == "failed"
139
- ):
140
- return job_response
158
+ job_status_resp = self.get_status(job_id)
159
+ job_status = job_status_resp.status
160
+ if job_status == "completed" or job_status == "failed":
161
+ return self.get(job_id)
141
162
  failures = 0
142
163
  except Exception as e:
143
164
  failures += 1
@@ -40,7 +40,8 @@ class SessionManager:
40
40
  self, params: SessionListParams = SessionListParams()
41
41
  ) -> SessionListResponse:
42
42
  response = self._client.transport.get(
43
- self._client._build_url("/sessions"), params=params.__dict__
43
+ self._client._build_url("/sessions"),
44
+ params=params.model_dump(exclude_none=True, by_alias=True),
44
45
  )
45
46
  return SessionListResponse(**response.data)
46
47
 
@@ -1,15 +1,15 @@
1
1
  from typing import Optional
2
2
 
3
+ from ..config import ClientConfig
4
+ from ..transport.sync import SyncTransport
5
+ from .base import HyperbrowserBase
6
+ from .managers.sync_manager.agents import Agents
7
+ from .managers.sync_manager.crawl import CrawlManager
8
+ from .managers.sync_manager.extension import ExtensionManager
3
9
  from .managers.sync_manager.extract import ExtractManager
4
10
  from .managers.sync_manager.profile import ProfileManager
5
- from .managers.sync_manager.session import SessionManager
6
11
  from .managers.sync_manager.scrape import ScrapeManager
7
- from .managers.sync_manager.crawl import CrawlManager
8
- from .managers.sync_manager.beta import Beta
9
- from .managers.sync_manager.extension import ExtensionManager
10
- from .base import HyperbrowserBase
11
- from ..transport.sync import SyncTransport
12
- from ..config import ClientConfig
12
+ from .managers.sync_manager.session import SessionManager
13
13
 
14
14
 
15
15
  class Hyperbrowser(HyperbrowserBase):
@@ -30,7 +30,7 @@ class Hyperbrowser(HyperbrowserBase):
30
30
  self.extract = ExtractManager(self)
31
31
  self.profiles = ProfileManager(self)
32
32
  self.extensions = ExtensionManager(self)
33
- self.beta = Beta(self)
33
+ self.agents = Agents(self)
34
34
 
35
35
  def close(self) -> None:
36
36
  self.transport.close()
@@ -1,142 +1,151 @@
1
+ from .agents.browser_use import (
2
+ BrowserUseTaskData,
3
+ BrowserUseTaskResponse,
4
+ BrowserUseTaskStatusResponse,
5
+ StartBrowserUseTaskParams,
6
+ StartBrowserUseTaskResponse,
7
+ )
1
8
  from .consts import (
2
- ScrapeFormat,
3
- ScrapeWaitUntil,
4
- ScrapePageStatus,
5
- ScrapeScreenshotFormat,
6
- RecordingStatus,
7
- DownloadsStatus,
9
+ ISO639_1,
8
10
  POLLING_ATTEMPTS,
11
+ BrowserUseLlm,
9
12
  Country,
13
+ DownloadsStatus,
10
14
  OperatingSystem,
11
15
  Platform,
12
- ISO639_1,
16
+ RecordingStatus,
17
+ ScrapeFormat,
18
+ ScrapePageStatus,
19
+ ScrapeScreenshotFormat,
20
+ ScrapeWaitUntil,
13
21
  State,
14
- BrowserUseLlm,
15
22
  )
16
23
  from .crawl import (
24
+ CrawledPage,
25
+ CrawlJobResponse,
17
26
  CrawlJobStatus,
27
+ CrawlJobStatusResponse,
18
28
  CrawlPageStatus,
29
+ GetCrawlJobParams,
19
30
  StartCrawlJobParams,
20
31
  StartCrawlJobResponse,
21
- CrawledPage,
22
- GetCrawlJobParams,
23
- CrawlJobResponse,
24
32
  )
25
33
  from .extension import CreateExtensionParams, ExtensionResponse
26
34
  from .extract import (
35
+ ExtractJobResponse,
27
36
  ExtractJobStatus,
37
+ ExtractJobStatusResponse,
28
38
  StartExtractJobParams,
29
39
  StartExtractJobResponse,
30
- ExtractJobResponse,
31
40
  )
32
41
  from .profile import (
33
42
  CreateProfileResponse,
34
- ProfileResponse,
35
43
  ProfileListParams,
36
44
  ProfileListResponse,
45
+ ProfileResponse,
37
46
  )
38
47
  from .scrape import (
39
- ScrapeJobStatus,
40
- ScreenshotOptions,
41
- ScrapeOptions,
42
- StartScrapeJobParams,
43
- StartScrapeJobResponse,
48
+ BatchScrapeJobResponse,
49
+ BatchScrapeJobStatusResponse,
50
+ GetBatchScrapeJobParams,
51
+ ScrapedPage,
44
52
  ScrapeJobData,
45
53
  ScrapeJobResponse,
54
+ ScrapeJobStatus,
55
+ ScrapeJobStatusResponse,
56
+ ScrapeOptions,
57
+ ScreenshotOptions,
46
58
  StartBatchScrapeJobParams,
47
- ScrapedPage,
48
- GetBatchScrapeJobParams,
49
59
  StartBatchScrapeJobResponse,
50
- BatchScrapeJobResponse,
60
+ StartScrapeJobParams,
61
+ StartScrapeJobResponse,
51
62
  )
52
63
  from .session import (
53
- SessionStatus,
54
64
  BasicResponse,
65
+ CreateSessionParams,
66
+ CreateSessionProfile,
67
+ GetSessionDownloadsUrlResponse,
68
+ GetSessionRecordingUrlResponse,
69
+ ScreenConfig,
55
70
  Session,
56
71
  SessionDetail,
57
72
  SessionListParams,
58
73
  SessionListResponse,
59
- ScreenConfig,
60
- CreateSessionProfile,
61
- CreateSessionParams,
62
74
  SessionRecording,
63
- GetSessionRecordingUrlResponse,
64
- GetSessionDownloadsUrlResponse,
65
- )
66
- from .beta.agents.browser_use import (
67
- StartBrowserUseTaskParams,
68
- StartBrowserUseTaskResponse,
69
- BrowserUseTaskStatusResponse,
70
- BrowserUseTaskData,
71
- BrowserUseTaskResponse,
75
+ SessionStatus,
72
76
  )
73
77
 
74
78
  __all__ = [
75
79
  # consts
76
- "BrowserUseLlm",
77
- "ScrapeFormat",
78
- "ScrapeWaitUntil",
79
- "ScrapePageStatus",
80
- "ScrapeScreenshotFormat",
81
- "RecordingStatus",
82
- "DownloadsStatus",
80
+ "ISO639_1",
83
81
  "POLLING_ATTEMPTS",
82
+ "BrowserUseLlm",
84
83
  "Country",
84
+ "DownloadsStatus",
85
85
  "OperatingSystem",
86
86
  "Platform",
87
- "ISO639_1",
87
+ "RecordingStatus",
88
+ "ScrapeFormat",
89
+ "ScrapePageStatus",
90
+ "ScrapeScreenshotFormat",
91
+ "ScrapeWaitUntil",
88
92
  "State",
93
+ # agents
94
+ "BrowserUseTaskStatus",
95
+ "BrowserUseTaskData",
96
+ "BrowserUseTaskResponse",
97
+ "BrowserUseTaskStatusResponse",
98
+ "StartBrowserUseTaskParams",
99
+ "StartBrowserUseTaskResponse",
89
100
  # crawl
101
+ "CrawledPage",
102
+ "CrawlJobResponse",
90
103
  "CrawlJobStatus",
104
+ "CrawlJobStatusResponse",
91
105
  "CrawlPageStatus",
106
+ "GetCrawlJobParams",
92
107
  "StartCrawlJobParams",
93
108
  "StartCrawlJobResponse",
94
- "CrawledPage",
95
- "GetCrawlJobParams",
96
- "CrawlJobResponse",
97
109
  # extension
98
110
  "CreateExtensionParams",
99
111
  "ExtensionResponse",
100
112
  # extract
113
+ "ExtractJobResponse",
101
114
  "ExtractJobStatus",
115
+ "ExtractJobStatusResponse",
102
116
  "StartExtractJobParams",
103
117
  "StartExtractJobResponse",
104
- "ExtractJobResponse",
105
118
  # profile
106
119
  "CreateProfileResponse",
107
- "ProfileResponse",
108
120
  "ProfileListParams",
109
121
  "ProfileListResponse",
122
+ "ProfileResponse",
110
123
  # scrape
111
- "ScrapeJobStatus",
112
- "ScreenshotOptions",
113
- "ScrapeOptions",
114
- "StartScrapeJobParams",
115
- "StartScrapeJobResponse",
124
+ "BatchScrapeJobResponse",
125
+ "BatchScrapeJobStatusResponse",
126
+ "GetBatchScrapeJobParams",
127
+ "ScrapedPage",
116
128
  "ScrapeJobData",
117
129
  "ScrapeJobResponse",
130
+ "ScrapeJobStatus",
131
+ "ScrapeJobStatusResponse",
132
+ "ScrapeOptions",
133
+ "ScreenshotOptions",
118
134
  "StartBatchScrapeJobParams",
119
- "ScrapedPage",
120
- "GetBatchScrapeJobParams",
121
135
  "StartBatchScrapeJobResponse",
122
- "BatchScrapeJobResponse",
136
+ "StartScrapeJobParams",
137
+ "StartScrapeJobResponse",
123
138
  # session
124
- "SessionStatus",
125
139
  "BasicResponse",
140
+ "CreateSessionParams",
141
+ "CreateSessionProfile",
142
+ "GetSessionDownloadsUrlResponse",
143
+ "GetSessionRecordingUrlResponse",
144
+ "ScreenConfig",
126
145
  "Session",
127
146
  "SessionDetail",
128
147
  "SessionListParams",
129
148
  "SessionListResponse",
130
- "ScreenConfig",
131
- "CreateSessionProfile",
132
- "CreateSessionParams",
133
149
  "SessionRecording",
134
- "GetSessionRecordingUrlResponse",
135
- "GetSessionDownloadsUrlResponse",
136
- # agents
137
- "StartBrowserUseTaskParams",
138
- "StartBrowserUseTaskResponse",
139
- "BrowserUseTaskStatusResponse",
140
- "BrowserUseTaskData",
141
- "BrowserUseTaskResponse",
150
+ "SessionStatus",
142
151
  ]
@@ -1,8 +1,9 @@
1
1
  from typing import Literal, Optional, Union
2
+
2
3
  from pydantic import BaseModel, ConfigDict, Field
3
4
 
4
- from ...consts import BrowserUseLlm
5
- from ...session import CreateSessionParams
5
+ from ..consts import BrowserUseLlm
6
+ from ..session import CreateSessionParams
6
7
 
7
8
  BrowserUseTaskStatus = Literal["pending", "running", "completed", "failed", "stopped"]
8
9
 
@@ -182,3 +183,4 @@ class BrowserUseTaskResponse(BaseModel):
182
183
  data: Optional[BrowserUseTaskData] = Field(default=None, alias="data")
183
184
  error: Optional[str] = Field(default=None, alias="error")
184
185
  live_url: Optional[str] = Field(default=None, alias="liveUrl")
186
+ live_url: Optional[str] = Field(default=None, alias="liveUrl")
@@ -47,6 +47,18 @@ class StartCrawlJobResponse(BaseModel):
47
47
  job_id: str = Field(alias="jobId")
48
48
 
49
49
 
50
+ class CrawlJobStatusResponse(BaseModel):
51
+ """
52
+ Response from getting the status of a crawl job.
53
+ """
54
+
55
+ model_config = ConfigDict(
56
+ populate_by_alias=True,
57
+ )
58
+
59
+ status: CrawlJobStatus
60
+
61
+
50
62
  class CrawledPage(BaseModel):
51
63
  """
52
64
  Data from a crawled page.
@@ -43,6 +43,18 @@ class StartExtractJobResponse(BaseModel):
43
43
  job_id: str = Field(alias="jobId")
44
44
 
45
45
 
46
+ class ExtractJobStatusResponse(BaseModel):
47
+ """
48
+ Response from getting the status of a extract job.
49
+ """
50
+
51
+ model_config = ConfigDict(
52
+ populate_by_alias=True,
53
+ )
54
+
55
+ status: ExtractJobStatus
56
+
57
+
46
58
  class ExtractJobResponse(BaseModel):
47
59
  """
48
60
  Response from a extract job.
@@ -78,6 +78,18 @@ class StartScrapeJobResponse(BaseModel):
78
78
  job_id: str = Field(alias="jobId")
79
79
 
80
80
 
81
+ class ScrapeJobStatusResponse(BaseModel):
82
+ """
83
+ Response from getting the status of a scrape job.
84
+ """
85
+
86
+ model_config = ConfigDict(
87
+ populate_by_alias=True,
88
+ )
89
+
90
+ status: ScrapeJobStatus
91
+
92
+
81
93
  class ScrapeJobData(BaseModel):
82
94
  """
83
95
  Data from a scraped site.
@@ -119,6 +131,18 @@ class StartBatchScrapeJobParams(BaseModel):
119
131
  )
120
132
 
121
133
 
134
+ class BatchScrapeJobStatusResponse(BaseModel):
135
+ """
136
+ Response from getting the status of a batch scrape job.
137
+ """
138
+
139
+ model_config = ConfigDict(
140
+ populate_by_alias=True,
141
+ )
142
+
143
+ status: ScrapeJobStatus
144
+
145
+
122
146
  class ScrapedPage(BaseModel):
123
147
  """
124
148
  A scraped page.
@@ -1,15 +1,18 @@
1
1
  import json
2
+ from hyperbrowser.models.beta.agents.browser_use import StartBrowserUseTaskParams
2
3
  from hyperbrowser.models.crawl import StartCrawlJobParams
3
4
  from hyperbrowser.models.extract import StartExtractJobParams
4
5
  from hyperbrowser.models.scrape import StartScrapeJobParams
5
6
  from hyperbrowser import Hyperbrowser, AsyncHyperbrowser
6
7
 
7
8
  from .openai import (
9
+ BROWSER_USE_TOOL_OPENAI,
8
10
  EXTRACT_TOOL_OPENAI,
9
11
  SCRAPE_TOOL_OPENAI,
10
12
  CRAWL_TOOL_OPENAI,
11
13
  )
12
14
  from .anthropic import (
15
+ BROWSER_USE_TOOL_ANTHROPIC,
13
16
  EXTRACT_TOOL_ANTHROPIC,
14
17
  SCRAPE_TOOL_ANTHROPIC,
15
18
  CRAWL_TOOL_ANTHROPIC,
@@ -79,8 +82,28 @@ class WebsiteExtractTool:
79
82
  return json.dumps(resp.data) if resp.data else ""
80
83
 
81
84
 
85
+ class BrowserUseTool:
86
+ openai_tool_definition = BROWSER_USE_TOOL_OPENAI
87
+ anthropic_tool_definition = BROWSER_USE_TOOL_ANTHROPIC
88
+
89
+ @staticmethod
90
+ def runnable(hb: Hyperbrowser, params: dict) -> str:
91
+ resp = hb.beta.agents.browser_use.start_and_wait(
92
+ params=StartBrowserUseTaskParams(**params)
93
+ )
94
+ return resp.data.final_result if resp.data and resp.data.final_result else ""
95
+
96
+ @staticmethod
97
+ async def async_runnable(hb: AsyncHyperbrowser, params: dict) -> str:
98
+ resp = await hb.beta.agents.browser_use.start_and_wait(
99
+ params=StartBrowserUseTaskParams(**params)
100
+ )
101
+ return resp.data.final_result if resp.data and resp.data.final_result else ""
102
+
103
+
82
104
  __all__ = [
83
105
  "WebsiteScrapeTool",
84
106
  "WebsiteCrawlTool",
85
107
  "WebsiteExtractTool",
108
+ "BrowserUseTool",
86
109
  ]
@@ -1,7 +1,12 @@
1
1
  from typing import Dict, Union, Optional
2
2
  from typing_extensions import Literal, Required, TypeAlias, TypedDict
3
3
 
4
- from hyperbrowser.tools.schema import CRAWL_SCHEMA, EXTRACT_SCHEMA, SCRAPE_SCHEMA
4
+ from hyperbrowser.tools.schema import (
5
+ BROWSER_USE_SCHEMA,
6
+ CRAWL_SCHEMA,
7
+ EXTRACT_SCHEMA,
8
+ SCRAPE_SCHEMA,
9
+ )
5
10
 
6
11
 
7
12
  class CacheControlEphemeralParam(TypedDict, total=False):
@@ -60,3 +65,9 @@ EXTRACT_TOOL_ANTHROPIC: ToolParam = {
60
65
  "name": "extract_data",
61
66
  "description": "Extract data in a structured format from multiple URLs in a single function call. IMPORTANT: When information must be gathered from multiple sources (such as comparing items, researching topics across sites, or answering questions that span multiple webpages), ALWAYS include all relevant URLs in ONE function call. This enables comprehensive answers with cross-referenced information. Returns data as a json string.",
62
67
  }
68
+
69
+ BROWSER_USE_TOOL_ANTHROPIC: ToolParam = {
70
+ "input_schema": BROWSER_USE_SCHEMA,
71
+ "name": "browser_use",
72
+ "description": "Have an AI agent use a browser to perform a task on the web.",
73
+ }
@@ -1,7 +1,12 @@
1
1
  from typing import Dict, Optional
2
2
  from typing_extensions import Literal, Required, TypedDict, TypeAlias
3
3
 
4
- from hyperbrowser.tools.schema import CRAWL_SCHEMA, EXTRACT_SCHEMA, SCRAPE_SCHEMA
4
+ from hyperbrowser.tools.schema import (
5
+ BROWSER_USE_SCHEMA,
6
+ CRAWL_SCHEMA,
7
+ EXTRACT_SCHEMA,
8
+ SCRAPE_SCHEMA,
9
+ )
5
10
 
6
11
  FunctionParameters: TypeAlias = Dict[str, object]
7
12
 
@@ -77,3 +82,13 @@ EXTRACT_TOOL_OPENAI: ChatCompletionToolParam = {
77
82
  "strict": True,
78
83
  },
79
84
  }
85
+
86
+ BROWSER_USE_TOOL_OPENAI: ChatCompletionToolParam = {
87
+ "type": "function",
88
+ "function": {
89
+ "name": "browser_use",
90
+ "description": "Have an AI agent use a browser to perform a task on the web.",
91
+ "parameters": BROWSER_USE_SCHEMA,
92
+ "strict": True,
93
+ },
94
+ }
@@ -111,3 +111,50 @@ EXTRACT_SCHEMA = {
111
111
  "required": ["urls", "prompt", "schema", "max_links"],
112
112
  "additionalProperties": False,
113
113
  }
114
+
115
+ BROWSER_USE_LLM_SCHEMA = {
116
+ "type": "string",
117
+ "enum": [
118
+ "gpt-4o",
119
+ "gpt-4o-mini",
120
+ "claude-3-7-sonnet-20250219",
121
+ "claude-3-5-sonnet-20241022",
122
+ "claude-3-5-haiku-20241022",
123
+ "gemini-2.0-flash",
124
+ ],
125
+ "default": "gemini-2.0-flash",
126
+ }
127
+
128
+ BROWSER_USE_SCHEMA = {
129
+ "type": "object",
130
+ "properties": {
131
+ "task": {
132
+ "type": "string",
133
+ "description": "The text description of the task to be performed by the agent.",
134
+ },
135
+ "llm": {
136
+ **BROWSER_USE_LLM_SCHEMA,
137
+ "description": "The language model (LLM) instance to use for generating actions. Default to gemini-2.0-flash.",
138
+ },
139
+ "planner_llm": {
140
+ **BROWSER_USE_LLM_SCHEMA,
141
+ "description": "The language model to use specifically for planning future actions, can differ from the main LLM. Default to gemini-2.0-flash.",
142
+ },
143
+ "page_extraction_llm": {
144
+ **BROWSER_USE_LLM_SCHEMA,
145
+ "description": "The language model to use for extracting structured data from webpages. Default to gemini-2.0-flash.",
146
+ },
147
+ "keep_browser_open": {
148
+ "type": "boolean",
149
+ "description": "When enabled, keeps the browser session open after task completion.",
150
+ },
151
+ },
152
+ "required": [
153
+ "task",
154
+ "llm",
155
+ "planner_llm",
156
+ "page_extraction_llm",
157
+ "keep_browser_open",
158
+ ],
159
+ "additionalProperties": False,
160
+ }
@@ -1,6 +1,6 @@
1
1
  [tool.poetry]
2
2
  name = "hyperbrowser"
3
- version = "0.33.0"
3
+ version = "0.34.0"
4
4
  description = "Python SDK for hyperbrowser"
5
5
  authors = ["Nikhil Shahi <nshahi1998@gmail.com>"]
6
6
  license = "MIT"
@@ -13,6 +13,7 @@ repository = "https://github.com/hyperbrowserai/python-sdk"
13
13
  python = "^3.8"
14
14
  pydantic = ">=2.0,<3"
15
15
  httpx = ">=0.23.0,<1"
16
+ jsonref = ">=1.1.0"
16
17
 
17
18
 
18
19
  [tool.poetry.group.dev.dependencies]
@@ -1,6 +0,0 @@
1
- from .agents import Agents
2
-
3
-
4
- class Beta:
5
- def __init__(self, client):
6
- self.agents = Agents(client)
@@ -1,6 +0,0 @@
1
- from .agents import Agents
2
-
3
-
4
- class Beta:
5
- def __init__(self, client):
6
- self.agents = Agents(client)
File without changes
File without changes