hishel 0.1.2__py3-none-any.whl → 0.1.4__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
hishel/__init__.py CHANGED
@@ -8,10 +8,50 @@ from ._serializers import *
8
8
  from ._sync import *
9
9
  from ._lfu_cache import *
10
10
 
11
+ __all__ = (
12
+ # Old API
13
+ "AsyncCacheClient",
14
+ "MockAsyncConnectionPool",
15
+ "MockAsyncTransport",
16
+ "AsyncCacheConnectionPool",
17
+ "AsyncBaseStorage",
18
+ "AsyncFileStorage",
19
+ "AsyncInMemoryStorage",
20
+ "AsyncRedisStorage",
21
+ "AsyncS3Storage",
22
+ "AsyncSQLiteStorage",
23
+ "AsyncCacheTransport",
24
+ "HEURISTICALLY_CACHEABLE_STATUS_CODES",
25
+ "Controller",
26
+ "CacheControlError",
27
+ "ParseError",
28
+ "ValidationError",
29
+ "CacheControl",
30
+ "Vary",
31
+ "BaseSerializer",
32
+ "JSONSerializer",
33
+ "Metadata",
34
+ "PickleSerializer",
35
+ "YAMLSerializer",
36
+ "clone_model",
37
+ "CacheClient",
38
+ "MockConnectionPool",
39
+ "MockTransport",
40
+ "CacheConnectionPool",
41
+ "BaseStorage",
42
+ "FileStorage",
43
+ "InMemoryStorage",
44
+ "RedisStorage",
45
+ "S3Storage",
46
+ "SQLiteStorage",
47
+ "CacheTransport",
48
+ "LFUCache",
49
+ )
11
50
 
12
51
  def install_cache() -> None: # pragma: no cover
13
52
  httpx.AsyncClient = AsyncCacheClient # type: ignore
14
53
  httpx.Client = CacheClient # type: ignore
15
54
 
16
55
 
17
- __version__ = "0.1.2"
56
+ __version__ = "0.1.4"
57
+
hishel/_async/_client.py CHANGED
@@ -2,7 +2,7 @@ import typing as tp
2
2
 
3
3
  import httpx
4
4
 
5
- from hishel._async._transports import AsyncCacheTransport
5
+ from ._transports import AsyncCacheTransport
6
6
 
7
7
  __all__ = ("AsyncCacheClient",)
8
8
 
@@ -4,7 +4,6 @@ import datetime
4
4
  import logging
5
5
  import os
6
6
  import time
7
- import typing as t
8
7
  import typing as tp
9
8
  import warnings
10
9
  from copy import deepcopy
@@ -24,13 +23,11 @@ except ImportError: # pragma: no cover
24
23
 
25
24
  from httpcore import Request, Response
26
25
 
27
- if t.TYPE_CHECKING: # pragma: no cover
26
+ if tp.TYPE_CHECKING: # pragma: no cover
28
27
  from typing_extensions import TypeAlias
29
28
 
30
- from hishel._serializers import BaseSerializer, clone_model
31
-
32
29
  from .._files import AsyncFileManager
33
- from .._serializers import JSONSerializer, Metadata
30
+ from .._serializers import BaseSerializer, JSONSerializer, Metadata, clone_model
34
31
  from .._synchronization import AsyncLock
35
32
  from .._utils import float_seconds_to_int_milliseconds
36
33
 
@@ -39,10 +36,10 @@ logger = logging.getLogger("hishel.storages")
39
36
  __all__ = (
40
37
  "AsyncBaseStorage",
41
38
  "AsyncFileStorage",
42
- "AsyncRedisStorage",
43
- "AsyncSQLiteStorage",
44
39
  "AsyncInMemoryStorage",
40
+ "AsyncRedisStorage",
45
41
  "AsyncS3Storage",
42
+ "AsyncSQLiteStorage",
46
43
  )
47
44
 
48
45
  StoredResponse: TypeAlias = tp.Tuple[Response, Request, Metadata]
@@ -106,8 +103,7 @@ class AsyncFileStorage(AsyncBaseStorage):
106
103
  self._base_path = Path(base_path) if base_path is not None else Path(".cache/hishel")
107
104
  self._gitignore_file = self._base_path / ".gitignore"
108
105
 
109
- if not self._base_path.is_dir():
110
- self._base_path.mkdir(parents=True)
106
+ self._base_path.mkdir(parents=True, exist_ok=True)
111
107
 
112
108
  if not self._gitignore_file.is_file():
113
109
  with open(self._gitignore_file, "w", encoding="utf-8") as f:
@@ -153,13 +149,13 @@ class AsyncFileStorage(AsyncBaseStorage):
153
149
  """
154
150
 
155
151
  if isinstance(key, Response): # pragma: no cover
156
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
152
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
157
153
 
158
154
  response_path = self._base_path / key
159
155
 
160
156
  async with self._lock:
161
157
  if response_path.exists():
162
- response_path.unlink()
158
+ response_path.unlink(missing_ok=True)
163
159
 
164
160
  async def update_metadata(self, key: str, response: Response, request: Request, metadata: Metadata) -> None:
165
161
  """
@@ -222,7 +218,7 @@ class AsyncFileStorage(AsyncBaseStorage):
222
218
  if response_path.is_file():
223
219
  age = time.time() - response_path.stat().st_mtime
224
220
  if age > self._ttl:
225
- response_path.unlink()
221
+ response_path.unlink(missing_ok=True)
226
222
  return
227
223
 
228
224
  self._last_cleaned = time.monotonic()
@@ -322,7 +318,7 @@ class AsyncSQLiteStorage(AsyncBaseStorage):
322
318
  assert self._connection
323
319
 
324
320
  if isinstance(key, Response): # pragma: no cover
325
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
321
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
326
322
 
327
323
  async with self._lock:
328
324
  await self._connection.execute("DELETE FROM cache WHERE key = ?", [key])
@@ -459,7 +455,7 @@ class AsyncRedisStorage(AsyncBaseStorage):
459
455
  """
460
456
 
461
457
  if isinstance(key, Response): # pragma: no cover
462
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
458
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
463
459
 
464
460
  await self._client.delete(key)
465
461
 
@@ -507,7 +503,7 @@ class AsyncRedisStorage(AsyncBaseStorage):
507
503
  return self._serializer.loads(cached_response)
508
504
 
509
505
  async def aclose(self) -> None: # pragma: no cover
510
- await self._client.close()
506
+ await self._client.aclose()
511
507
 
512
508
 
513
509
  class AsyncInMemoryStorage(AsyncBaseStorage):
@@ -572,7 +568,7 @@ class AsyncInMemoryStorage(AsyncBaseStorage):
572
568
  """
573
569
 
574
570
  if isinstance(key, Response): # pragma: no cover
575
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
571
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
576
572
 
577
573
  async with self._lock:
578
574
  self._cache.remove_key(key)
@@ -654,6 +650,8 @@ class AsyncS3Storage(AsyncBaseStorage): # pragma: no cover
654
650
  :type check_ttl_every: tp.Union[int, float]
655
651
  :param client: A client for S3, defaults to None
656
652
  :type client: tp.Optional[tp.Any], optional
653
+ :param path_prefix: A path prefix to use for S3 object keys, defaults to "hishel-"
654
+ :type path_prefix: str, optional
657
655
  """
658
656
 
659
657
  def __init__(
@@ -663,6 +661,7 @@ class AsyncS3Storage(AsyncBaseStorage): # pragma: no cover
663
661
  ttl: tp.Optional[tp.Union[int, float]] = None,
664
662
  check_ttl_every: tp.Union[int, float] = 60,
665
663
  client: tp.Optional[tp.Any] = None,
664
+ path_prefix: str = "hishel-",
666
665
  ) -> None:
667
666
  super().__init__(serializer, ttl)
668
667
 
@@ -680,6 +679,7 @@ class AsyncS3Storage(AsyncBaseStorage): # pragma: no cover
680
679
  bucket_name=bucket_name,
681
680
  is_binary=self._serializer.is_binary,
682
681
  check_ttl_every=check_ttl_every,
682
+ path_prefix=path_prefix,
683
683
  )
684
684
  self._lock = AsyncLock()
685
685
 
@@ -716,7 +716,7 @@ class AsyncS3Storage(AsyncBaseStorage): # pragma: no cover
716
716
  """
717
717
 
718
718
  if isinstance(key, Response): # pragma: no cover
719
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
719
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
720
720
 
721
721
  async with self._lock:
722
722
  await self._s3_manager.remove_entry(key)
@@ -8,11 +8,10 @@ import httpx
8
8
  from httpx import AsyncByteStream, Request, Response
9
9
  from httpx._exceptions import ConnectError
10
10
 
11
- from hishel._utils import extract_header_values_decoded, normalized_url
12
-
13
11
  from .._controller import Controller, allowed_stale
14
12
  from .._headers import parse_cache_control
15
13
  from .._serializers import JSONSerializer, Metadata
14
+ from .._utils import extract_header_values_decoded, normalized_url
16
15
  from ._storages import AsyncBaseStorage, AsyncFileStorage
17
16
 
18
17
  if tp.TYPE_CHECKING: # pragma: no cover
@@ -211,11 +210,17 @@ class AsyncCacheTransport(httpx.AsyncBaseTransport):
211
210
  )
212
211
 
213
212
  regular_response = await self._transport.handle_async_request(request)
214
- assert isinstance(regular_response.stream, tp.AsyncIterable)
213
+ try:
214
+ # Prefer already-read content, if available
215
+ stream = fake_stream(regular_response.content)
216
+ except httpx.ResponseNotRead:
217
+ # Fall back to stream if not yet read
218
+ assert isinstance(regular_response.stream, tp.AsyncIterable)
219
+ stream = regular_response.stream
215
220
  httpcore_regular_response = httpcore.Response(
216
221
  status=regular_response.status_code,
217
222
  headers=regular_response.headers.raw,
218
- content=AsyncCacheStream(regular_response.stream),
223
+ content=AsyncCacheStream(stream),
219
224
  extensions=regular_response.extensions,
220
225
  )
221
226
  await httpcore_regular_response.aread()
hishel/_controller.py CHANGED
@@ -3,8 +3,7 @@ import typing as tp
3
3
 
4
4
  from httpcore import Request, Response
5
5
 
6
- from hishel._headers import Vary, parse_cache_control
7
-
6
+ from ._headers import Vary, parse_cache_control
8
7
  from ._utils import (
9
8
  BaseClock,
10
9
  Clock,
@@ -21,7 +20,7 @@ logger = logging.getLogger("hishel.controller")
21
20
  HEURISTICALLY_CACHEABLE_STATUS_CODES = (200, 203, 204, 206, 300, 301, 308, 404, 405, 410, 414, 501)
22
21
  HTTP_METHODS = ["GET", "HEAD", "POST", "PUT", "DELETE", "CONNECT", "OPTIONS", "TRACE", "PATCH"]
23
22
 
24
- __all__ = ("Controller", "HEURISTICALLY_CACHEABLE_STATUS_CODES")
23
+ __all__ = ("HEURISTICALLY_CACHEABLE_STATUS_CODES", "Controller")
25
24
 
26
25
 
27
26
  def get_updated_headers(
@@ -0,0 +1,53 @@
1
+ from contextlib import contextmanager
2
+ from typing import Any, Iterator
3
+
4
+ class Database: ...
5
+
6
+ class Transaction:
7
+ def get(self, key: bytes, *, db: Database | None = None) -> bytes | None:
8
+ """
9
+ Get the value associated with the given key.
10
+ """
11
+ pass
12
+
13
+ def put(self, key: bytes, value: bytes, *, db: Database | None = None, dupdata: bool = False) -> None:
14
+ """
15
+ Put a key-value pair into the database.
16
+ """
17
+ pass
18
+
19
+ def delete(self, key: bytes, *, db: Database | None = None) -> bool:
20
+ """
21
+ Delete the key from the database.
22
+ """
23
+ pass
24
+
25
+ def cursor(self, db: Database) -> Any:
26
+ """
27
+ Create a cursor for iterating over key-value pairs in the database.
28
+ """
29
+ pass
30
+
31
+ class Environment:
32
+ @contextmanager
33
+ def begin(self, *, db: Database | None = None, write: bool = False) -> Iterator[Transaction]:
34
+ """
35
+ Begin a transaction in the environment.
36
+ """
37
+ raise NotImplementedError("It's only for type hinting purposes")
38
+
39
+ def open_db(self, key: bytes, dupsort: bool = False) -> Database:
40
+ """
41
+ Open a database within the environment.
42
+ """
43
+ raise NotImplementedError("It's only for type hinting purposes")
44
+
45
+ def open(
46
+ path: str,
47
+ *,
48
+ max_dbs: int = 0,
49
+ ) -> Environment:
50
+ """
51
+ Open an LMDB environment at the specified path.
52
+ """
53
+ raise NotImplementedError("It's only for type hinting purposes")
hishel/_s3.py CHANGED
@@ -11,16 +11,22 @@ def get_timestamp_in_ms() -> float:
11
11
 
12
12
  class S3Manager:
13
13
  def __init__(
14
- self, client: tp.Any, bucket_name: str, check_ttl_every: tp.Union[int, float], is_binary: bool = False
14
+ self,
15
+ client: tp.Any,
16
+ bucket_name: str,
17
+ check_ttl_every: tp.Union[int, float],
18
+ is_binary: bool = False,
19
+ path_prefix: str = "hishel-",
15
20
  ):
16
21
  self._client = client
17
22
  self._bucket_name = bucket_name
18
23
  self._is_binary = is_binary
19
24
  self._last_cleaned = time.monotonic()
20
25
  self._check_ttl_every = check_ttl_every
26
+ self._path_prefix = path_prefix
21
27
 
22
28
  def write_to(self, path: str, data: tp.Union[bytes, str], only_metadata: bool = False) -> None:
23
- path = "hishel-" + path
29
+ path = self._path_prefix + path
24
30
  if isinstance(data, str):
25
31
  data = data.encode("utf-8")
26
32
 
@@ -43,7 +49,7 @@ class S3Manager:
43
49
  )
44
50
 
45
51
  def read_from(self, path: str) -> tp.Union[bytes, str]:
46
- path = "hishel-" + path
52
+ path = self._path_prefix + path
47
53
  response = self._client.get_object(
48
54
  Bucket=self._bucket_name,
49
55
  Key=path,
@@ -57,7 +63,7 @@ class S3Manager:
57
63
  return tp.cast(str, content.decode("utf-8"))
58
64
 
59
65
  def remove_expired(self, ttl: int, key: str) -> None:
60
- path = "hishel-" + key
66
+ path = self._path_prefix + key
61
67
 
62
68
  if time.monotonic() - self._last_cleaned < self._check_ttl_every:
63
69
  try:
@@ -72,7 +78,7 @@ class S3Manager:
72
78
 
73
79
  self._last_cleaned = time.monotonic()
74
80
  for obj in self._client.list_objects(Bucket=self._bucket_name).get("Contents", []):
75
- if not obj["Key"].startswith("hishel-"): # pragma: no cover
81
+ if not obj["Key"].startswith(self._path_prefix): # pragma: no cover
76
82
  continue
77
83
 
78
84
  try:
@@ -88,15 +94,20 @@ class S3Manager:
88
94
  self._client.delete_object(Bucket=self._bucket_name, Key=obj["Key"])
89
95
 
90
96
  def remove_entry(self, key: str) -> None:
91
- path = "hishel-" + key
97
+ path = self._path_prefix + key
92
98
  self._client.delete_object(Bucket=self._bucket_name, Key=path)
93
99
 
94
100
 
95
101
  class AsyncS3Manager: # pragma: no cover
96
102
  def __init__(
97
- self, client: tp.Any, bucket_name: str, check_ttl_every: tp.Union[int, float], is_binary: bool = False
103
+ self,
104
+ client: tp.Any,
105
+ bucket_name: str,
106
+ check_ttl_every: tp.Union[int, float],
107
+ is_binary: bool = False,
108
+ path_prefix: str = "hishel-",
98
109
  ):
99
- self._sync_manager = S3Manager(client, bucket_name, check_ttl_every, is_binary)
110
+ self._sync_manager = S3Manager(client, bucket_name, check_ttl_every, is_binary, path_prefix)
100
111
 
101
112
  async def write_to(self, path: str, data: tp.Union[bytes, str], only_metadata: bool = False) -> None:
102
113
  return await to_thread.run_sync(self._sync_manager.write_to, path, data, only_metadata)
hishel/_serializers.py CHANGED
@@ -6,7 +6,7 @@ from datetime import datetime
6
6
 
7
7
  from httpcore import Request, Response
8
8
 
9
- from hishel._utils import normalized_url
9
+ from ._utils import normalized_url
10
10
 
11
11
  try:
12
12
  import yaml
@@ -17,7 +17,7 @@ HEADERS_ENCODING = "iso-8859-1"
17
17
  KNOWN_RESPONSE_EXTENSIONS = ("http_version", "reason_phrase")
18
18
  KNOWN_REQUEST_EXTENSIONS = ("timeout", "sni_hostname")
19
19
 
20
- __all__ = ("PickleSerializer", "JSONSerializer", "YAMLSerializer", "BaseSerializer", "clone_model")
20
+ __all__ = ("BaseSerializer", "JSONSerializer", "Metadata", "PickleSerializer", "YAMLSerializer", "clone_model")
21
21
 
22
22
  T = tp.TypeVar("T", Request, Response)
23
23
 
hishel/_sync/_client.py CHANGED
@@ -2,7 +2,7 @@ import typing as tp
2
2
 
3
3
  import httpx
4
4
 
5
- from hishel._sync._transports import CacheTransport
5
+ from ._transports import CacheTransport
6
6
 
7
7
  __all__ = ("CacheClient",)
8
8
 
hishel/_sync/_storages.py CHANGED
@@ -4,7 +4,6 @@ import datetime
4
4
  import logging
5
5
  import os
6
6
  import time
7
- import typing as t
8
7
  import typing as tp
9
8
  import warnings
10
9
  from copy import deepcopy
@@ -24,13 +23,11 @@ except ImportError: # pragma: no cover
24
23
 
25
24
  from httpcore import Request, Response
26
25
 
27
- if t.TYPE_CHECKING: # pragma: no cover
26
+ if tp.TYPE_CHECKING: # pragma: no cover
28
27
  from typing_extensions import TypeAlias
29
28
 
30
- from hishel._serializers import BaseSerializer, clone_model
31
-
32
29
  from .._files import FileManager
33
- from .._serializers import JSONSerializer, Metadata
30
+ from .._serializers import BaseSerializer, JSONSerializer, Metadata, clone_model
34
31
  from .._synchronization import Lock
35
32
  from .._utils import float_seconds_to_int_milliseconds
36
33
 
@@ -39,10 +36,10 @@ logger = logging.getLogger("hishel.storages")
39
36
  __all__ = (
40
37
  "BaseStorage",
41
38
  "FileStorage",
42
- "RedisStorage",
43
- "SQLiteStorage",
44
39
  "InMemoryStorage",
40
+ "RedisStorage",
45
41
  "S3Storage",
42
+ "SQLiteStorage",
46
43
  )
47
44
 
48
45
  StoredResponse: TypeAlias = tp.Tuple[Response, Request, Metadata]
@@ -106,8 +103,7 @@ class FileStorage(BaseStorage):
106
103
  self._base_path = Path(base_path) if base_path is not None else Path(".cache/hishel")
107
104
  self._gitignore_file = self._base_path / ".gitignore"
108
105
 
109
- if not self._base_path.is_dir():
110
- self._base_path.mkdir(parents=True)
106
+ self._base_path.mkdir(parents=True, exist_ok=True)
111
107
 
112
108
  if not self._gitignore_file.is_file():
113
109
  with open(self._gitignore_file, "w", encoding="utf-8") as f:
@@ -153,13 +149,13 @@ class FileStorage(BaseStorage):
153
149
  """
154
150
 
155
151
  if isinstance(key, Response): # pragma: no cover
156
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
152
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
157
153
 
158
154
  response_path = self._base_path / key
159
155
 
160
156
  with self._lock:
161
157
  if response_path.exists():
162
- response_path.unlink()
158
+ response_path.unlink(missing_ok=True)
163
159
 
164
160
  def update_metadata(self, key: str, response: Response, request: Request, metadata: Metadata) -> None:
165
161
  """
@@ -222,7 +218,7 @@ class FileStorage(BaseStorage):
222
218
  if response_path.is_file():
223
219
  age = time.time() - response_path.stat().st_mtime
224
220
  if age > self._ttl:
225
- response_path.unlink()
221
+ response_path.unlink(missing_ok=True)
226
222
  return
227
223
 
228
224
  self._last_cleaned = time.monotonic()
@@ -322,7 +318,7 @@ class SQLiteStorage(BaseStorage):
322
318
  assert self._connection
323
319
 
324
320
  if isinstance(key, Response): # pragma: no cover
325
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
321
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
326
322
 
327
323
  with self._lock:
328
324
  self._connection.execute("DELETE FROM cache WHERE key = ?", [key])
@@ -459,7 +455,7 @@ class RedisStorage(BaseStorage):
459
455
  """
460
456
 
461
457
  if isinstance(key, Response): # pragma: no cover
462
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
458
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
463
459
 
464
460
  self._client.delete(key)
465
461
 
@@ -572,7 +568,7 @@ class InMemoryStorage(BaseStorage):
572
568
  """
573
569
 
574
570
  if isinstance(key, Response): # pragma: no cover
575
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
571
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
576
572
 
577
573
  with self._lock:
578
574
  self._cache.remove_key(key)
@@ -654,6 +650,8 @@ class S3Storage(BaseStorage): # pragma: no cover
654
650
  :type check_ttl_every: tp.Union[int, float]
655
651
  :param client: A client for S3, defaults to None
656
652
  :type client: tp.Optional[tp.Any], optional
653
+ :param path_prefix: A path prefix to use for S3 object keys, defaults to "hishel-"
654
+ :type path_prefix: str, optional
657
655
  """
658
656
 
659
657
  def __init__(
@@ -663,6 +661,7 @@ class S3Storage(BaseStorage): # pragma: no cover
663
661
  ttl: tp.Optional[tp.Union[int, float]] = None,
664
662
  check_ttl_every: tp.Union[int, float] = 60,
665
663
  client: tp.Optional[tp.Any] = None,
664
+ path_prefix: str = "hishel-",
666
665
  ) -> None:
667
666
  super().__init__(serializer, ttl)
668
667
 
@@ -680,6 +679,7 @@ class S3Storage(BaseStorage): # pragma: no cover
680
679
  bucket_name=bucket_name,
681
680
  is_binary=self._serializer.is_binary,
682
681
  check_ttl_every=check_ttl_every,
682
+ path_prefix=path_prefix,
683
683
  )
684
684
  self._lock = Lock()
685
685
 
@@ -716,7 +716,7 @@ class S3Storage(BaseStorage): # pragma: no cover
716
716
  """
717
717
 
718
718
  if isinstance(key, Response): # pragma: no cover
719
- key = t.cast(str, key.extensions["cache_metadata"]["cache_key"])
719
+ key = tp.cast(str, key.extensions["cache_metadata"]["cache_key"])
720
720
 
721
721
  with self._lock:
722
722
  self._s3_manager.remove_entry(key)
@@ -8,11 +8,10 @@ import httpx
8
8
  from httpx import SyncByteStream, Request, Response
9
9
  from httpx._exceptions import ConnectError
10
10
 
11
- from hishel._utils import extract_header_values_decoded, normalized_url
12
-
13
11
  from .._controller import Controller, allowed_stale
14
12
  from .._headers import parse_cache_control
15
13
  from .._serializers import JSONSerializer, Metadata
14
+ from .._utils import extract_header_values_decoded, normalized_url
16
15
  from ._storages import BaseStorage, FileStorage
17
16
 
18
17
  if tp.TYPE_CHECKING: # pragma: no cover
@@ -211,11 +210,17 @@ class CacheTransport(httpx.BaseTransport):
211
210
  )
212
211
 
213
212
  regular_response = self._transport.handle_request(request)
214
- assert isinstance(regular_response.stream, tp.Iterable)
213
+ try:
214
+ # Prefer already-read content, if available
215
+ stream = fake_stream(regular_response.content)
216
+ except httpx.ResponseNotRead:
217
+ # Fall back to stream if not yet read
218
+ assert isinstance(regular_response.stream, tp.Iterable)
219
+ stream = regular_response.stream
215
220
  httpcore_regular_response = httpcore.Response(
216
221
  status=regular_response.status_code,
217
222
  headers=regular_response.headers.raw,
218
- content=CacheStream(regular_response.stream),
223
+ content=CacheStream(stream),
219
224
  extensions=regular_response.extensions,
220
225
  )
221
226
  httpcore_regular_response.read()