elasticsearch9 9.0.3__py3-none-any.whl → 9.0.4__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -22,6 +22,9 @@ from elastic_transport import ObjectApiResponse
22
22
  from ._base import NamespacedClient
23
23
  from .utils import SKIP_IN_PATH, _quote, _rewrite_parameters
24
24
 
25
+ if t.TYPE_CHECKING:
26
+ from elasticsearch.esql import ESQLBase
27
+
25
28
 
26
29
  class EsqlClient(NamespacedClient):
27
30
 
@@ -44,7 +47,7 @@ class EsqlClient(NamespacedClient):
44
47
  def async_query(
45
48
  self,
46
49
  *,
47
- query: t.Optional[str] = None,
50
+ query: t.Optional[t.Union[str, "ESQLBase"]] = None,
48
51
  columnar: t.Optional[bool] = None,
49
52
  delimiter: t.Optional[str] = None,
50
53
  drop_null_columns: t.Optional[bool] = None,
@@ -99,7 +102,12 @@ class EsqlClient(NamespacedClient):
99
102
  which has the name of all the columns.
100
103
  :param filter: Specify a Query DSL query in the filter parameter to filter the
101
104
  set of documents that an ES|QL query runs on.
102
- :param format: A short version of the Accept header, for example `json` or `yaml`.
105
+ :param format: A short version of the Accept header, e.g. json, yaml. `csv`,
106
+ `tsv`, and `txt` formats will return results in a tabular format, excluding
107
+ other metadata fields from the response. For async requests, nothing will
108
+ be returned if the async query doesn't finish within the timeout. The query
109
+ ID and running status are available in the `X-Elasticsearch-Async-Id` and
110
+ `X-Elasticsearch-Async-Is-Running` HTTP headers of the response, respectively.
103
111
  :param include_ccs_metadata: When set to `true` and performing a cross-cluster
104
112
  query, the response will include an extra `_clusters` object with information
105
113
  about the clusters that participated in the search along with info such as
@@ -151,7 +159,7 @@ class EsqlClient(NamespacedClient):
151
159
  __query["pretty"] = pretty
152
160
  if not __body:
153
161
  if query is not None:
154
- __body["query"] = query
162
+ __body["query"] = str(query)
155
163
  if columnar is not None:
156
164
  __body["columnar"] = columnar
157
165
  if filter is not None:
@@ -389,7 +397,7 @@ class EsqlClient(NamespacedClient):
389
397
  def query(
390
398
  self,
391
399
  *,
392
- query: t.Optional[str] = None,
400
+ query: t.Optional[t.Union[str, "ESQLBase"]] = None,
393
401
  columnar: t.Optional[bool] = None,
394
402
  delimiter: t.Optional[str] = None,
395
403
  drop_null_columns: t.Optional[bool] = None,
@@ -438,7 +446,9 @@ class EsqlClient(NamespacedClient):
438
446
  `all_columns` which has the name of all columns.
439
447
  :param filter: Specify a Query DSL query in the filter parameter to filter the
440
448
  set of documents that an ES|QL query runs on.
441
- :param format: A short version of the Accept header, e.g. json, yaml.
449
+ :param format: A short version of the Accept header, e.g. json, yaml. `csv`,
450
+ `tsv`, and `txt` formats will return results in a tabular format, excluding
451
+ other metadata fields from the response.
442
452
  :param include_ccs_metadata: When set to `true` and performing a cross-cluster
443
453
  query, the response will include an extra `_clusters` object with information
444
454
  about the clusters that participated in the search along with info such as
@@ -476,7 +486,7 @@ class EsqlClient(NamespacedClient):
476
486
  __query["pretty"] = pretty
477
487
  if not __body:
478
488
  if query is not None:
479
- __body["query"] = query
489
+ __body["query"] = str(query)
480
490
  if columnar is not None:
481
491
  __body["columnar"] = columnar
482
492
  if filter is not None:
@@ -3715,7 +3715,7 @@ class IndicesClient(NamespacedClient):
3715
3715
  <li>Change a field's mapping using reindexing</li>
3716
3716
  <li>Rename a field using a field alias</li>
3717
3717
  </ul>
3718
- <p>Learn how to use the update mapping API with practical examples in the <a href="https://www.elastic.co/docs//manage-data/data-store/mapping/update-mappings-examples">Update mapping API examples</a> guide.</p>
3718
+ <p>Learn how to use the update mapping API with practical examples in the <a href="https://www.elastic.co/docs/manage-data/data-store/mapping/update-mappings-examples">Update mapping API examples</a> guide.</p>
3719
3719
 
3720
3720
 
3721
3721
  `<https://www.elastic.co/docs/api/doc/elasticsearch/v9/operation/operation-indices-put-mapping>`_
@@ -4587,7 +4587,7 @@ class IndicesClient(NamespacedClient):
4587
4587
  .. raw:: html
4588
4588
 
4589
4589
  <p>Roll over to a new index.
4590
- TIP: It is recommended to use the index lifecycle rollover action to automate rollovers.</p>
4590
+ TIP: We recommend using the index lifecycle rollover action to automate rollovers. However, Serverless does not support Index Lifecycle Management (ILM), so don't use this approach in the Serverless context.</p>
4591
4591
  <p>The rollover API creates a new index for a data stream or index alias.
4592
4592
  The API behavior depends on the rollover target.</p>
4593
4593
  <p><strong>Roll over a data stream</strong></p>
@@ -283,7 +283,7 @@ class SqlClient(NamespacedClient):
283
283
  keep_alive: t.Optional[t.Union[str, t.Literal[-1], t.Literal[0]]] = None,
284
284
  keep_on_completion: t.Optional[bool] = None,
285
285
  page_timeout: t.Optional[t.Union[str, t.Literal[-1], t.Literal[0]]] = None,
286
- params: t.Optional[t.Mapping[str, t.Any]] = None,
286
+ params: t.Optional[t.Sequence[t.Any]] = None,
287
287
  pretty: t.Optional[bool] = None,
288
288
  query: t.Optional[str] = None,
289
289
  request_timeout: t.Optional[t.Union[str, t.Literal[-1], t.Literal[0]]] = None,
@@ -602,6 +602,66 @@ class TransformClient(NamespacedClient):
602
602
  path_parts=__path_parts,
603
603
  )
604
604
 
605
+ @_rewrite_parameters()
606
+ def set_upgrade_mode(
607
+ self,
608
+ *,
609
+ enabled: t.Optional[bool] = None,
610
+ error_trace: t.Optional[bool] = None,
611
+ filter_path: t.Optional[t.Union[str, t.Sequence[str]]] = None,
612
+ human: t.Optional[bool] = None,
613
+ pretty: t.Optional[bool] = None,
614
+ timeout: t.Optional[t.Union[str, t.Literal[-1], t.Literal[0]]] = None,
615
+ ) -> ObjectApiResponse[t.Any]:
616
+ """
617
+ .. raw:: html
618
+
619
+ <p>Set upgrade_mode for transform indices.
620
+ Sets a cluster wide upgrade_mode setting that prepares transform
621
+ indices for an upgrade.
622
+ When upgrading your cluster, in some circumstances you must restart your
623
+ nodes and reindex your transform indices. In those circumstances,
624
+ there must be no transforms running. You can close the transforms,
625
+ do the upgrade, then open all the transforms again. Alternatively,
626
+ you can use this API to temporarily halt tasks associated with the transforms
627
+ and prevent new transforms from opening. You can also use this API
628
+ during upgrades that do not require you to reindex your transform
629
+ indices, though stopping transforms is not a requirement in that case.
630
+ You can see the current value for the upgrade_mode setting by using the get
631
+ transform info API.</p>
632
+
633
+
634
+ `<https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-transform-set-upgrade-mode>`_
635
+
636
+ :param enabled: When `true`, it enables `upgrade_mode` which temporarily halts
637
+ all transform tasks and prohibits new transform tasks from starting.
638
+ :param timeout: The time to wait for the request to be completed.
639
+ """
640
+ __path_parts: t.Dict[str, str] = {}
641
+ __path = "/_transform/set_upgrade_mode"
642
+ __query: t.Dict[str, t.Any] = {}
643
+ if enabled is not None:
644
+ __query["enabled"] = enabled
645
+ if error_trace is not None:
646
+ __query["error_trace"] = error_trace
647
+ if filter_path is not None:
648
+ __query["filter_path"] = filter_path
649
+ if human is not None:
650
+ __query["human"] = human
651
+ if pretty is not None:
652
+ __query["pretty"] = pretty
653
+ if timeout is not None:
654
+ __query["timeout"] = timeout
655
+ __headers = {"accept": "application/json"}
656
+ return self.perform_request( # type: ignore[return-value]
657
+ "POST",
658
+ __path,
659
+ params=__query,
660
+ headers=__headers,
661
+ endpoint_id="transform.set_upgrade_mode",
662
+ path_parts=__path_parts,
663
+ )
664
+
605
665
  @_rewrite_parameters(
606
666
  parameter_aliases={"from": "from_"},
607
667
  )
@@ -15,4 +15,4 @@
15
15
  # specific language governing permissions and limitations
16
16
  # under the License.
17
17
 
18
- __versionstr__ = "9.0.3"
18
+ __versionstr__ = "9.0.4"
@@ -20,6 +20,7 @@ from typing import (
20
20
  TYPE_CHECKING,
21
21
  Any,
22
22
  AsyncIterable,
23
+ AsyncIterator,
23
24
  Dict,
24
25
  List,
25
26
  Optional,
@@ -42,6 +43,7 @@ from .search import AsyncSearch
42
43
 
43
44
  if TYPE_CHECKING:
44
45
  from elasticsearch import AsyncElasticsearch
46
+ from elasticsearch.esql.esql import ESQLBase
45
47
 
46
48
 
47
49
  class AsyncIndexMeta(DocumentMeta):
@@ -520,3 +522,85 @@ class AsyncDocument(DocumentBase, metaclass=AsyncIndexMeta):
520
522
  return action
521
523
 
522
524
  return await async_bulk(es, Generate(actions), **kwargs)
525
+
526
+ @classmethod
527
+ async def esql_execute(
528
+ cls,
529
+ query: "ESQLBase",
530
+ return_additional: bool = False,
531
+ ignore_missing_fields: bool = False,
532
+ using: Optional[AsyncUsingType] = None,
533
+ **kwargs: Any,
534
+ ) -> AsyncIterator[Union[Self, Tuple[Self, Dict[str, Any]]]]:
535
+ """
536
+ Execute the given ES|QL query and return an iterator of 2-element tuples,
537
+ where the first element is an instance of this ``Document`` and the
538
+ second a dictionary with any remaining columns requested in the query.
539
+
540
+ :arg query: an ES|QL query object created with the ``esql_from()`` method.
541
+ :arg return_additional: if ``False`` (the default), this method returns
542
+ document objects. If set to ``True``, the method returns tuples with
543
+ a document in the first element and a dictionary with any additional
544
+ columns returned by the query in the second element.
545
+ :arg ignore_missing_fields: if ``False`` (the default), all the fields of
546
+ the document must be present in the query, or else an exception is
547
+ raised. Set to ``True`` to allow missing fields, which will result in
548
+ partially initialized document objects.
549
+ :arg using: connection alias to use, defaults to ``'default'``
550
+ :arg kwargs: additional options for the ``client.esql.query()`` function.
551
+ """
552
+ es = cls._get_connection(using)
553
+ response = await es.esql.query(query=str(query), **kwargs)
554
+ query_columns = [col["name"] for col in response.body.get("columns", [])]
555
+
556
+ # Here we get the list of columns defined in the document, which are the
557
+ # columns that we will take from each result to assemble the document
558
+ # object.
559
+ # When `for_esql=False` is passed below by default, the list will include
560
+ # nested fields, which ES|QL does not return, causing an error. When passing
561
+ # `ignore_missing_fields=True` the list will be generated with
562
+ # `for_esql=True`, so the error will not occur, but the documents will
563
+ # not have any Nested objects in them.
564
+ doc_fields = set(cls._get_field_names(for_esql=ignore_missing_fields))
565
+ if not ignore_missing_fields and not doc_fields.issubset(set(query_columns)):
566
+ raise ValueError(
567
+ f"Not all fields of {cls.__name__} were returned by the query. "
568
+ "Make sure your document does not use Nested fields, which are "
569
+ "currently not supported in ES|QL. To force the query to be "
570
+ "evaluated in spite of the missing fields, pass set the "
571
+ "ignore_missing_fields=True option in the esql_execute() call."
572
+ )
573
+ non_doc_fields: set[str] = set(query_columns) - doc_fields - {"_id"}
574
+ index_id = query_columns.index("_id")
575
+
576
+ results = response.body.get("values", [])
577
+ for column_values in results:
578
+ # create a dictionary with all the document fields, expanding the
579
+ # dot notation returned by ES|QL into the recursive dictionaries
580
+ # used by Document.from_dict()
581
+ doc_dict: Dict[str, Any] = {}
582
+ for col, val in zip(query_columns, column_values):
583
+ if col in doc_fields:
584
+ cols = col.split(".")
585
+ d = doc_dict
586
+ for c in cols[:-1]:
587
+ if c not in d:
588
+ d[c] = {}
589
+ d = d[c]
590
+ d[cols[-1]] = val
591
+
592
+ # create the document instance
593
+ obj = cls(meta={"_id": column_values[index_id]})
594
+ obj._from_dict(doc_dict)
595
+
596
+ if return_additional:
597
+ # build a dict with any other values included in the response
598
+ other = {
599
+ col: val
600
+ for col, val in zip(query_columns, column_values)
601
+ if col in non_doc_fields
602
+ }
603
+
604
+ yield obj, other
605
+ else:
606
+ yield obj
@@ -21,6 +21,7 @@ from typing import (
21
21
  Any,
22
22
  Dict,
23
23
  Iterable,
24
+ Iterator,
24
25
  List,
25
26
  Optional,
26
27
  Tuple,
@@ -42,6 +43,7 @@ from .search import Search
42
43
 
43
44
  if TYPE_CHECKING:
44
45
  from elasticsearch import Elasticsearch
46
+ from elasticsearch.esql.esql import ESQLBase
45
47
 
46
48
 
47
49
  class IndexMeta(DocumentMeta):
@@ -512,3 +514,85 @@ class Document(DocumentBase, metaclass=IndexMeta):
512
514
  return action
513
515
 
514
516
  return bulk(es, Generate(actions), **kwargs)
517
+
518
+ @classmethod
519
+ def esql_execute(
520
+ cls,
521
+ query: "ESQLBase",
522
+ return_additional: bool = False,
523
+ ignore_missing_fields: bool = False,
524
+ using: Optional[UsingType] = None,
525
+ **kwargs: Any,
526
+ ) -> Iterator[Union[Self, Tuple[Self, Dict[str, Any]]]]:
527
+ """
528
+ Execute the given ES|QL query and return an iterator of 2-element tuples,
529
+ where the first element is an instance of this ``Document`` and the
530
+ second a dictionary with any remaining columns requested in the query.
531
+
532
+ :arg query: an ES|QL query object created with the ``esql_from()`` method.
533
+ :arg return_additional: if ``False`` (the default), this method returns
534
+ document objects. If set to ``True``, the method returns tuples with
535
+ a document in the first element and a dictionary with any additional
536
+ columns returned by the query in the second element.
537
+ :arg ignore_missing_fields: if ``False`` (the default), all the fields of
538
+ the document must be present in the query, or else an exception is
539
+ raised. Set to ``True`` to allow missing fields, which will result in
540
+ partially initialized document objects.
541
+ :arg using: connection alias to use, defaults to ``'default'``
542
+ :arg kwargs: additional options for the ``client.esql.query()`` function.
543
+ """
544
+ es = cls._get_connection(using)
545
+ response = es.esql.query(query=str(query), **kwargs)
546
+ query_columns = [col["name"] for col in response.body.get("columns", [])]
547
+
548
+ # Here we get the list of columns defined in the document, which are the
549
+ # columns that we will take from each result to assemble the document
550
+ # object.
551
+ # When `for_esql=False` is passed below by default, the list will include
552
+ # nested fields, which ES|QL does not return, causing an error. When passing
553
+ # `ignore_missing_fields=True` the list will be generated with
554
+ # `for_esql=True`, so the error will not occur, but the documents will
555
+ # not have any Nested objects in them.
556
+ doc_fields = set(cls._get_field_names(for_esql=ignore_missing_fields))
557
+ if not ignore_missing_fields and not doc_fields.issubset(set(query_columns)):
558
+ raise ValueError(
559
+ f"Not all fields of {cls.__name__} were returned by the query. "
560
+ "Make sure your document does not use Nested fields, which are "
561
+ "currently not supported in ES|QL. To force the query to be "
562
+ "evaluated in spite of the missing fields, pass set the "
563
+ "ignore_missing_fields=True option in the esql_execute() call."
564
+ )
565
+ non_doc_fields: set[str] = set(query_columns) - doc_fields - {"_id"}
566
+ index_id = query_columns.index("_id")
567
+
568
+ results = response.body.get("values", [])
569
+ for column_values in results:
570
+ # create a dictionary with all the document fields, expanding the
571
+ # dot notation returned by ES|QL into the recursive dictionaries
572
+ # used by Document.from_dict()
573
+ doc_dict: Dict[str, Any] = {}
574
+ for col, val in zip(query_columns, column_values):
575
+ if col in doc_fields:
576
+ cols = col.split(".")
577
+ d = doc_dict
578
+ for c in cols[:-1]:
579
+ if c not in d:
580
+ d[c] = {}
581
+ d = d[c]
582
+ d[cols[-1]] = val
583
+
584
+ # create the document instance
585
+ obj = cls(meta={"_id": column_values[index_id]})
586
+ obj._from_dict(doc_dict)
587
+
588
+ if return_additional:
589
+ # build a dict with any other values included in the response
590
+ other = {
591
+ col: val
592
+ for col, val in zip(query_columns, column_values)
593
+ if col in non_doc_fields
594
+ }
595
+
596
+ yield obj, other
597
+ else:
598
+ yield obj
@@ -28,6 +28,7 @@ from typing import (
28
28
  List,
29
29
  Optional,
30
30
  Tuple,
31
+ Type,
31
32
  TypeVar,
32
33
  Union,
33
34
  get_args,
@@ -49,6 +50,7 @@ from .utils import DOC_META_FIELDS, ObjectBase
49
50
  if TYPE_CHECKING:
50
51
  from elastic_transport import ObjectApiResponse
51
52
 
53
+ from ..esql.esql import ESQLBase
52
54
  from .index_base import IndexBase
53
55
 
54
56
 
@@ -602,3 +604,44 @@ class DocumentBase(ObjectBase):
602
604
 
603
605
  meta["_source"] = d
604
606
  return meta
607
+
608
+ @classmethod
609
+ def _get_field_names(
610
+ cls, for_esql: bool = False, nested_class: Optional[Type[InnerDoc]] = None
611
+ ) -> List[str]:
612
+ """Return the list of field names used by this document.
613
+ If the document has nested objects, their fields are reported using dot
614
+ notation. If the ``for_esql`` argument is set to ``True``, the list omits
615
+ nested fields, which are currently unsupported in ES|QL.
616
+ """
617
+ fields = []
618
+ class_ = nested_class or cls
619
+ for field_name in class_._doc_type.mapping:
620
+ field = class_._doc_type.mapping[field_name]
621
+ if isinstance(field, Object):
622
+ if for_esql and isinstance(field, Nested):
623
+ # ES|QL does not recognize Nested fields at this time
624
+ continue
625
+ sub_fields = cls._get_field_names(
626
+ for_esql=for_esql, nested_class=field._doc_class
627
+ )
628
+ for sub_field in sub_fields:
629
+ fields.append(f"{field_name}.{sub_field}")
630
+ else:
631
+ fields.append(field_name)
632
+ return fields
633
+
634
+ @classmethod
635
+ def esql_from(cls) -> "ESQLBase":
636
+ """Return a base ES|QL query for instances of this document class.
637
+
638
+ The returned query is initialized with ``FROM`` and ``KEEP`` statements,
639
+ and can be completed as desired.
640
+ """
641
+ from ..esql import ESQL # here to avoid circular imports
642
+
643
+ return (
644
+ ESQL.from_(cls)
645
+ .metadata("_id")
646
+ .keep("_id", *tuple(cls._get_field_names(for_esql=True)))
647
+ )
@@ -119,9 +119,16 @@ class Field(DslBase):
119
119
  def __getitem__(self, subfield: str) -> "Field":
120
120
  return cast(Field, self._params.get("fields", {})[subfield])
121
121
 
122
- def _serialize(self, data: Any) -> Any:
122
+ def _serialize(self, data: Any, skip_empty: bool) -> Any:
123
123
  return data
124
124
 
125
+ def _safe_serialize(self, data: Any, skip_empty: bool) -> Any:
126
+ try:
127
+ return self._serialize(data, skip_empty)
128
+ except TypeError:
129
+ # older method signature, without skip_empty
130
+ return self._serialize(data) # type: ignore[call-arg]
131
+
125
132
  def _deserialize(self, data: Any) -> Any:
126
133
  return data
127
134
 
@@ -133,10 +140,16 @@ class Field(DslBase):
133
140
  return AttrList([])
134
141
  return self._empty()
135
142
 
136
- def serialize(self, data: Any) -> Any:
143
+ def serialize(self, data: Any, skip_empty: bool = True) -> Any:
137
144
  if isinstance(data, (list, AttrList, tuple)):
138
- return list(map(self._serialize, cast(Iterable[Any], data)))
139
- return self._serialize(data)
145
+ return list(
146
+ map(
147
+ self._safe_serialize,
148
+ cast(Iterable[Any], data),
149
+ [skip_empty] * len(data),
150
+ )
151
+ )
152
+ return self._safe_serialize(data, skip_empty)
140
153
 
141
154
  def deserialize(self, data: Any) -> Any:
142
155
  if isinstance(data, (list, AttrList, tuple)):
@@ -186,7 +199,7 @@ class RangeField(Field):
186
199
  data = {k: self._core_field.deserialize(v) for k, v in data.items()} # type: ignore[union-attr]
187
200
  return Range(data)
188
201
 
189
- def _serialize(self, data: Any) -> Optional[Dict[str, Any]]:
202
+ def _serialize(self, data: Any, skip_empty: bool) -> Optional[Dict[str, Any]]:
190
203
  if data is None:
191
204
  return None
192
205
  if not isinstance(data, collections.abc.Mapping):
@@ -550,7 +563,7 @@ class Object(Field):
550
563
  return self._wrap(data)
551
564
 
552
565
  def _serialize(
553
- self, data: Optional[Union[Dict[str, Any], "InnerDoc"]]
566
+ self, data: Optional[Union[Dict[str, Any], "InnerDoc"]], skip_empty: bool
554
567
  ) -> Optional[Dict[str, Any]]:
555
568
  if data is None:
556
569
  return None
@@ -559,7 +572,7 @@ class Object(Field):
559
572
  if isinstance(data, collections.abc.Mapping):
560
573
  return data
561
574
 
562
- return data.to_dict()
575
+ return data.to_dict(skip_empty=skip_empty)
563
576
 
564
577
  def clean(self, data: Any) -> Any:
565
578
  data = super().clean(data)
@@ -768,7 +781,7 @@ class Binary(Field):
768
781
  def _deserialize(self, data: Any) -> bytes:
769
782
  return base64.b64decode(data)
770
783
 
771
- def _serialize(self, data: Any) -> Optional[str]:
784
+ def _serialize(self, data: Any, skip_empty: bool) -> Optional[str]:
772
785
  if data is None:
773
786
  return None
774
787
  return base64.b64encode(data).decode()
@@ -2619,7 +2632,7 @@ class Ip(Field):
2619
2632
  # the ipaddress library for pypy only accepts unicode.
2620
2633
  return ipaddress.ip_address(unicode(data))
2621
2634
 
2622
- def _serialize(self, data: Any) -> Optional[str]:
2635
+ def _serialize(self, data: Any, skip_empty: bool) -> Optional[str]:
2623
2636
  if data is None:
2624
2637
  return None
2625
2638
  return str(data)
@@ -3367,7 +3380,7 @@ class Percolator(Field):
3367
3380
  def _deserialize(self, data: Any) -> "Query":
3368
3381
  return Q(data) # type: ignore[no-any-return]
3369
3382
 
3370
- def _serialize(self, data: Any) -> Optional[Dict[str, Any]]:
3383
+ def _serialize(self, data: Any, skip_empty: bool) -> Optional[Dict[str, Any]]:
3371
3384
  if data is None:
3372
3385
  return None
3373
3386
  return data.to_dict() # type: ignore[no-any-return]
@@ -63,7 +63,7 @@ class BucketData(AggResponse[_R]):
63
63
  )
64
64
 
65
65
  def __iter__(self) -> Iterator["Agg"]: # type: ignore[override]
66
- return iter(self.buckets) # type: ignore[arg-type]
66
+ return iter(self.buckets)
67
67
 
68
68
  def __len__(self) -> int:
69
69
  return len(self.buckets)
@@ -4390,7 +4390,7 @@ class ArrayPercentilesItem(AttrDict[Any]):
4390
4390
  :arg value_as_string:
4391
4391
  """
4392
4392
 
4393
- key: str
4393
+ key: float
4394
4394
  value: Union[float, None]
4395
4395
  value_as_string: str
4396
4396
 
@@ -5236,7 +5236,9 @@ class HdrPercentileRanksAggregate(AttrDict[Any]):
5236
5236
  :arg meta:
5237
5237
  """
5238
5238
 
5239
- values: Union[Mapping[str, Union[str, int, None]], Sequence["ArrayPercentilesItem"]]
5239
+ values: Union[
5240
+ Mapping[str, Union[str, float, None]], Sequence["ArrayPercentilesItem"]
5241
+ ]
5240
5242
  meta: Mapping[str, Any]
5241
5243
 
5242
5244
 
@@ -5246,7 +5248,9 @@ class HdrPercentilesAggregate(AttrDict[Any]):
5246
5248
  :arg meta:
5247
5249
  """
5248
5250
 
5249
- values: Union[Mapping[str, Union[str, int, None]], Sequence["ArrayPercentilesItem"]]
5251
+ values: Union[
5252
+ Mapping[str, Union[str, float, None]], Sequence["ArrayPercentilesItem"]
5253
+ ]
5250
5254
  meta: Mapping[str, Any]
5251
5255
 
5252
5256
 
@@ -5753,7 +5757,9 @@ class PercentilesBucketAggregate(AttrDict[Any]):
5753
5757
  :arg meta:
5754
5758
  """
5755
5759
 
5756
- values: Union[Mapping[str, Union[str, int, None]], Sequence["ArrayPercentilesItem"]]
5760
+ values: Union[
5761
+ Mapping[str, Union[str, float, None]], Sequence["ArrayPercentilesItem"]
5762
+ ]
5757
5763
  meta: Mapping[str, Any]
5758
5764
 
5759
5765
 
@@ -5954,17 +5960,19 @@ class SearchProfile(AttrDict[Any]):
5954
5960
  class ShardFailure(AttrDict[Any]):
5955
5961
  """
5956
5962
  :arg reason: (required)
5957
- :arg shard: (required)
5958
5963
  :arg index:
5959
5964
  :arg node:
5965
+ :arg shard:
5960
5966
  :arg status:
5967
+ :arg primary:
5961
5968
  """
5962
5969
 
5963
5970
  reason: "ErrorCause"
5964
- shard: int
5965
5971
  index: str
5966
5972
  node: str
5973
+ shard: int
5967
5974
  status: str
5975
+ primary: bool
5968
5976
 
5969
5977
 
5970
5978
  class ShardProfile(AttrDict[Any]):
@@ -6288,7 +6296,9 @@ class TDigestPercentileRanksAggregate(AttrDict[Any]):
6288
6296
  :arg meta:
6289
6297
  """
6290
6298
 
6291
- values: Union[Mapping[str, Union[str, int, None]], Sequence["ArrayPercentilesItem"]]
6299
+ values: Union[
6300
+ Mapping[str, Union[str, float, None]], Sequence["ArrayPercentilesItem"]
6301
+ ]
6292
6302
  meta: Mapping[str, Any]
6293
6303
 
6294
6304
 
@@ -6298,7 +6308,9 @@ class TDigestPercentilesAggregate(AttrDict[Any]):
6298
6308
  :arg meta:
6299
6309
  """
6300
6310
 
6301
- values: Union[Mapping[str, Union[str, int, None]], Sequence["ArrayPercentilesItem"]]
6311
+ values: Union[
6312
+ Mapping[str, Union[str, float, None]], Sequence["ArrayPercentilesItem"]
6313
+ ]
6302
6314
  meta: Mapping[str, Any]
6303
6315
 
6304
6316
 
@@ -603,7 +603,7 @@ class ObjectBase(AttrDict[Any]):
603
603
  # if this is a mapped field,
604
604
  f = self.__get_field(k)
605
605
  if f and f._coerce:
606
- v = f.serialize(v)
606
+ v = f.serialize(v, skip_empty=skip_empty)
607
607
 
608
608
  # if someone assigned AttrList, unwrap it
609
609
  if isinstance(v, AttrList):
@@ -15,4 +15,5 @@
15
15
  # specific language governing permissions and limitations
16
16
  # under the License.
17
17
 
18
- from .esql import ESQL, and_, not_, or_ # noqa: F401
18
+ from ..dsl import E # noqa: F401
19
+ from .esql import ESQL, ESQLBase, and_, not_, or_ # noqa: F401