diff --git a/elasticsearch/_async/client/snapshot.py b/elasticsearch/_async/client/snapshot.py index 0f4e8530c..d9bed38b2 100644 --- a/elasticsearch/_async/client/snapshot.py +++ b/elasticsearch/_async/client/snapshot.py @@ -802,6 +802,10 @@ async def repository_analyze( This allows you to demonstrate to your storage supplier that a repository analysis failure must only be caused by an incompatibility with AWS S3 and cannot be attributed to a problem in Elasticsearch. Please do not report Elasticsearch issues involving third-party storage systems unless you can demonstrate that the same issue exists when analysing a repository that uses the reference implementation of the same storage protocol. You will need to work with the supplier of your storage system to address the incompatibilities that Elasticsearch detects.

+

The analysis may also report a failure if your repository experienced a service disruption while the analysis was running. + In practice, occasional service disruptions are inevitable, but the analysis cannot itself distinguish such disruptions from incorrect behavior so must report all deviations from the expected behavior as failures. + If you are certain that you can ascribe an analysis failure to such a service disruption, wait for your service provider to resolve the disruption and then re-run the analysis. + Elasticsearch will be unable to create or restore snapshots during repository service disruptions, so you must ensure that these events occur only very rarely.

If the analysis is successful, the API returns details of the testing process, optionally including how long each operation took. You can use this information to determine the performance of your storage system. If any operation fails or returns an incorrect result, the API returns an error. diff --git a/elasticsearch/_sync/client/snapshot.py b/elasticsearch/_sync/client/snapshot.py index 97b9ecea7..475d26ba7 100644 --- a/elasticsearch/_sync/client/snapshot.py +++ b/elasticsearch/_sync/client/snapshot.py @@ -802,6 +802,10 @@ def repository_analyze( This allows you to demonstrate to your storage supplier that a repository analysis failure must only be caused by an incompatibility with AWS S3 and cannot be attributed to a problem in Elasticsearch. Please do not report Elasticsearch issues involving third-party storage systems unless you can demonstrate that the same issue exists when analysing a repository that uses the reference implementation of the same storage protocol. You will need to work with the supplier of your storage system to address the incompatibilities that Elasticsearch detects.

+

The analysis may also report a failure if your repository experienced a service disruption while the analysis was running. + In practice, occasional service disruptions are inevitable, but the analysis cannot itself distinguish such disruptions from incorrect behavior so must report all deviations from the expected behavior as failures. + If you are certain that you can ascribe an analysis failure to such a service disruption, wait for your service provider to resolve the disruption and then re-run the analysis. + Elasticsearch will be unable to create or restore snapshots during repository service disruptions, so you must ensure that these events occur only very rarely.

If the analysis is successful, the API returns details of the testing process, optionally including how long each operation took. You can use this information to determine the performance of your storage system. If any operation fails or returns an incorrect result, the API returns an error. diff --git a/elasticsearch/_version.py b/elasticsearch/_version.py index 314961bed..61fe3047a 100644 --- a/elasticsearch/_version.py +++ b/elasticsearch/_version.py @@ -16,5 +16,5 @@ # under the License. __versionstr__ = "9.4.0" -__es_specification_commit__ = "9926f2cb48e44e4e3540fbe48303dbfef72d8bd7" +__es_specification_commit__ = "fcf537e4be958d56e9c7cafe9076afdc8a91ffc1" _SERVERLESS_API_VERSION = "2023-10-31" diff --git a/elasticsearch/dsl/types.py b/elasticsearch/dsl/types.py index f0fc9020d..102535fa0 100644 --- a/elasticsearch/dsl/types.py +++ b/elasticsearch/dsl/types.py @@ -656,6 +656,7 @@ class FieldSort(AttrDict[Any]): "keyword", "text", "search_as_you_type", + "wildcard", "date", "date_nanos", "boolean", @@ -721,6 +722,7 @@ def __init__( "keyword", "text", "search_as_you_type", + "wildcard", "date", "date_nanos", "boolean", diff --git a/elasticsearch/serializer.py b/elasticsearch/serializer.py index c281f1348..4596f8cdc 100644 --- a/elasticsearch/serializer.py +++ b/elasticsearch/serializer.py @@ -54,7 +54,7 @@ __all__.append("PyArrowSerializer") except ImportError: - pa = None + pa = None # type: ignore[assignment] class JsonSerializer(_JsonSerializer): diff --git a/test_elasticsearch/test_async/test_server/test_vectorstore/test_vectorstore.py b/test_elasticsearch/test_async/test_server/test_vectorstore/test_vectorstore.py index 35ea51816..fdc86887d 100644 --- a/test_elasticsearch/test_async/test_server/test_vectorstore/test_vectorstore.py +++ b/test_elasticsearch/test_async/test_server/test_vectorstore/test_vectorstore.py @@ -1065,14 +1065,9 @@ async def test_metadata_mapping( "dims": 10, "index": True, "index_options": { - "bits": 4, - "cluster_size": 384, - "default_visit_percentage": 0.0, - "flat_index_threshold": -1, - "rescore_vector": { - "oversample": 3.0, - }, - "type": "bbq_disk", + "ef_construction": 100, + "m": 16, + "type": "int8_hnsw", }, "similarity": "cosine", } diff --git a/test_elasticsearch/test_server/test_vectorstore/test_vectorstore.py b/test_elasticsearch/test_server/test_vectorstore/test_vectorstore.py index 6fdf93fff..fffe9cf05 100644 --- a/test_elasticsearch/test_server/test_vectorstore/test_vectorstore.py +++ b/test_elasticsearch/test_server/test_vectorstore/test_vectorstore.py @@ -1045,14 +1045,9 @@ def test_metadata_mapping(self, sync_client: Elasticsearch, index: str) -> None: "dims": 10, "index": True, "index_options": { - "bits": 4, - "cluster_size": 384, - "default_visit_percentage": 0.0, - "flat_index_threshold": -1, - "rescore_vector": { - "oversample": 3.0, - }, - "type": "bbq_disk", + "ef_construction": 100, + "m": 16, + "type": "int8_hnsw", }, "similarity": "cosine", }