Still testing
This commit is contained in:
@ -56,7 +56,7 @@ interface AsyncSearchSubmitRequest extends <<RequestBase>> {
|
||||
allow_partial_search_results?: boolean
|
||||
analyzer?: string
|
||||
analyze_wildcard?: boolean
|
||||
batched_reduce_size?: number
|
||||
batched_reduce_size?: <<_long, long>>
|
||||
ccs_minimize_roundtrips?: boolean
|
||||
default_operator?: QueryDslOperator
|
||||
df?: string
|
||||
@ -64,17 +64,17 @@ interface AsyncSearchSubmitRequest extends <<RequestBase>> {
|
||||
ignore_throttled?: boolean
|
||||
ignore_unavailable?: boolean
|
||||
lenient?: boolean
|
||||
max_concurrent_shard_requests?: number
|
||||
max_concurrent_shard_requests?: <<_long, long>>
|
||||
min_compatible_shard_node?: <<VersionString>>
|
||||
preference?: string
|
||||
pre_filter_shard_size?: number
|
||||
pre_filter_shard_size?: <<_long, long>>
|
||||
request_cache?: boolean
|
||||
routing?: <<Routing>>
|
||||
scroll?: <<Duration>>
|
||||
search_type?: <<SearchType>>
|
||||
suggest_field?: <<Field>>
|
||||
suggest_mode?: <<SuggestMode>>
|
||||
suggest_size?: number
|
||||
suggest_size?: <<_long, long>>
|
||||
suggest_text?: string
|
||||
typed_keys?: boolean
|
||||
rest_total_hits_as_int?: boolean
|
||||
@ -82,31 +82,31 @@ interface AsyncSearchSubmitRequest extends <<RequestBase>> {
|
||||
_source_includes?: <<Fields>>
|
||||
q?: string
|
||||
aggregations?: Record<string, AggregationsAggregationContainer>
|
||||
/** @alias aggregations */
|
||||
pass:[/**] @alias aggregations */
|
||||
aggs?: Record<string, AggregationsAggregationContainer>
|
||||
collapse?: SearchFieldCollapse
|
||||
explain?: boolean
|
||||
ext?: Record<string, any>
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
highlight?: SearchHighlight
|
||||
track_total_hits?: SearchTrackHits
|
||||
indices_boost?: Record<<<IndexName>>, number>[]
|
||||
indices_boost?: Record<<<IndexName>>, <<_double, double>>>[]
|
||||
docvalue_fields?: (QueryDslFieldAndFormat | <<Field>>)[]
|
||||
knn?: <<KnnSearch>> | <<KnnSearch>>[]
|
||||
min_score?: number
|
||||
min_score?: <<_double, double>>
|
||||
post_filter?: QueryDslQueryContainer
|
||||
profile?: boolean
|
||||
query?: QueryDslQueryContainer
|
||||
rescore?: SearchRescore | SearchRescore[]
|
||||
script_fields?: Record<string, <<ScriptField>>>
|
||||
search_after?: <<SortResults>>
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
slice?: <<SlicedScroll>>
|
||||
sort?: <<Sort>>
|
||||
_source?: SearchSourceConfig
|
||||
fields?: (QueryDslFieldAndFormat | <<Field>>)[]
|
||||
suggest?: SearchSuggester
|
||||
terminate_after?: number
|
||||
terminate_after?: <<_long, long>>
|
||||
timeout?: string
|
||||
track_scores?: boolean
|
||||
version?: boolean
|
||||
|
||||
@ -75,8 +75,8 @@ interface BulkRequest<TDocument = unknown, TPartialDocument = unknown> extends <
|
||||
interface BulkResponse {
|
||||
errors: boolean
|
||||
items: Partial<Record<BulkOperationType, BulkResponseItem>>[]
|
||||
took: number
|
||||
ingest_took?: number
|
||||
took: <<_long, long>>
|
||||
ingest_took?: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -53,8 +53,8 @@ interface CatMlTrainedModelsRequest extends CatCatRequestBase {
|
||||
bytes?: <<Bytes>>
|
||||
h?: CatCatTrainedModelsColumns
|
||||
s?: CatCatTrainedModelsColumns
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -50,11 +50,11 @@ Get transforms. Returns configuration and usage information about transforms. CA
|
||||
interface CatTransformsRequest extends CatCatRequestBase {
|
||||
transform_id?: <<Id>>
|
||||
allow_no_match?: boolean
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
h?: CatCatTransformColumns
|
||||
s?: CatCatTransformColumns
|
||||
time?: <<TimeUnit>>
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -51,14 +51,14 @@ interface CcrFollowRequest extends <<RequestBase>> {
|
||||
index: <<IndexName>>
|
||||
wait_for_active_shards?: <<WaitForActiveShards>>
|
||||
leader_index?: <<IndexName>>
|
||||
max_outstanding_read_requests?: number
|
||||
max_outstanding_write_requests?: number
|
||||
max_read_request_operation_count?: number
|
||||
max_outstanding_read_requests?: <<_long, long>>
|
||||
max_outstanding_write_requests?: <<_long, long>>
|
||||
max_read_request_operation_count?: <<_long, long>>
|
||||
max_read_request_size?: string
|
||||
max_retry_delay?: <<Duration>>
|
||||
max_write_buffer_count?: number
|
||||
max_write_buffer_count?: <<_long, long>>
|
||||
max_write_buffer_size?: string
|
||||
max_write_request_operation_count?: number
|
||||
max_write_request_operation_count?: <<_long, long>>
|
||||
max_write_request_size?: string
|
||||
read_poll_timeout?: <<Duration>>
|
||||
remote_cluster?: string
|
||||
|
||||
@ -53,16 +53,16 @@ interface CcrPutAutoFollowPatternRequest extends <<RequestBase>> {
|
||||
follow_index_pattern?: <<IndexPattern>>
|
||||
leader_index_patterns?: <<IndexPatterns>>
|
||||
leader_index_exclusion_patterns?: <<IndexPatterns>>
|
||||
max_outstanding_read_requests?: number
|
||||
max_outstanding_read_requests?: <<_integer, integer>>
|
||||
settings?: Record<string, any>
|
||||
max_outstanding_write_requests?: number
|
||||
max_outstanding_write_requests?: <<_integer, integer>>
|
||||
read_poll_timeout?: <<Duration>>
|
||||
max_read_request_operation_count?: number
|
||||
max_read_request_operation_count?: <<_integer, integer>>
|
||||
max_read_request_size?: <<ByteSize>>
|
||||
max_retry_delay?: <<Duration>>
|
||||
max_write_buffer_count?: number
|
||||
max_write_buffer_count?: <<_integer, integer>>
|
||||
max_write_buffer_size?: <<ByteSize>>
|
||||
max_write_request_operation_count?: number
|
||||
max_write_request_operation_count?: <<_integer, integer>>
|
||||
max_write_request_size?: <<ByteSize>>
|
||||
}
|
||||
|
||||
|
||||
@ -49,14 +49,14 @@ Resumes a follower index that has been paused
|
||||
++++
|
||||
interface CcrResumeFollowRequest extends <<RequestBase>> {
|
||||
index: <<IndexName>>
|
||||
max_outstanding_read_requests?: number
|
||||
max_outstanding_write_requests?: number
|
||||
max_read_request_operation_count?: number
|
||||
max_outstanding_read_requests?: <<_long, long>>
|
||||
max_outstanding_write_requests?: <<_long, long>>
|
||||
max_read_request_operation_count?: <<_long, long>>
|
||||
max_read_request_size?: string
|
||||
max_retry_delay?: <<Duration>>
|
||||
max_write_buffer_count?: number
|
||||
max_write_buffer_count?: <<_long, long>>
|
||||
max_write_buffer_size?: string
|
||||
max_write_request_operation_count?: number
|
||||
max_write_request_operation_count?: <<_long, long>>
|
||||
max_write_request_size?: string
|
||||
read_poll_timeout?: <<Duration>>
|
||||
}
|
||||
|
||||
@ -64,7 +64,7 @@ interface ClearScrollRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface ClearScrollResponse {
|
||||
succeeded: boolean
|
||||
num_freed: number
|
||||
num_freed: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -64,7 +64,7 @@ interface ClosePointInTimeRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface ClosePointInTimeResponse {
|
||||
succeeded: boolean
|
||||
num_freed: number
|
||||
num_freed: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -53,7 +53,7 @@ interface ClusterAllocationExplainRequest extends <<RequestBase>> {
|
||||
current_node?: string
|
||||
index?: <<IndexName>>
|
||||
primary?: boolean
|
||||
shard?: number
|
||||
shard?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -90,7 +90,7 @@ interface ClusterAllocationExplainResponse {
|
||||
rebalance_explanation?: string
|
||||
remaining_delay?: <<Duration>>
|
||||
remaining_delay_in_millis?: <<DurationValue>><<<UnitMillis>>>
|
||||
shard: number
|
||||
shard: <<_integer, integer>>
|
||||
unassigned_info?: ClusterAllocationExplainUnassignedInformation
|
||||
note?: string
|
||||
}
|
||||
|
||||
@ -56,7 +56,7 @@ interface ClusterHealthRequest extends <<RequestBase>> {
|
||||
timeout?: <<Duration>>
|
||||
wait_for_active_shards?: <<WaitForActiveShards>>
|
||||
wait_for_events?: <<WaitForEvents>>
|
||||
wait_for_nodes?: string | number
|
||||
wait_for_nodes?: string | <<_integer, integer>>
|
||||
wait_for_no_initializing_shards?: boolean
|
||||
wait_for_no_relocating_shards?: boolean
|
||||
wait_for_status?: <<HealthStatus>>
|
||||
|
||||
@ -48,8 +48,8 @@ Get all connectors. Get information about all connectors.
|
||||
<pre>
|
||||
++++
|
||||
interface ConnectorListRequest extends <<RequestBase>> {
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
index_name?: <<Indices>>
|
||||
connector_name?: <<Names>>
|
||||
service_type?: <<Names>>
|
||||
@ -68,7 +68,7 @@ interface ConnectorListRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface ConnectorListResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
results: ConnectorConnector[]
|
||||
}
|
||||
|
||||
|
||||
@ -48,8 +48,8 @@ Get all connector sync jobs. Get information about all stored connector sync job
|
||||
<pre>
|
||||
++++
|
||||
interface ConnectorSyncJobListRequest extends <<RequestBase>> {
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
status?: ConnectorSyncStatus
|
||||
connector_id?: <<Id>>
|
||||
job_type?: ConnectorSyncJobType | ConnectorSyncJobType[]
|
||||
@ -67,7 +67,7 @@ interface ConnectorSyncJobListRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface ConnectorSyncJobListResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
results: ConnectorConnectorSyncJob[]
|
||||
}
|
||||
|
||||
|
||||
@ -58,10 +58,10 @@ interface CountRequest extends <<RequestBase>> {
|
||||
ignore_throttled?: boolean
|
||||
ignore_unavailable?: boolean
|
||||
lenient?: boolean
|
||||
min_score?: number
|
||||
min_score?: <<_double, double>>
|
||||
preference?: string
|
||||
routing?: <<Routing>>
|
||||
terminate_after?: number
|
||||
terminate_after?: <<_long, long>>
|
||||
q?: string
|
||||
query?: QueryDslQueryContainer
|
||||
}
|
||||
@ -78,7 +78,7 @@ interface CountRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface CountResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
_shards: <<ShardStatistics>>
|
||||
}
|
||||
|
||||
|
||||
@ -50,7 +50,7 @@ Delete a document. Removes a JSON document from the specified index.
|
||||
interface DeleteRequest extends <<RequestBase>> {
|
||||
id: <<Id>>
|
||||
index: <<IndexName>>
|
||||
if_primary_term?: number
|
||||
if_primary_term?: <<_long, long>>
|
||||
if_seq_no?: <<SequenceNumber>>
|
||||
refresh?: <<Refresh>>
|
||||
routing?: <<Routing>>
|
||||
|
||||
@ -56,28 +56,28 @@ interface DeleteByQueryRequest extends <<RequestBase>> {
|
||||
default_operator?: QueryDslOperator
|
||||
df?: string
|
||||
expand_wildcards?: <<ExpandWildcards>>
|
||||
from?: number
|
||||
from?: <<_long, long>>
|
||||
ignore_unavailable?: boolean
|
||||
lenient?: boolean
|
||||
preference?: string
|
||||
refresh?: boolean
|
||||
request_cache?: boolean
|
||||
requests_per_second?: number
|
||||
requests_per_second?: <<_float, float>>
|
||||
routing?: <<Routing>>
|
||||
q?: string
|
||||
scroll?: <<Duration>>
|
||||
scroll_size?: number
|
||||
scroll_size?: <<_long, long>>
|
||||
search_timeout?: <<Duration>>
|
||||
search_type?: <<SearchType>>
|
||||
slices?: <<Slices>>
|
||||
sort?: string[]
|
||||
stats?: string[]
|
||||
terminate_after?: number
|
||||
terminate_after?: <<_long, long>>
|
||||
timeout?: <<Duration>>
|
||||
version?: boolean
|
||||
wait_for_active_shards?: <<WaitForActiveShards>>
|
||||
wait_for_completion?: boolean
|
||||
max_docs?: number
|
||||
max_docs?: <<_long, long>>
|
||||
query?: QueryDslQueryContainer
|
||||
slice?: <<SlicedScroll>>
|
||||
}
|
||||
@ -94,13 +94,13 @@ interface DeleteByQueryRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface DeleteByQueryResponse {
|
||||
batches?: number
|
||||
deleted?: number
|
||||
batches?: <<_long, long>>
|
||||
deleted?: <<_long, long>>
|
||||
failures?: <<BulkIndexByScrollFailure>>[]
|
||||
noops?: number
|
||||
requests_per_second?: number
|
||||
noops?: <<_long, long>>
|
||||
requests_per_second?: <<_float, float>>
|
||||
retries?: <<Retries>>
|
||||
slice_id?: number
|
||||
slice_id?: <<_integer, integer>>
|
||||
task?: <<TaskId>>
|
||||
throttled?: <<Duration>>
|
||||
throttled_millis?: <<DurationValue>><<<UnitMillis>>>
|
||||
@ -108,8 +108,8 @@ interface DeleteByQueryResponse {
|
||||
throttled_until_millis?: <<DurationValue>><<<UnitMillis>>>
|
||||
timed_out?: boolean
|
||||
took?: <<DurationValue>><<<UnitMillis>>>
|
||||
total?: number
|
||||
version_conflicts?: number
|
||||
total?: <<_long, long>>
|
||||
version_conflicts?: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -49,7 +49,7 @@ Throttle a delete by query operation. Change the number of requests per second f
|
||||
++++
|
||||
interface DeleteByQueryRethrottleRequest extends <<RequestBase>> {
|
||||
task_id: <<TaskId>>
|
||||
requests_per_second?: number
|
||||
requests_per_second?: <<_float, float>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -68,7 +68,7 @@ interface EqlGetStatusResponse {
|
||||
is_running: boolean
|
||||
start_time_in_millis?: <<EpochTime>><<<UnitMillis>>>
|
||||
expiration_time_in_millis?: <<EpochTime>><<<UnitMillis>>>
|
||||
completion_status?: number
|
||||
completion_status?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -57,12 +57,12 @@ interface EqlSearchRequest extends <<RequestBase>> {
|
||||
event_category_field?: <<Field>>
|
||||
tiebreaker_field?: <<Field>>
|
||||
timestamp_field?: <<Field>>
|
||||
fetch_size?: number
|
||||
fetch_size?: <<_uint, uint>>
|
||||
filter?: QueryDslQueryContainer | QueryDslQueryContainer[]
|
||||
keep_alive?: <<Duration>>
|
||||
keep_on_completion?: boolean
|
||||
wait_for_completion_timeout?: <<Duration>>
|
||||
size?: number
|
||||
size?: <<_uint, uint>>
|
||||
fields?: QueryDslFieldAndFormat | <<Field>> | (QueryDslFieldAndFormat | <<Field>>)[]
|
||||
result_position?: EqlSearchResultPosition
|
||||
runtime_mappings?: MappingRuntimeFields
|
||||
|
||||
@ -52,9 +52,9 @@ interface FleetMsearchRequest extends <<RequestBase>> {
|
||||
expand_wildcards?: <<ExpandWildcards>>
|
||||
ignore_throttled?: boolean
|
||||
ignore_unavailable?: boolean
|
||||
max_concurrent_searches?: number
|
||||
max_concurrent_shard_requests?: number
|
||||
pre_filter_shard_size?: number
|
||||
max_concurrent_searches?: <<_long, long>>
|
||||
max_concurrent_shard_requests?: <<_long, long>>
|
||||
pre_filter_shard_size?: <<_long, long>>
|
||||
search_type?: <<SearchType>>
|
||||
rest_total_hits_as_int?: boolean
|
||||
typed_keys?: boolean
|
||||
|
||||
@ -50,7 +50,7 @@ interface FleetSearchRequest extends <<RequestBase>> {
|
||||
allow_no_indices?: boolean
|
||||
analyzer?: string
|
||||
analyze_wildcard?: boolean
|
||||
batched_reduce_size?: number
|
||||
batched_reduce_size?: <<_long, long>>
|
||||
ccs_minimize_roundtrips?: boolean
|
||||
default_operator?: QueryDslOperator
|
||||
df?: string
|
||||
@ -58,17 +58,17 @@ interface FleetSearchRequest extends <<RequestBase>> {
|
||||
ignore_throttled?: boolean
|
||||
ignore_unavailable?: boolean
|
||||
lenient?: boolean
|
||||
max_concurrent_shard_requests?: number
|
||||
max_concurrent_shard_requests?: <<_long, long>>
|
||||
min_compatible_shard_node?: <<VersionString>>
|
||||
preference?: string
|
||||
pre_filter_shard_size?: number
|
||||
pre_filter_shard_size?: <<_long, long>>
|
||||
request_cache?: boolean
|
||||
routing?: <<Routing>>
|
||||
scroll?: <<Duration>>
|
||||
search_type?: <<SearchType>>
|
||||
suggest_field?: <<Field>>
|
||||
suggest_mode?: <<SuggestMode>>
|
||||
suggest_size?: number
|
||||
suggest_size?: <<_long, long>>
|
||||
suggest_text?: string
|
||||
typed_keys?: boolean
|
||||
rest_total_hits_as_int?: boolean
|
||||
@ -78,30 +78,30 @@ interface FleetSearchRequest extends <<RequestBase>> {
|
||||
wait_for_checkpoints?: FleetCheckpoint[]
|
||||
allow_partial_search_results?: boolean
|
||||
aggregations?: Record<string, AggregationsAggregationContainer>
|
||||
/** @alias aggregations */
|
||||
pass:[/**] @alias aggregations */
|
||||
aggs?: Record<string, AggregationsAggregationContainer>
|
||||
collapse?: SearchFieldCollapse
|
||||
explain?: boolean
|
||||
ext?: Record<string, any>
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
highlight?: SearchHighlight
|
||||
track_total_hits?: SearchTrackHits
|
||||
indices_boost?: Record<<<IndexName>>, number>[]
|
||||
indices_boost?: Record<<<IndexName>>, <<_double, double>>>[]
|
||||
docvalue_fields?: (QueryDslFieldAndFormat | <<Field>>)[]
|
||||
min_score?: number
|
||||
min_score?: <<_double, double>>
|
||||
post_filter?: QueryDslQueryContainer
|
||||
profile?: boolean
|
||||
query?: QueryDslQueryContainer
|
||||
rescore?: SearchRescore | SearchRescore[]
|
||||
script_fields?: Record<string, <<ScriptField>>>
|
||||
search_after?: <<SortResults>>
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
slice?: <<SlicedScroll>>
|
||||
sort?: <<Sort>>
|
||||
_source?: SearchSourceConfig
|
||||
fields?: (QueryDslFieldAndFormat | <<Field>>)[]
|
||||
suggest?: SearchSuggester
|
||||
terminate_after?: number
|
||||
terminate_after?: <<_long, long>>
|
||||
timeout?: string
|
||||
track_scores?: boolean
|
||||
version?: boolean
|
||||
@ -124,15 +124,15 @@ interface FleetSearchRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface FleetSearchResponse<TDocument = unknown> {
|
||||
took: number
|
||||
took: <<_long, long>>
|
||||
timed_out: boolean
|
||||
_shards: <<ShardStatistics>>
|
||||
hits: SearchHitsMetadata<TDocument>
|
||||
aggregations?: Record<<<AggregateName>>, AggregationsAggregate>
|
||||
_clusters?: <<ClusterStatistics>>
|
||||
fields?: Record<string, any>
|
||||
max_score?: number
|
||||
num_reduce_phases?: number
|
||||
max_score?: <<_double, double>>
|
||||
num_reduce_phases?: <<_long, long>>
|
||||
profile?: SearchProfile
|
||||
pit_id?: <<Id>>
|
||||
_scroll_id?: <<ScrollId>>
|
||||
|
||||
@ -72,7 +72,7 @@ interface GraphExploreResponse {
|
||||
connections: GraphConnection[]
|
||||
failures: <<ShardFailure>>[]
|
||||
timed_out: boolean
|
||||
took: number
|
||||
took: <<_long, long>>
|
||||
vertices: GraphVertex[]
|
||||
}
|
||||
|
||||
|
||||
@ -51,7 +51,7 @@ interface HealthReportRequest extends <<RequestBase>> {
|
||||
feature?: string | string[]
|
||||
timeout?: <<Duration>>
|
||||
verbose?: boolean
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -50,7 +50,7 @@ Index a document. Adds a JSON document to the specified data stream or index and
|
||||
interface IndexRequest<TDocument = unknown> extends <<RequestBase>> {
|
||||
id?: <<Id>>
|
||||
index: <<IndexName>>
|
||||
if_primary_term?: number
|
||||
if_primary_term?: <<_long, long>>
|
||||
if_seq_no?: <<SequenceNumber>>
|
||||
op_type?: <<OpType>>
|
||||
pipeline?: string
|
||||
|
||||
@ -65,11 +65,11 @@ interface IndicesDataStreamsStatsRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface IndicesDataStreamsStatsResponse {
|
||||
_shards: <<ShardStatistics>>
|
||||
backing_indices: number
|
||||
data_stream_count: number
|
||||
backing_indices: <<_integer, integer>>
|
||||
data_stream_count: <<_integer, integer>>
|
||||
data_streams: IndicesDataStreamsStatsDataStreamsStatsItem[]
|
||||
total_store_sizes?: <<ByteSize>>
|
||||
total_store_size_bytes: number
|
||||
total_store_size_bytes: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -53,7 +53,7 @@ interface IndicesForcemergeRequest extends <<RequestBase>> {
|
||||
expand_wildcards?: <<ExpandWildcards>>
|
||||
flush?: boolean
|
||||
ignore_unavailable?: boolean
|
||||
max_num_segments?: number
|
||||
max_num_segments?: <<_long, long>>
|
||||
only_expunge_deletes?: boolean
|
||||
wait_for_completion?: boolean
|
||||
}
|
||||
|
||||
@ -56,7 +56,7 @@ interface IndicesPutIndexTemplateRequest extends <<RequestBase>> {
|
||||
composed_of?: <<Name>>[]
|
||||
template?: IndicesPutIndexTemplateIndexTemplateMapping
|
||||
data_stream?: IndicesDataStreamVisibility
|
||||
priority?: number
|
||||
priority?: <<_long, long>>
|
||||
version?: <<VersionNumber>>
|
||||
_meta?: <<Metadata>>
|
||||
allow_auto_create?: boolean
|
||||
|
||||
@ -55,7 +55,7 @@ interface IndicesPutTemplateRequest extends <<RequestBase>> {
|
||||
aliases?: Record<<<IndexName>>, IndicesAlias>
|
||||
index_patterns?: string | string[]
|
||||
mappings?: MappingTypeMapping
|
||||
order?: number
|
||||
order?: <<_integer, integer>>
|
||||
settings?: IndicesIndexSettings
|
||||
version?: <<VersionNumber>>
|
||||
}
|
||||
|
||||
@ -57,7 +57,7 @@ interface IndicesSimulateTemplateRequest extends <<RequestBase>> {
|
||||
composed_of?: <<Name>>[]
|
||||
template?: IndicesPutIndexTemplateIndexTemplateMapping
|
||||
data_stream?: IndicesDataStreamVisibility
|
||||
priority?: number
|
||||
priority?: <<_long, long>>
|
||||
version?: <<VersionNumber>>
|
||||
_meta?: <<Metadata>>
|
||||
ignore_missing_component_templates?: string[]
|
||||
|
||||
@ -70,12 +70,12 @@ interface KnnSearchRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface KnnSearchResponse<TDocument = unknown> {
|
||||
took: number
|
||||
took: <<_long, long>>
|
||||
timed_out: boolean
|
||||
_shards: <<ShardStatistics>>
|
||||
hits: SearchHitsMetadata<TDocument>
|
||||
fields?: Record<string, any>
|
||||
max_score?: number
|
||||
max_score?: <<_double, double>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -1,3 +1,5 @@
|
||||
|
||||
|
||||
[[reference-main-index]]
|
||||
== API Reference
|
||||
|
||||
|
||||
@ -49,7 +49,7 @@ Delete expired ML data. Deletes all job results, model snapshots and forecast da
|
||||
++++
|
||||
interface MlDeleteExpiredDataRequest extends <<RequestBase>> {
|
||||
job_id?: <<Id>>
|
||||
requests_per_second?: number
|
||||
requests_per_second?: <<_float, float>>
|
||||
timeout?: <<Duration>>
|
||||
}
|
||||
|
||||
|
||||
@ -49,8 +49,8 @@ Estimate job model memory usage. Makes an estimation of the memory usage for an
|
||||
++++
|
||||
interface MlEstimateModelMemoryRequest extends <<RequestBase>> {
|
||||
analysis_config?: MlAnalysisConfig
|
||||
max_bucket_cardinality?: Record<<<Field>>, number>
|
||||
overall_cardinality?: Record<<<Field>>, number>
|
||||
max_bucket_cardinality?: Record<<<Field>>, <<_long, long>>>
|
||||
overall_cardinality?: Record<<<Field>>, <<_long, long>>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -54,7 +54,7 @@ interface MlExplainDataFrameAnalyticsRequest extends <<RequestBase>> {
|
||||
analysis?: MlDataframeAnalysisContainer
|
||||
description?: string
|
||||
model_memory_limit?: string
|
||||
max_num_threads?: number
|
||||
max_num_threads?: <<_integer, integer>>
|
||||
analyzed_fields?: MlDataframeAnalysisAnalyzedFields | string[]
|
||||
allow_lazy_start?: boolean
|
||||
}
|
||||
|
||||
@ -69,7 +69,7 @@ interface MlFlushJobRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface MlFlushJobResponse {
|
||||
flushed: boolean
|
||||
last_finalized_bucket_end?: number
|
||||
last_finalized_bucket_end?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -50,9 +50,9 @@ Get anomaly detection job results for buckets. The API presents a chronological
|
||||
interface MlGetBucketsRequest extends <<RequestBase>> {
|
||||
job_id: <<Id>>
|
||||
timestamp?: <<DateTime>>
|
||||
from?: number
|
||||
size?: number
|
||||
anomaly_score?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
anomaly_score?: <<_double, double>>
|
||||
desc?: boolean
|
||||
end?: <<DateTime>>
|
||||
exclude_interim?: boolean
|
||||
@ -75,7 +75,7 @@ interface MlGetBucketsRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface MlGetBucketsResponse {
|
||||
buckets: MlBucketSummary[]
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -50,9 +50,9 @@ Get info about events in calendars.
|
||||
interface MlGetCalendarEventsRequest extends <<RequestBase>> {
|
||||
calendar_id: <<Id>>
|
||||
end?: <<DateTime>>
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
job_id?: <<Id>>
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
start?: <<DateTime>>
|
||||
}
|
||||
|
||||
@ -68,7 +68,7 @@ interface MlGetCalendarEventsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetCalendarEventsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
events: MlCalendarEvent[]
|
||||
}
|
||||
|
||||
|
||||
@ -49,8 +49,8 @@ Get calendar configuration info.
|
||||
++++
|
||||
interface MlGetCalendarsRequest extends <<RequestBase>> {
|
||||
calendar_id?: <<Id>>
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
page?: MlPage
|
||||
}
|
||||
|
||||
@ -67,7 +67,7 @@ interface MlGetCalendarsRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface MlGetCalendarsResponse {
|
||||
calendars: MlGetCalendarsCalendar[]
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -50,9 +50,9 @@ Get anomaly detection job results for categories.
|
||||
interface MlGetCategoriesRequest extends <<RequestBase>> {
|
||||
job_id: <<Id>>
|
||||
category_id?: <<CategoryId>>
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
partition_field_value?: string
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
page?: MlPage
|
||||
}
|
||||
|
||||
@ -69,7 +69,7 @@ interface MlGetCategoriesRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface MlGetCategoriesResponse {
|
||||
categories: MlCategory[]
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -50,8 +50,8 @@ Get data frame analytics job configuration info. You can get information for mul
|
||||
interface MlGetDataFrameAnalyticsRequest extends <<RequestBase>> {
|
||||
id?: <<Id>>
|
||||
allow_no_match?: boolean
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
exclude_generated?: boolean
|
||||
}
|
||||
|
||||
@ -67,7 +67,7 @@ interface MlGetDataFrameAnalyticsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetDataFrameAnalyticsResponse {
|
||||
count: number
|
||||
count: <<_integer, integer>>
|
||||
data_frame_analytics: MlDataframeAnalyticsSummary[]
|
||||
}
|
||||
|
||||
|
||||
@ -50,8 +50,8 @@ Get data frame analytics jobs usage info.
|
||||
interface MlGetDataFrameAnalyticsStatsRequest extends <<RequestBase>> {
|
||||
id?: <<Id>>
|
||||
allow_no_match?: boolean
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
verbose?: boolean
|
||||
}
|
||||
|
||||
@ -67,7 +67,7 @@ interface MlGetDataFrameAnalyticsStatsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetDataFrameAnalyticsStatsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
data_frame_analytics: MlDataframeAnalytics[]
|
||||
}
|
||||
|
||||
|
||||
@ -64,7 +64,7 @@ interface MlGetDatafeedStatsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetDatafeedStatsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
datafeeds: MlDatafeedStats[]
|
||||
}
|
||||
|
||||
|
||||
@ -65,7 +65,7 @@ interface MlGetDatafeedsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetDatafeedsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
datafeeds: MlDatafeed[]
|
||||
}
|
||||
|
||||
|
||||
@ -49,8 +49,8 @@ Get filters. You can get a single filter or all filters.
|
||||
++++
|
||||
interface MlGetFiltersRequest extends <<RequestBase>> {
|
||||
filter_id?: <<Ids>>
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -65,7 +65,7 @@ interface MlGetFiltersRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetFiltersResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
filters: MlFilter[]
|
||||
}
|
||||
|
||||
|
||||
@ -52,9 +52,9 @@ interface MlGetInfluencersRequest extends <<RequestBase>> {
|
||||
desc?: boolean
|
||||
end?: <<DateTime>>
|
||||
exclude_interim?: boolean
|
||||
influencer_score?: number
|
||||
from?: number
|
||||
size?: number
|
||||
influencer_score?: <<_double, double>>
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
sort?: <<Field>>
|
||||
start?: <<DateTime>>
|
||||
page?: MlPage
|
||||
@ -72,7 +72,7 @@ interface MlGetInfluencersRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetInfluencersResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
influencers: MlInfluencer[]
|
||||
}
|
||||
|
||||
|
||||
@ -64,7 +64,7 @@ interface MlGetJobStatsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetJobStatsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
jobs: MlJobStats[]
|
||||
}
|
||||
|
||||
|
||||
@ -65,7 +65,7 @@ interface MlGetJobsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetJobsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
jobs: MlJob[]
|
||||
}
|
||||
|
||||
|
||||
@ -65,7 +65,7 @@ interface MlGetModelSnapshotUpgradeStatsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetModelSnapshotUpgradeStatsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
model_snapshot_upgrades: MlModelSnapshotUpgrade[]
|
||||
}
|
||||
|
||||
|
||||
@ -50,8 +50,8 @@ Get model snapshots info.
|
||||
interface MlGetModelSnapshotsRequest extends <<RequestBase>> {
|
||||
job_id: <<Id>>
|
||||
snapshot_id?: <<Id>>
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
desc?: boolean
|
||||
end?: <<DateTime>>
|
||||
page?: MlPage
|
||||
@ -71,7 +71,7 @@ interface MlGetModelSnapshotsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetModelSnapshotsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
model_snapshots: MlModelSnapshot[]
|
||||
}
|
||||
|
||||
|
||||
@ -53,9 +53,9 @@ interface MlGetOverallBucketsRequest extends <<RequestBase>> {
|
||||
bucket_span?: <<Duration>>
|
||||
end?: <<DateTime>>
|
||||
exclude_interim?: boolean
|
||||
overall_score?: number | string
|
||||
overall_score?: <<_double, double>> | string
|
||||
start?: <<DateTime>>
|
||||
top_n?: number
|
||||
top_n?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -70,7 +70,7 @@ interface MlGetOverallBucketsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetOverallBucketsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
overall_buckets: MlOverallBucket[]
|
||||
}
|
||||
|
||||
|
||||
@ -49,13 +49,13 @@ Get anomaly records for an anomaly detection job. Records contain the detailed a
|
||||
++++
|
||||
interface MlGetRecordsRequest extends <<RequestBase>> {
|
||||
job_id: <<Id>>
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
desc?: boolean
|
||||
end?: <<DateTime>>
|
||||
exclude_interim?: boolean
|
||||
page?: MlPage
|
||||
record_score?: number
|
||||
record_score?: <<_double, double>>
|
||||
sort?: <<Field>>
|
||||
start?: <<DateTime>>
|
||||
}
|
||||
@ -72,7 +72,7 @@ interface MlGetRecordsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetRecordsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
records: MlAnomaly[]
|
||||
}
|
||||
|
||||
|
||||
@ -52,9 +52,9 @@ interface MlGetTrainedModelsRequest extends <<RequestBase>> {
|
||||
allow_no_match?: boolean
|
||||
decompress_definition?: boolean
|
||||
exclude_generated?: boolean
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
include?: MlInclude
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
tags?: string | string[]
|
||||
}
|
||||
|
||||
@ -70,7 +70,7 @@ interface MlGetTrainedModelsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetTrainedModelsResponse {
|
||||
count: number
|
||||
count: <<_integer, integer>>
|
||||
trained_model_configs: MlTrainedModelConfig[]
|
||||
}
|
||||
|
||||
|
||||
@ -50,8 +50,8 @@ Get trained models usage info. You can get usage information for multiple traine
|
||||
interface MlGetTrainedModelsStatsRequest extends <<RequestBase>> {
|
||||
model_id?: <<Ids>>
|
||||
allow_no_match?: boolean
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -66,7 +66,7 @@ interface MlGetTrainedModelsStatsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlGetTrainedModelsStatsResponse {
|
||||
count: number
|
||||
count: <<_integer, integer>>
|
||||
trained_model_stats: MlTrainedModelStats[]
|
||||
}
|
||||
|
||||
|
||||
@ -66,21 +66,21 @@ interface MlPostDataRequest<TData = unknown> extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface MlPostDataResponse {
|
||||
bucket_count: number
|
||||
earliest_record_timestamp: number
|
||||
empty_bucket_count: number
|
||||
input_bytes: number
|
||||
input_field_count: number
|
||||
input_record_count: number
|
||||
invalid_date_count: number
|
||||
bucket_count: <<_long, long>>
|
||||
earliest_record_timestamp: <<_long, long>>
|
||||
empty_bucket_count: <<_long, long>>
|
||||
input_bytes: <<_long, long>>
|
||||
input_field_count: <<_long, long>>
|
||||
input_record_count: <<_long, long>>
|
||||
invalid_date_count: <<_long, long>>
|
||||
job_id: <<Id>>
|
||||
last_data_time: number
|
||||
latest_record_timestamp: number
|
||||
missing_field_count: number
|
||||
out_of_order_timestamp_count: number
|
||||
processed_field_count: number
|
||||
processed_record_count: number
|
||||
sparse_bucket_count: number
|
||||
last_data_time: <<_integer, integer>>
|
||||
latest_record_timestamp: <<_long, long>>
|
||||
missing_field_count: <<_long, long>>
|
||||
out_of_order_timestamp_count: <<_long, long>>
|
||||
processed_field_count: <<_long, long>>
|
||||
processed_record_count: <<_long, long>>
|
||||
sparse_bucket_count: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -54,7 +54,7 @@ interface MlPutDataFrameAnalyticsRequest extends <<RequestBase>> {
|
||||
analyzed_fields?: MlDataframeAnalysisAnalyzedFields | string[]
|
||||
description?: string
|
||||
dest: MlDataframeAnalyticsDestination
|
||||
max_num_threads?: number
|
||||
max_num_threads?: <<_integer, integer>>
|
||||
model_memory_limit?: string
|
||||
source: MlDataframeAnalyticsSource
|
||||
headers?: <<HttpHeaders>>
|
||||
@ -81,7 +81,7 @@ interface MlPutDataFrameAnalyticsResponse {
|
||||
description?: string
|
||||
dest: MlDataframeAnalyticsDestination
|
||||
id: <<Id>>
|
||||
max_num_threads: number
|
||||
max_num_threads: <<_integer, integer>>
|
||||
model_memory_limit: string
|
||||
source: MlDataframeAnalyticsSource
|
||||
version: <<VersionString>>
|
||||
|
||||
@ -58,16 +58,16 @@ interface MlPutDatafeedRequest extends <<RequestBase>> {
|
||||
delayed_data_check_config?: MlDelayedDataCheckConfig
|
||||
frequency?: <<Duration>>
|
||||
indices?: <<Indices>>
|
||||
/** @alias indices */
|
||||
pass:[/**] @alias indices */
|
||||
indexes?: <<Indices>>
|
||||
indices_options?: <<IndicesOptions>>
|
||||
job_id?: <<Id>>
|
||||
max_empty_searches?: number
|
||||
max_empty_searches?: <<_integer, integer>>
|
||||
query?: QueryDslQueryContainer
|
||||
query_delay?: <<Duration>>
|
||||
runtime_mappings?: MappingRuntimeFields
|
||||
script_fields?: Record<string, <<ScriptField>>>
|
||||
scroll_size?: number
|
||||
scroll_size?: <<_integer, integer>>
|
||||
headers?: <<HttpHeaders>>
|
||||
}
|
||||
|
||||
@ -92,12 +92,12 @@ interface MlPutDatafeedResponse {
|
||||
indices: string[]
|
||||
job_id: <<Id>>
|
||||
indices_options?: <<IndicesOptions>>
|
||||
max_empty_searches?: number
|
||||
max_empty_searches?: <<_integer, integer>>
|
||||
query: QueryDslQueryContainer
|
||||
query_delay: <<Duration>>
|
||||
runtime_mappings?: MappingRuntimeFields
|
||||
script_fields?: Record<string, <<ScriptField>>>
|
||||
scroll_size: number
|
||||
scroll_size: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -54,16 +54,16 @@ interface MlPutJobRequest extends <<RequestBase>> {
|
||||
analysis_limits?: MlAnalysisLimits
|
||||
background_persist_interval?: <<Duration>>
|
||||
custom_settings?: MlCustomSettings
|
||||
daily_model_snapshot_retention_after_days?: number
|
||||
daily_model_snapshot_retention_after_days?: <<_long, long>>
|
||||
data_description: MlDataDescription
|
||||
datafeed_config?: MlDatafeedConfig
|
||||
description?: string
|
||||
groups?: string[]
|
||||
model_plot_config?: MlModelPlotConfig
|
||||
model_snapshot_retention_days?: number
|
||||
renormalization_window_days?: number
|
||||
model_snapshot_retention_days?: <<_long, long>>
|
||||
renormalization_window_days?: <<_long, long>>
|
||||
results_index_name?: <<IndexName>>
|
||||
results_retention_days?: number
|
||||
results_retention_days?: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -84,7 +84,7 @@ interface MlPutJobResponse {
|
||||
background_persist_interval?: <<Duration>>
|
||||
create_time: <<DateTime>>
|
||||
custom_settings?: MlCustomSettings
|
||||
daily_model_snapshot_retention_after_days: number
|
||||
daily_model_snapshot_retention_after_days: <<_long, long>>
|
||||
data_description: MlDataDescription
|
||||
datafeed_config?: MlDatafeed
|
||||
description?: string
|
||||
@ -94,10 +94,10 @@ interface MlPutJobResponse {
|
||||
job_version: string
|
||||
model_plot_config?: MlModelPlotConfig
|
||||
model_snapshot_id?: <<Id>>
|
||||
model_snapshot_retention_days: number
|
||||
renormalization_window_days?: number
|
||||
model_snapshot_retention_days: <<_long, long>>
|
||||
renormalization_window_days?: <<_long, long>>
|
||||
results_index_name: string
|
||||
results_retention_days?: number
|
||||
results_retention_days?: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -58,7 +58,7 @@ interface MlPutTrainedModelRequest extends <<RequestBase>> {
|
||||
input?: MlPutTrainedModelInput
|
||||
metadata?: any
|
||||
model_type?: MlTrainedModelType
|
||||
model_size_bytes?: number
|
||||
model_size_bytes?: <<_long, long>>
|
||||
platform_architecture?: string
|
||||
tags?: string[]
|
||||
prefix_strings?: MlTrainedModelPrefixStrings
|
||||
|
||||
@ -49,10 +49,10 @@ Create part of a trained model definition.
|
||||
++++
|
||||
interface MlPutTrainedModelDefinitionPartRequest extends <<RequestBase>> {
|
||||
model_id: <<Id>>
|
||||
part: number
|
||||
part: <<_integer, integer>>
|
||||
definition: string
|
||||
total_definition_length: number
|
||||
total_parts: number
|
||||
total_definition_length: <<_long, long>>
|
||||
total_parts: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -51,7 +51,7 @@ interface MlPutTrainedModelVocabularyRequest extends <<RequestBase>> {
|
||||
model_id: <<Id>>
|
||||
vocabulary: string[]
|
||||
merges?: string[]
|
||||
scores?: number[]
|
||||
scores?: <<_double, double>>[]
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -51,10 +51,10 @@ interface MlStartTrainedModelDeploymentRequest extends <<RequestBase>> {
|
||||
model_id: <<Id>>
|
||||
cache_size?: <<ByteSize>>
|
||||
deployment_id?: string
|
||||
number_of_allocations?: number
|
||||
number_of_allocations?: <<_integer, integer>>
|
||||
priority?: MlTrainingPriority
|
||||
queue_capacity?: number
|
||||
threads_per_allocation?: number
|
||||
queue_capacity?: <<_integer, integer>>
|
||||
threads_per_allocation?: <<_integer, integer>>
|
||||
timeout?: <<Duration>>
|
||||
wait_for?: MlDeploymentAllocationState
|
||||
}
|
||||
|
||||
@ -51,7 +51,7 @@ interface MlUpdateDataFrameAnalyticsRequest extends <<RequestBase>> {
|
||||
id: <<Id>>
|
||||
description?: string
|
||||
model_memory_limit?: string
|
||||
max_num_threads?: number
|
||||
max_num_threads?: <<_integer, integer>>
|
||||
allow_lazy_start?: boolean
|
||||
}
|
||||
|
||||
@ -71,11 +71,11 @@ interface MlUpdateDataFrameAnalyticsResponse {
|
||||
allow_lazy_start: boolean
|
||||
analysis: MlDataframeAnalysisContainer
|
||||
analyzed_fields?: MlDataframeAnalysisAnalyzedFields | string[]
|
||||
create_time: number
|
||||
create_time: <<_long, long>>
|
||||
description?: string
|
||||
dest: MlDataframeAnalyticsDestination
|
||||
id: <<Id>>
|
||||
max_num_threads: number
|
||||
max_num_threads: <<_integer, integer>>
|
||||
model_memory_limit: string
|
||||
source: MlDataframeAnalyticsSource
|
||||
version: <<VersionString>>
|
||||
|
||||
@ -58,16 +58,16 @@ interface MlUpdateDatafeedRequest extends <<RequestBase>> {
|
||||
delayed_data_check_config?: MlDelayedDataCheckConfig
|
||||
frequency?: <<Duration>>
|
||||
indices?: string[]
|
||||
/** @alias indices */
|
||||
pass:[/**] @alias indices */
|
||||
indexes?: string[]
|
||||
indices_options?: <<IndicesOptions>>
|
||||
job_id?: <<Id>>
|
||||
max_empty_searches?: number
|
||||
max_empty_searches?: <<_integer, integer>>
|
||||
query?: QueryDslQueryContainer
|
||||
query_delay?: <<Duration>>
|
||||
runtime_mappings?: MappingRuntimeFields
|
||||
script_fields?: Record<string, <<ScriptField>>>
|
||||
scroll_size?: number
|
||||
scroll_size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -91,12 +91,12 @@ interface MlUpdateDatafeedResponse {
|
||||
indices: string[]
|
||||
indices_options?: <<IndicesOptions>>
|
||||
job_id: <<Id>>
|
||||
max_empty_searches?: number
|
||||
max_empty_searches?: <<_integer, integer>>
|
||||
query: QueryDslQueryContainer
|
||||
query_delay: <<Duration>>
|
||||
runtime_mappings?: MappingRuntimeFields
|
||||
script_fields?: Record<string, <<ScriptField>>>
|
||||
scroll_size: number
|
||||
scroll_size: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -57,10 +57,10 @@ interface MlUpdateJobRequest extends <<RequestBase>> {
|
||||
description?: string
|
||||
model_plot_config?: MlModelPlotConfig
|
||||
model_prune_window?: <<Duration>>
|
||||
daily_model_snapshot_retention_after_days?: number
|
||||
model_snapshot_retention_days?: number
|
||||
renormalization_window_days?: number
|
||||
results_retention_days?: number
|
||||
daily_model_snapshot_retention_after_days?: <<_long, long>>
|
||||
model_snapshot_retention_days?: <<_long, long>>
|
||||
renormalization_window_days?: <<_long, long>>
|
||||
results_retention_days?: <<_long, long>>
|
||||
groups?: string[]
|
||||
detectors?: MlDetector[]
|
||||
per_partition_categorization?: MlPerPartitionCategorization
|
||||
@ -85,7 +85,7 @@ interface MlUpdateJobResponse {
|
||||
create_time: <<EpochTime>><<<UnitMillis>>>
|
||||
finished_time?: <<EpochTime>><<<UnitMillis>>>
|
||||
custom_settings?: Record<string, string>
|
||||
daily_model_snapshot_retention_after_days: number
|
||||
daily_model_snapshot_retention_after_days: <<_long, long>>
|
||||
data_description: MlDataDescription
|
||||
datafeed_config?: MlDatafeed
|
||||
description?: string
|
||||
@ -95,10 +95,10 @@ interface MlUpdateJobResponse {
|
||||
job_version: <<VersionString>>
|
||||
model_plot_config?: MlModelPlotConfig
|
||||
model_snapshot_id?: <<Id>>
|
||||
model_snapshot_retention_days: number
|
||||
renormalization_window_days?: number
|
||||
model_snapshot_retention_days: <<_long, long>>
|
||||
renormalization_window_days?: <<_long, long>>
|
||||
results_index_name: <<IndexName>>
|
||||
results_retention_days?: number
|
||||
results_retention_days?: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -49,7 +49,7 @@ Update a trained model deployment.
|
||||
++++
|
||||
interface MlUpdateTrainedModelDeploymentRequest extends <<RequestBase>> {
|
||||
model_id: <<Id>>
|
||||
number_of_allocations?: number
|
||||
number_of_allocations?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -70,7 +70,7 @@ interface MonitoringBulkResponse {
|
||||
error?: <<ErrorCause>>
|
||||
errors: boolean
|
||||
ignored: boolean
|
||||
took: number
|
||||
took: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -55,9 +55,9 @@ interface MsearchRequest extends <<RequestBase>> {
|
||||
ignore_throttled?: boolean
|
||||
ignore_unavailable?: boolean
|
||||
include_named_queries_score?: boolean
|
||||
max_concurrent_searches?: number
|
||||
max_concurrent_shard_requests?: number
|
||||
pre_filter_shard_size?: number
|
||||
max_concurrent_searches?: <<_long, long>>
|
||||
max_concurrent_shard_requests?: <<_long, long>>
|
||||
pre_filter_shard_size?: <<_long, long>>
|
||||
rest_total_hits_as_int?: boolean
|
||||
routing?: <<Routing>>
|
||||
search_type?: <<SearchType>>
|
||||
|
||||
@ -50,7 +50,7 @@ Run multiple templated searches.
|
||||
interface MsearchTemplateRequest extends <<RequestBase>> {
|
||||
index?: <<Indices>>
|
||||
ccs_minimize_roundtrips?: boolean
|
||||
max_concurrent_searches?: number
|
||||
max_concurrent_searches?: <<_long, long>>
|
||||
search_type?: <<SearchType>>
|
||||
rest_total_hits_as_int?: boolean
|
||||
typed_keys?: boolean
|
||||
|
||||
@ -49,7 +49,7 @@ You can use this API to clear the archived repositories metering information in
|
||||
++++
|
||||
interface NodesClearRepositoriesMeteringArchiveRequest extends <<RequestBase>> {
|
||||
node_id: <<NodeIds>>
|
||||
max_archive_version: number
|
||||
max_archive_version: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -51,9 +51,9 @@ interface NodesHotThreadsRequest extends <<RequestBase>> {
|
||||
node_id?: <<NodeIds>>
|
||||
ignore_idle_threads?: boolean
|
||||
interval?: <<Duration>>
|
||||
snapshots?: number
|
||||
snapshots?: <<_long, long>>
|
||||
master_timeout?: <<Duration>>
|
||||
threads?: number
|
||||
threads?: <<_long, long>>
|
||||
timeout?: <<Duration>>
|
||||
type?: <<ThreadType>>
|
||||
sort?: <<ThreadType>>
|
||||
|
||||
@ -48,8 +48,8 @@ Returns summarized information about existing query rulesets.
|
||||
<pre>
|
||||
++++
|
||||
interface QueryRulesListRulesetsRequest extends <<RequestBase>> {
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -64,7 +64,7 @@ interface QueryRulesListRulesetsRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface QueryRulesListRulesetsResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
results: QueryRulesListRulesetsQueryRulesetListItem[]
|
||||
}
|
||||
|
||||
|
||||
@ -53,7 +53,7 @@ interface QueryRulesPutRuleRequest extends <<RequestBase>> {
|
||||
type: QueryRulesQueryRuleType
|
||||
criteria: QueryRulesQueryRuleCriteria | QueryRulesQueryRuleCriteria[]
|
||||
actions: QueryRulesQueryRuleActions
|
||||
priority?: number
|
||||
priority?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -64,7 +64,7 @@ interface QueryRulesTestRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface QueryRulesTestResponse {
|
||||
total_matched_rules: number
|
||||
total_matched_rules: <<_integer, integer>>
|
||||
matched_rules: QueryRulesTestQueryRulesetMatchedRule[]
|
||||
}
|
||||
|
||||
|
||||
@ -69,7 +69,7 @@ interface RankEvalRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface RankEvalResponse {
|
||||
metric_score: number
|
||||
metric_score: <<_double, double>>
|
||||
details: Record<<<Id>>, RankEvalRankEvalMetricDetail>
|
||||
failures: Record<string, any>
|
||||
}
|
||||
|
||||
@ -49,7 +49,7 @@ Reindex documents. Copies documents from a source to a destination. The source c
|
||||
++++
|
||||
interface ReindexRequest extends <<RequestBase>> {
|
||||
refresh?: boolean
|
||||
requests_per_second?: number
|
||||
requests_per_second?: <<_float, float>>
|
||||
scroll?: <<Duration>>
|
||||
slices?: <<Slices>>
|
||||
timeout?: <<Duration>>
|
||||
@ -58,9 +58,9 @@ interface ReindexRequest extends <<RequestBase>> {
|
||||
require_alias?: boolean
|
||||
conflicts?: <<Conflicts>>
|
||||
dest: ReindexDestination
|
||||
max_docs?: number
|
||||
max_docs?: <<_long, long>>
|
||||
script?: <<Script>> | string
|
||||
size?: number
|
||||
size?: <<_long, long>>
|
||||
source: ReindexSource
|
||||
}
|
||||
|
||||
@ -76,22 +76,22 @@ interface ReindexRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface ReindexResponse {
|
||||
batches?: number
|
||||
created?: number
|
||||
deleted?: number
|
||||
batches?: <<_long, long>>
|
||||
created?: <<_long, long>>
|
||||
deleted?: <<_long, long>>
|
||||
failures?: <<BulkIndexByScrollFailure>>[]
|
||||
noops?: number
|
||||
noops?: <<_long, long>>
|
||||
retries?: <<Retries>>
|
||||
requests_per_second?: number
|
||||
slice_id?: number
|
||||
requests_per_second?: <<_float, float>>
|
||||
slice_id?: <<_integer, integer>>
|
||||
task?: <<TaskId>>
|
||||
throttled_millis?: <<EpochTime>><<<UnitMillis>>>
|
||||
throttled_until_millis?: <<EpochTime>><<<UnitMillis>>>
|
||||
timed_out?: boolean
|
||||
took?: <<DurationValue>><<<UnitMillis>>>
|
||||
total?: number
|
||||
updated?: number
|
||||
version_conflicts?: number
|
||||
total?: <<_long, long>>
|
||||
updated?: <<_long, long>>
|
||||
version_conflicts?: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -49,7 +49,7 @@ Throttle a reindex operation. Change the number of requests per second for a par
|
||||
++++
|
||||
interface ReindexRethrottleRequest extends <<RequestBase>> {
|
||||
task_id: <<Id>>
|
||||
requests_per_second?: number
|
||||
requests_per_second?: <<_float, float>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -53,7 +53,7 @@ interface RollupPutJobRequest extends <<RequestBase>> {
|
||||
groups: RollupGroupings
|
||||
index_pattern: string
|
||||
metrics?: RollupFieldMetric[]
|
||||
page_size: number
|
||||
page_size: <<_integer, integer>>
|
||||
rollup_index: <<IndexName>>
|
||||
timeout?: <<Duration>>
|
||||
headers?: <<HttpHeaders>>
|
||||
|
||||
@ -52,10 +52,10 @@ interface RollupRollupSearchRequest extends <<RequestBase>> {
|
||||
rest_total_hits_as_int?: boolean
|
||||
typed_keys?: boolean
|
||||
aggregations?: Record<string, AggregationsAggregationContainer>
|
||||
/** @alias aggregations */
|
||||
pass:[/**] @alias aggregations */
|
||||
aggs?: Record<string, AggregationsAggregationContainer>
|
||||
query?: QueryDslQueryContainer
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -70,7 +70,7 @@ interface RollupRollupSearchRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface RollupRollupSearchResponse<TDocument = unknown, TAggregations = Record<<<AggregateName>>, AggregationsAggregate>> {
|
||||
took: number
|
||||
took: <<_long, long>>
|
||||
timed_out: boolean
|
||||
terminated_early?: boolean
|
||||
_shards: <<ShardStatistics>>
|
||||
|
||||
@ -53,7 +53,7 @@ interface SearchRequest extends <<RequestBase>> {
|
||||
allow_partial_search_results?: boolean
|
||||
analyzer?: string
|
||||
analyze_wildcard?: boolean
|
||||
batched_reduce_size?: number
|
||||
batched_reduce_size?: <<_long, long>>
|
||||
ccs_minimize_roundtrips?: boolean
|
||||
default_operator?: QueryDslOperator
|
||||
df?: string
|
||||
@ -62,17 +62,17 @@ interface SearchRequest extends <<RequestBase>> {
|
||||
ignore_unavailable?: boolean
|
||||
include_named_queries_score?: boolean
|
||||
lenient?: boolean
|
||||
max_concurrent_shard_requests?: number
|
||||
max_concurrent_shard_requests?: <<_long, long>>
|
||||
min_compatible_shard_node?: <<VersionString>>
|
||||
preference?: string
|
||||
pre_filter_shard_size?: number
|
||||
pre_filter_shard_size?: <<_long, long>>
|
||||
request_cache?: boolean
|
||||
routing?: <<Routing>>
|
||||
scroll?: <<Duration>>
|
||||
search_type?: <<SearchType>>
|
||||
suggest_field?: <<Field>>
|
||||
suggest_mode?: <<SuggestMode>>
|
||||
suggest_size?: number
|
||||
suggest_size?: <<_long, long>>
|
||||
suggest_text?: string
|
||||
typed_keys?: boolean
|
||||
rest_total_hits_as_int?: boolean
|
||||
@ -81,19 +81,19 @@ interface SearchRequest extends <<RequestBase>> {
|
||||
q?: string
|
||||
force_synthetic_source?: boolean
|
||||
aggregations?: Record<string, AggregationsAggregationContainer>
|
||||
/** @alias aggregations */
|
||||
pass:[/**] @alias aggregations */
|
||||
aggs?: Record<string, AggregationsAggregationContainer>
|
||||
collapse?: SearchFieldCollapse
|
||||
explain?: boolean
|
||||
ext?: Record<string, any>
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
highlight?: SearchHighlight
|
||||
track_total_hits?: SearchTrackHits
|
||||
indices_boost?: Record<<<IndexName>>, number>[]
|
||||
indices_boost?: Record<<<IndexName>>, <<_double, double>>>[]
|
||||
docvalue_fields?: (QueryDslFieldAndFormat | <<Field>>)[]
|
||||
knn?: <<KnnSearch>> | <<KnnSearch>>[]
|
||||
rank?: <<RankContainer>>
|
||||
min_score?: number
|
||||
min_score?: <<_double, double>>
|
||||
post_filter?: QueryDslQueryContainer
|
||||
profile?: boolean
|
||||
query?: QueryDslQueryContainer
|
||||
@ -101,13 +101,13 @@ interface SearchRequest extends <<RequestBase>> {
|
||||
retriever?: <<RetrieverContainer>>
|
||||
script_fields?: Record<string, <<ScriptField>>>
|
||||
search_after?: <<SortResults>>
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
slice?: <<SlicedScroll>>
|
||||
sort?: <<Sort>>
|
||||
_source?: SearchSourceConfig
|
||||
fields?: (QueryDslFieldAndFormat | <<Field>>)[]
|
||||
suggest?: SearchSuggester
|
||||
terminate_after?: number
|
||||
terminate_after?: <<_long, long>>
|
||||
timeout?: string
|
||||
track_scores?: boolean
|
||||
version?: boolean
|
||||
|
||||
@ -49,8 +49,8 @@ Returns the existing search applications.
|
||||
++++
|
||||
interface SearchApplicationListRequest extends <<RequestBase>> {
|
||||
q?: string
|
||||
from?: number
|
||||
size?: number
|
||||
from?: <<_integer, integer>>
|
||||
size?: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
@ -65,7 +65,7 @@ interface SearchApplicationListRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface SearchApplicationListResponse {
|
||||
count: number
|
||||
count: <<_long, long>>
|
||||
results: SearchApplicationListSearchApplicationListItem[]
|
||||
}
|
||||
|
||||
|
||||
@ -54,16 +54,16 @@ interface SearchMvtRequest extends <<RequestBase>> {
|
||||
x: SearchMvtCoordinate
|
||||
y: SearchMvtCoordinate
|
||||
aggs?: Record<string, AggregationsAggregationContainer>
|
||||
buffer?: number
|
||||
buffer?: <<_integer, integer>>
|
||||
exact_bounds?: boolean
|
||||
extent?: number
|
||||
extent?: <<_integer, integer>>
|
||||
fields?: <<Fields>>
|
||||
grid_agg?: SearchMvtGridAggregationType
|
||||
grid_precision?: number
|
||||
grid_precision?: <<_integer, integer>>
|
||||
grid_type?: SearchMvtGridType
|
||||
query?: QueryDslQueryContainer
|
||||
runtime_mappings?: MappingRuntimeFields
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
sort?: <<Sort>>
|
||||
track_total_hits?: SearchTrackHits
|
||||
with_labels?: boolean
|
||||
|
||||
@ -79,15 +79,15 @@ interface SearchTemplateRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface SearchTemplateResponse<TDocument = unknown> {
|
||||
took: number
|
||||
took: <<_long, long>>
|
||||
timed_out: boolean
|
||||
_shards: <<ShardStatistics>>
|
||||
hits: SearchHitsMetadata<TDocument>
|
||||
aggregations?: Record<<<AggregateName>>, AggregationsAggregate>
|
||||
_clusters?: <<ClusterStatistics>>
|
||||
fields?: Record<string, any>
|
||||
max_score?: number
|
||||
num_reduce_phases?: number
|
||||
max_score?: <<_double, double>>
|
||||
num_reduce_phases?: <<_long, long>>
|
||||
profile?: SearchProfile
|
||||
pit_id?: <<Id>>
|
||||
_scroll_id?: <<ScrollId>>
|
||||
|
||||
@ -68,7 +68,7 @@ interface SecurityCreateApiKeyRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface SecurityCreateApiKeyResponse {
|
||||
api_key: string
|
||||
expiration?: number
|
||||
expiration?: <<_long, long>>
|
||||
id: <<Id>>
|
||||
name: <<Name>>
|
||||
encoded: string
|
||||
|
||||
@ -65,7 +65,7 @@ interface SecurityGetServiceCredentialsRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface SecurityGetServiceCredentialsResponse {
|
||||
service_account: string
|
||||
count: number
|
||||
count: <<_integer, integer>>
|
||||
tokens: Record<string, <<Metadata>>>
|
||||
nodes_credentials: SecurityGetServiceCredentialsNodesCredentials
|
||||
}
|
||||
|
||||
@ -69,7 +69,7 @@ interface SecurityGetTokenRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface SecurityGetTokenResponse {
|
||||
access_token: string
|
||||
expires_in: number
|
||||
expires_in: <<_long, long>>
|
||||
scope?: string
|
||||
type: string
|
||||
refresh_token?: string
|
||||
|
||||
@ -68,7 +68,7 @@ interface SecurityInvalidateApiKeyRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface SecurityInvalidateApiKeyResponse {
|
||||
error_count: number
|
||||
error_count: <<_integer, integer>>
|
||||
error_details?: <<ErrorCause>>[]
|
||||
invalidated_api_keys: string[]
|
||||
previously_invalidated_api_keys: string[]
|
||||
|
||||
@ -66,10 +66,10 @@ interface SecurityInvalidateTokenRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface SecurityInvalidateTokenResponse {
|
||||
error_count: number
|
||||
error_count: <<_long, long>>
|
||||
error_details?: <<ErrorCause>>[]
|
||||
invalidated_tokens: number
|
||||
previously_invalidated_tokens: number
|
||||
invalidated_tokens: <<_long, long>>
|
||||
previously_invalidated_tokens: <<_long, long>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -52,12 +52,12 @@ interface SecurityQueryApiKeysRequest extends <<RequestBase>> {
|
||||
with_profile_uid?: boolean
|
||||
typed_keys?: boolean
|
||||
aggregations?: Record<string, SecurityQueryApiKeysApiKeyAggregationContainer>
|
||||
/** @alias aggregations */
|
||||
pass:[/**] @alias aggregations */
|
||||
aggs?: Record<string, SecurityQueryApiKeysApiKeyAggregationContainer>
|
||||
query?: SecurityQueryApiKeysApiKeyQueryContainer
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
sort?: <<Sort>>
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
search_after?: <<SortResults>>
|
||||
}
|
||||
|
||||
@ -73,8 +73,8 @@ interface SecurityQueryApiKeysRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface SecurityQueryApiKeysResponse {
|
||||
total: number
|
||||
count: number
|
||||
total: <<_integer, integer>>
|
||||
count: <<_integer, integer>>
|
||||
api_keys: SecurityApiKey[]
|
||||
aggregations?: Record<<<AggregateName>>, SecurityQueryApiKeysApiKeyAggregate>
|
||||
}
|
||||
|
||||
@ -49,9 +49,9 @@ Find roles with a query. Get roles in a paginated manner. You can optionally fil
|
||||
++++
|
||||
interface SecurityQueryRoleRequest extends <<RequestBase>> {
|
||||
query?: SecurityQueryRoleRoleQueryContainer
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
sort?: <<Sort>>
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
search_after?: <<SortResults>>
|
||||
}
|
||||
|
||||
@ -67,8 +67,8 @@ interface SecurityQueryRoleRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface SecurityQueryRoleResponse {
|
||||
total: number
|
||||
count: number
|
||||
total: <<_integer, integer>>
|
||||
count: <<_integer, integer>>
|
||||
roles: SecurityQueryRoleQueryRole[]
|
||||
}
|
||||
|
||||
|
||||
@ -50,9 +50,9 @@ Find users with a query. Get information for users in a paginated manner. You ca
|
||||
interface SecurityQueryUserRequest extends <<RequestBase>> {
|
||||
with_profile_uid?: boolean
|
||||
query?: SecurityQueryUserUserQueryContainer
|
||||
from?: number
|
||||
from?: <<_integer, integer>>
|
||||
sort?: <<Sort>>
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
search_after?: <<SortResults>>
|
||||
}
|
||||
|
||||
@ -68,8 +68,8 @@ interface SecurityQueryUserRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface SecurityQueryUserResponse {
|
||||
total: number
|
||||
count: number
|
||||
total: <<_integer, integer>>
|
||||
count: <<_integer, integer>>
|
||||
users: SecurityQueryUserQueryUser[]
|
||||
}
|
||||
|
||||
|
||||
@ -67,7 +67,7 @@ interface SecuritySamlAuthenticateRequest extends <<RequestBase>> {
|
||||
interface SecuritySamlAuthenticateResponse {
|
||||
access_token: string
|
||||
username: string
|
||||
expires_in: number
|
||||
expires_in: <<_integer, integer>>
|
||||
refresh_token: string
|
||||
realm: string
|
||||
}
|
||||
|
||||
@ -65,7 +65,7 @@ interface SecuritySamlInvalidateRequest extends <<RequestBase>> {
|
||||
<pre>
|
||||
++++
|
||||
interface SecuritySamlInvalidateResponse {
|
||||
invalidated: number
|
||||
invalidated: <<_integer, integer>>
|
||||
realm: string
|
||||
redirect: string
|
||||
}
|
||||
|
||||
@ -49,7 +49,7 @@ Suggest a user profile. Get suggestions for user profiles that match specified s
|
||||
++++
|
||||
interface SecuritySuggestUserProfilesRequest extends <<RequestBase>> {
|
||||
name?: string
|
||||
size?: number
|
||||
size?: <<_long, long>>
|
||||
data?: string | string[]
|
||||
hint?: SecuritySuggestUserProfilesHint
|
||||
}
|
||||
@ -67,7 +67,7 @@ interface SecuritySuggestUserProfilesRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface SecuritySuggestUserProfilesResponse {
|
||||
total: SecuritySuggestUserProfilesTotalUserProfiles
|
||||
took: number
|
||||
took: <<_long, long>>
|
||||
profiles: SecurityUserProfile[]
|
||||
}
|
||||
|
||||
|
||||
@ -50,7 +50,7 @@ Update user profile data. Update specific data for the user profile that is asso
|
||||
interface SecurityUpdateUserProfileDataRequest extends <<RequestBase>> {
|
||||
uid: SecurityUserProfileId
|
||||
if_seq_no?: <<SequenceNumber>>
|
||||
if_primary_term?: number
|
||||
if_primary_term?: <<_long, long>>
|
||||
refresh?: <<Refresh>>
|
||||
labels?: Record<string, any>
|
||||
data?: Record<string, any>
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
@ -64,13 +64,13 @@ interface SlmGetStatsRequest extends <<RequestBase>> {
|
||||
interface SlmGetStatsResponse {
|
||||
retention_deletion_time: <<Duration>>
|
||||
retention_deletion_time_millis: <<DurationValue>><<<UnitMillis>>>
|
||||
retention_failed: number
|
||||
retention_runs: number
|
||||
retention_timed_out: number
|
||||
total_snapshots_deleted: number
|
||||
total_snapshot_deletion_failures: number
|
||||
total_snapshots_failed: number
|
||||
total_snapshots_taken: number
|
||||
retention_failed: <<_long, long>>
|
||||
retention_runs: <<_long, long>>
|
||||
retention_timed_out: <<_long, long>>
|
||||
total_snapshots_deleted: <<_long, long>>
|
||||
total_snapshot_deletion_failures: <<_long, long>>
|
||||
total_snapshots_failed: <<_long, long>>
|
||||
total_snapshots_taken: <<_long, long>>
|
||||
policy_stats: string[]
|
||||
}
|
||||
|
||||
|
||||
@ -57,10 +57,10 @@ interface SnapshotGetRequest extends <<RequestBase>> {
|
||||
index_names?: boolean
|
||||
include_repository?: boolean
|
||||
sort?: SnapshotSnapshotSort
|
||||
size?: number
|
||||
size?: <<_integer, integer>>
|
||||
order?: <<SortOrder>>
|
||||
after?: string
|
||||
offset?: number
|
||||
offset?: <<_integer, integer>>
|
||||
from_sort_value?: string
|
||||
slm_policy_filter?: <<Name>>
|
||||
}
|
||||
@ -79,8 +79,8 @@ interface SnapshotGetRequest extends <<RequestBase>> {
|
||||
interface SnapshotGetResponse {
|
||||
responses?: SnapshotGetSnapshotResponseItem[]
|
||||
snapshots?: SnapshotSnapshotInfo[]
|
||||
total: number
|
||||
remaining: number
|
||||
total: <<_integer, integer>>
|
||||
remaining: <<_integer, integer>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -68,7 +68,7 @@ interface SqlGetAsyncStatusResponse {
|
||||
is_partial: boolean
|
||||
start_time_in_millis: <<EpochTime>><<<UnitMillis>>>
|
||||
expiration_time_in_millis: <<EpochTime>><<<UnitMillis>>>
|
||||
completion_status?: number
|
||||
completion_status?: <<_uint, uint>>
|
||||
}
|
||||
|
||||
[pass]
|
||||
|
||||
@ -52,7 +52,7 @@ interface SqlQueryRequest extends <<RequestBase>> {
|
||||
catalog?: string
|
||||
columnar?: boolean
|
||||
cursor?: string
|
||||
fetch_size?: number
|
||||
fetch_size?: <<_integer, integer>>
|
||||
filter?: QueryDslQueryContainer
|
||||
query?: string
|
||||
request_timeout?: <<Duration>>
|
||||
|
||||
@ -48,7 +48,7 @@ Translates SQL into Elasticsearch queries
|
||||
<pre>
|
||||
++++
|
||||
interface SqlTranslateRequest extends <<RequestBase>> {
|
||||
fetch_size?: number
|
||||
fetch_size?: <<_integer, integer>>
|
||||
filter?: QueryDslQueryContainer
|
||||
query: string
|
||||
time_zone?: <<TimeZone>>
|
||||
@ -67,7 +67,7 @@ interface SqlTranslateRequest extends <<RequestBase>> {
|
||||
++++
|
||||
interface SqlTranslateResponse {
|
||||
aggregations?: Record<string, AggregationsAggregationContainer>
|
||||
size?: number
|
||||
size?: <<_long, long>>
|
||||
_source?: SearchSourceConfig
|
||||
fields?: (QueryDslFieldAndFormat | <<Field>>)[]
|
||||
query?: QueryDslQueryContainer
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user