Still testing

This commit is contained in:
Josh Mock
2024-11-20 11:39:21 -06:00
parent be8fc5ae83
commit 15a43f3747
115 changed files with 609 additions and 818 deletions

View File

@ -56,7 +56,7 @@ interface AsyncSearchSubmitRequest extends <<RequestBase>> {
allow_partial_search_results?: boolean
analyzer?: string
analyze_wildcard?: boolean
batched_reduce_size?: number
batched_reduce_size?: <<_long, long>>
ccs_minimize_roundtrips?: boolean
default_operator?: QueryDslOperator
df?: string
@ -64,17 +64,17 @@ interface AsyncSearchSubmitRequest extends <<RequestBase>> {
ignore_throttled?: boolean
ignore_unavailable?: boolean
lenient?: boolean
max_concurrent_shard_requests?: number
max_concurrent_shard_requests?: <<_long, long>>
min_compatible_shard_node?: <<VersionString>>
preference?: string
pre_filter_shard_size?: number
pre_filter_shard_size?: <<_long, long>>
request_cache?: boolean
routing?: <<Routing>>
scroll?: <<Duration>>
search_type?: <<SearchType>>
suggest_field?: <<Field>>
suggest_mode?: <<SuggestMode>>
suggest_size?: number
suggest_size?: <<_long, long>>
suggest_text?: string
typed_keys?: boolean
rest_total_hits_as_int?: boolean
@ -82,31 +82,31 @@ interface AsyncSearchSubmitRequest extends <<RequestBase>> {
_source_includes?: <<Fields>>
q?: string
aggregations?: Record<string, AggregationsAggregationContainer>
/** @alias aggregations */
pass:[/**] @alias aggregations */
aggs?: Record<string, AggregationsAggregationContainer>
collapse?: SearchFieldCollapse
explain?: boolean
ext?: Record<string, any>
from?: number
from?: <<_integer, integer>>
highlight?: SearchHighlight
track_total_hits?: SearchTrackHits
indices_boost?: Record<<<IndexName>>, number>[]
indices_boost?: Record<<<IndexName>>, <<_double, double>>>[]
docvalue_fields?: (QueryDslFieldAndFormat | <<Field>>)[]
knn?: <<KnnSearch>> | <<KnnSearch>>[]
min_score?: number
min_score?: <<_double, double>>
post_filter?: QueryDslQueryContainer
profile?: boolean
query?: QueryDslQueryContainer
rescore?: SearchRescore | SearchRescore[]
script_fields?: Record<string, <<ScriptField>>>
search_after?: <<SortResults>>
size?: number
size?: <<_integer, integer>>
slice?: <<SlicedScroll>>
sort?: <<Sort>>
_source?: SearchSourceConfig
fields?: (QueryDslFieldAndFormat | <<Field>>)[]
suggest?: SearchSuggester
terminate_after?: number
terminate_after?: <<_long, long>>
timeout?: string
track_scores?: boolean
version?: boolean

View File

@ -75,8 +75,8 @@ interface BulkRequest<TDocument = unknown, TPartialDocument = unknown> extends <
interface BulkResponse {
errors: boolean
items: Partial<Record<BulkOperationType, BulkResponseItem>>[]
took: number
ingest_took?: number
took: <<_long, long>>
ingest_took?: <<_long, long>>
}
[pass]

View File

@ -53,8 +53,8 @@ interface CatMlTrainedModelsRequest extends CatCatRequestBase {
bytes?: <<Bytes>>
h?: CatCatTrainedModelsColumns
s?: CatCatTrainedModelsColumns
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
}
[pass]

View File

@ -50,11 +50,11 @@ Get transforms. Returns configuration and usage information about transforms. CA
interface CatTransformsRequest extends CatCatRequestBase {
transform_id?: <<Id>>
allow_no_match?: boolean
from?: number
from?: <<_integer, integer>>
h?: CatCatTransformColumns
s?: CatCatTransformColumns
time?: <<TimeUnit>>
size?: number
size?: <<_integer, integer>>
}
[pass]

View File

@ -51,14 +51,14 @@ interface CcrFollowRequest extends <<RequestBase>> {
index: <<IndexName>>
wait_for_active_shards?: <<WaitForActiveShards>>
leader_index?: <<IndexName>>
max_outstanding_read_requests?: number
max_outstanding_write_requests?: number
max_read_request_operation_count?: number
max_outstanding_read_requests?: <<_long, long>>
max_outstanding_write_requests?: <<_long, long>>
max_read_request_operation_count?: <<_long, long>>
max_read_request_size?: string
max_retry_delay?: <<Duration>>
max_write_buffer_count?: number
max_write_buffer_count?: <<_long, long>>
max_write_buffer_size?: string
max_write_request_operation_count?: number
max_write_request_operation_count?: <<_long, long>>
max_write_request_size?: string
read_poll_timeout?: <<Duration>>
remote_cluster?: string

View File

@ -53,16 +53,16 @@ interface CcrPutAutoFollowPatternRequest extends <<RequestBase>> {
follow_index_pattern?: <<IndexPattern>>
leader_index_patterns?: <<IndexPatterns>>
leader_index_exclusion_patterns?: <<IndexPatterns>>
max_outstanding_read_requests?: number
max_outstanding_read_requests?: <<_integer, integer>>
settings?: Record<string, any>
max_outstanding_write_requests?: number
max_outstanding_write_requests?: <<_integer, integer>>
read_poll_timeout?: <<Duration>>
max_read_request_operation_count?: number
max_read_request_operation_count?: <<_integer, integer>>
max_read_request_size?: <<ByteSize>>
max_retry_delay?: <<Duration>>
max_write_buffer_count?: number
max_write_buffer_count?: <<_integer, integer>>
max_write_buffer_size?: <<ByteSize>>
max_write_request_operation_count?: number
max_write_request_operation_count?: <<_integer, integer>>
max_write_request_size?: <<ByteSize>>
}

View File

@ -49,14 +49,14 @@ Resumes a follower index that has been paused
++++
interface CcrResumeFollowRequest extends <<RequestBase>> {
index: <<IndexName>>
max_outstanding_read_requests?: number
max_outstanding_write_requests?: number
max_read_request_operation_count?: number
max_outstanding_read_requests?: <<_long, long>>
max_outstanding_write_requests?: <<_long, long>>
max_read_request_operation_count?: <<_long, long>>
max_read_request_size?: string
max_retry_delay?: <<Duration>>
max_write_buffer_count?: number
max_write_buffer_count?: <<_long, long>>
max_write_buffer_size?: string
max_write_request_operation_count?: number
max_write_request_operation_count?: <<_long, long>>
max_write_request_size?: string
read_poll_timeout?: <<Duration>>
}

View File

@ -64,7 +64,7 @@ interface ClearScrollRequest extends <<RequestBase>> {
++++
interface ClearScrollResponse {
succeeded: boolean
num_freed: number
num_freed: <<_integer, integer>>
}
[pass]

View File

@ -64,7 +64,7 @@ interface ClosePointInTimeRequest extends <<RequestBase>> {
++++
interface ClosePointInTimeResponse {
succeeded: boolean
num_freed: number
num_freed: <<_integer, integer>>
}
[pass]

View File

@ -53,7 +53,7 @@ interface ClusterAllocationExplainRequest extends <<RequestBase>> {
current_node?: string
index?: <<IndexName>>
primary?: boolean
shard?: number
shard?: <<_integer, integer>>
}
[pass]
@ -90,7 +90,7 @@ interface ClusterAllocationExplainResponse {
rebalance_explanation?: string
remaining_delay?: <<Duration>>
remaining_delay_in_millis?: <<DurationValue>><<<UnitMillis>>>
shard: number
shard: <<_integer, integer>>
unassigned_info?: ClusterAllocationExplainUnassignedInformation
note?: string
}

View File

@ -56,7 +56,7 @@ interface ClusterHealthRequest extends <<RequestBase>> {
timeout?: <<Duration>>
wait_for_active_shards?: <<WaitForActiveShards>>
wait_for_events?: <<WaitForEvents>>
wait_for_nodes?: string | number
wait_for_nodes?: string | <<_integer, integer>>
wait_for_no_initializing_shards?: boolean
wait_for_no_relocating_shards?: boolean
wait_for_status?: <<HealthStatus>>

View File

@ -48,8 +48,8 @@ Get all connectors. Get information about all connectors.
<pre>
++++
interface ConnectorListRequest extends <<RequestBase>> {
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
index_name?: <<Indices>>
connector_name?: <<Names>>
service_type?: <<Names>>
@ -68,7 +68,7 @@ interface ConnectorListRequest extends <<RequestBase>> {
<pre>
++++
interface ConnectorListResponse {
count: number
count: <<_long, long>>
results: ConnectorConnector[]
}

View File

@ -48,8 +48,8 @@ Get all connector sync jobs. Get information about all stored connector sync job
<pre>
++++
interface ConnectorSyncJobListRequest extends <<RequestBase>> {
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
status?: ConnectorSyncStatus
connector_id?: <<Id>>
job_type?: ConnectorSyncJobType | ConnectorSyncJobType[]
@ -67,7 +67,7 @@ interface ConnectorSyncJobListRequest extends <<RequestBase>> {
<pre>
++++
interface ConnectorSyncJobListResponse {
count: number
count: <<_long, long>>
results: ConnectorConnectorSyncJob[]
}

View File

@ -58,10 +58,10 @@ interface CountRequest extends <<RequestBase>> {
ignore_throttled?: boolean
ignore_unavailable?: boolean
lenient?: boolean
min_score?: number
min_score?: <<_double, double>>
preference?: string
routing?: <<Routing>>
terminate_after?: number
terminate_after?: <<_long, long>>
q?: string
query?: QueryDslQueryContainer
}
@ -78,7 +78,7 @@ interface CountRequest extends <<RequestBase>> {
<pre>
++++
interface CountResponse {
count: number
count: <<_long, long>>
_shards: <<ShardStatistics>>
}

View File

@ -50,7 +50,7 @@ Delete a document. Removes a JSON document from the specified index.
interface DeleteRequest extends <<RequestBase>> {
id: <<Id>>
index: <<IndexName>>
if_primary_term?: number
if_primary_term?: <<_long, long>>
if_seq_no?: <<SequenceNumber>>
refresh?: <<Refresh>>
routing?: <<Routing>>

View File

@ -56,28 +56,28 @@ interface DeleteByQueryRequest extends <<RequestBase>> {
default_operator?: QueryDslOperator
df?: string
expand_wildcards?: <<ExpandWildcards>>
from?: number
from?: <<_long, long>>
ignore_unavailable?: boolean
lenient?: boolean
preference?: string
refresh?: boolean
request_cache?: boolean
requests_per_second?: number
requests_per_second?: <<_float, float>>
routing?: <<Routing>>
q?: string
scroll?: <<Duration>>
scroll_size?: number
scroll_size?: <<_long, long>>
search_timeout?: <<Duration>>
search_type?: <<SearchType>>
slices?: <<Slices>>
sort?: string[]
stats?: string[]
terminate_after?: number
terminate_after?: <<_long, long>>
timeout?: <<Duration>>
version?: boolean
wait_for_active_shards?: <<WaitForActiveShards>>
wait_for_completion?: boolean
max_docs?: number
max_docs?: <<_long, long>>
query?: QueryDslQueryContainer
slice?: <<SlicedScroll>>
}
@ -94,13 +94,13 @@ interface DeleteByQueryRequest extends <<RequestBase>> {
<pre>
++++
interface DeleteByQueryResponse {
batches?: number
deleted?: number
batches?: <<_long, long>>
deleted?: <<_long, long>>
failures?: <<BulkIndexByScrollFailure>>[]
noops?: number
requests_per_second?: number
noops?: <<_long, long>>
requests_per_second?: <<_float, float>>
retries?: <<Retries>>
slice_id?: number
slice_id?: <<_integer, integer>>
task?: <<TaskId>>
throttled?: <<Duration>>
throttled_millis?: <<DurationValue>><<<UnitMillis>>>
@ -108,8 +108,8 @@ interface DeleteByQueryResponse {
throttled_until_millis?: <<DurationValue>><<<UnitMillis>>>
timed_out?: boolean
took?: <<DurationValue>><<<UnitMillis>>>
total?: number
version_conflicts?: number
total?: <<_long, long>>
version_conflicts?: <<_long, long>>
}
[pass]

View File

@ -49,7 +49,7 @@ Throttle a delete by query operation. Change the number of requests per second f
++++
interface DeleteByQueryRethrottleRequest extends <<RequestBase>> {
task_id: <<TaskId>>
requests_per_second?: number
requests_per_second?: <<_float, float>>
}
[pass]

View File

@ -68,7 +68,7 @@ interface EqlGetStatusResponse {
is_running: boolean
start_time_in_millis?: <<EpochTime>><<<UnitMillis>>>
expiration_time_in_millis?: <<EpochTime>><<<UnitMillis>>>
completion_status?: number
completion_status?: <<_integer, integer>>
}
[pass]

View File

@ -57,12 +57,12 @@ interface EqlSearchRequest extends <<RequestBase>> {
event_category_field?: <<Field>>
tiebreaker_field?: <<Field>>
timestamp_field?: <<Field>>
fetch_size?: number
fetch_size?: <<_uint, uint>>
filter?: QueryDslQueryContainer | QueryDslQueryContainer[]
keep_alive?: <<Duration>>
keep_on_completion?: boolean
wait_for_completion_timeout?: <<Duration>>
size?: number
size?: <<_uint, uint>>
fields?: QueryDslFieldAndFormat | <<Field>> | (QueryDslFieldAndFormat | <<Field>>)[]
result_position?: EqlSearchResultPosition
runtime_mappings?: MappingRuntimeFields

View File

@ -52,9 +52,9 @@ interface FleetMsearchRequest extends <<RequestBase>> {
expand_wildcards?: <<ExpandWildcards>>
ignore_throttled?: boolean
ignore_unavailable?: boolean
max_concurrent_searches?: number
max_concurrent_shard_requests?: number
pre_filter_shard_size?: number
max_concurrent_searches?: <<_long, long>>
max_concurrent_shard_requests?: <<_long, long>>
pre_filter_shard_size?: <<_long, long>>
search_type?: <<SearchType>>
rest_total_hits_as_int?: boolean
typed_keys?: boolean

View File

@ -50,7 +50,7 @@ interface FleetSearchRequest extends <<RequestBase>> {
allow_no_indices?: boolean
analyzer?: string
analyze_wildcard?: boolean
batched_reduce_size?: number
batched_reduce_size?: <<_long, long>>
ccs_minimize_roundtrips?: boolean
default_operator?: QueryDslOperator
df?: string
@ -58,17 +58,17 @@ interface FleetSearchRequest extends <<RequestBase>> {
ignore_throttled?: boolean
ignore_unavailable?: boolean
lenient?: boolean
max_concurrent_shard_requests?: number
max_concurrent_shard_requests?: <<_long, long>>
min_compatible_shard_node?: <<VersionString>>
preference?: string
pre_filter_shard_size?: number
pre_filter_shard_size?: <<_long, long>>
request_cache?: boolean
routing?: <<Routing>>
scroll?: <<Duration>>
search_type?: <<SearchType>>
suggest_field?: <<Field>>
suggest_mode?: <<SuggestMode>>
suggest_size?: number
suggest_size?: <<_long, long>>
suggest_text?: string
typed_keys?: boolean
rest_total_hits_as_int?: boolean
@ -78,30 +78,30 @@ interface FleetSearchRequest extends <<RequestBase>> {
wait_for_checkpoints?: FleetCheckpoint[]
allow_partial_search_results?: boolean
aggregations?: Record<string, AggregationsAggregationContainer>
/** @alias aggregations */
pass:[/**] @alias aggregations */
aggs?: Record<string, AggregationsAggregationContainer>
collapse?: SearchFieldCollapse
explain?: boolean
ext?: Record<string, any>
from?: number
from?: <<_integer, integer>>
highlight?: SearchHighlight
track_total_hits?: SearchTrackHits
indices_boost?: Record<<<IndexName>>, number>[]
indices_boost?: Record<<<IndexName>>, <<_double, double>>>[]
docvalue_fields?: (QueryDslFieldAndFormat | <<Field>>)[]
min_score?: number
min_score?: <<_double, double>>
post_filter?: QueryDslQueryContainer
profile?: boolean
query?: QueryDslQueryContainer
rescore?: SearchRescore | SearchRescore[]
script_fields?: Record<string, <<ScriptField>>>
search_after?: <<SortResults>>
size?: number
size?: <<_integer, integer>>
slice?: <<SlicedScroll>>
sort?: <<Sort>>
_source?: SearchSourceConfig
fields?: (QueryDslFieldAndFormat | <<Field>>)[]
suggest?: SearchSuggester
terminate_after?: number
terminate_after?: <<_long, long>>
timeout?: string
track_scores?: boolean
version?: boolean
@ -124,15 +124,15 @@ interface FleetSearchRequest extends <<RequestBase>> {
<pre>
++++
interface FleetSearchResponse<TDocument = unknown> {
took: number
took: <<_long, long>>
timed_out: boolean
_shards: <<ShardStatistics>>
hits: SearchHitsMetadata<TDocument>
aggregations?: Record<<<AggregateName>>, AggregationsAggregate>
_clusters?: <<ClusterStatistics>>
fields?: Record<string, any>
max_score?: number
num_reduce_phases?: number
max_score?: <<_double, double>>
num_reduce_phases?: <<_long, long>>
profile?: SearchProfile
pit_id?: <<Id>>
_scroll_id?: <<ScrollId>>

View File

@ -72,7 +72,7 @@ interface GraphExploreResponse {
connections: GraphConnection[]
failures: <<ShardFailure>>[]
timed_out: boolean
took: number
took: <<_long, long>>
vertices: GraphVertex[]
}

View File

@ -51,7 +51,7 @@ interface HealthReportRequest extends <<RequestBase>> {
feature?: string | string[]
timeout?: <<Duration>>
verbose?: boolean
size?: number
size?: <<_integer, integer>>
}
[pass]

View File

@ -50,7 +50,7 @@ Index a document. Adds a JSON document to the specified data stream or index and
interface IndexRequest<TDocument = unknown> extends <<RequestBase>> {
id?: <<Id>>
index: <<IndexName>>
if_primary_term?: number
if_primary_term?: <<_long, long>>
if_seq_no?: <<SequenceNumber>>
op_type?: <<OpType>>
pipeline?: string

View File

@ -65,11 +65,11 @@ interface IndicesDataStreamsStatsRequest extends <<RequestBase>> {
++++
interface IndicesDataStreamsStatsResponse {
_shards: <<ShardStatistics>>
backing_indices: number
data_stream_count: number
backing_indices: <<_integer, integer>>
data_stream_count: <<_integer, integer>>
data_streams: IndicesDataStreamsStatsDataStreamsStatsItem[]
total_store_sizes?: <<ByteSize>>
total_store_size_bytes: number
total_store_size_bytes: <<_long, long>>
}
[pass]

View File

@ -53,7 +53,7 @@ interface IndicesForcemergeRequest extends <<RequestBase>> {
expand_wildcards?: <<ExpandWildcards>>
flush?: boolean
ignore_unavailable?: boolean
max_num_segments?: number
max_num_segments?: <<_long, long>>
only_expunge_deletes?: boolean
wait_for_completion?: boolean
}

View File

@ -56,7 +56,7 @@ interface IndicesPutIndexTemplateRequest extends <<RequestBase>> {
composed_of?: <<Name>>[]
template?: IndicesPutIndexTemplateIndexTemplateMapping
data_stream?: IndicesDataStreamVisibility
priority?: number
priority?: <<_long, long>>
version?: <<VersionNumber>>
_meta?: <<Metadata>>
allow_auto_create?: boolean

View File

@ -55,7 +55,7 @@ interface IndicesPutTemplateRequest extends <<RequestBase>> {
aliases?: Record<<<IndexName>>, IndicesAlias>
index_patterns?: string | string[]
mappings?: MappingTypeMapping
order?: number
order?: <<_integer, integer>>
settings?: IndicesIndexSettings
version?: <<VersionNumber>>
}

View File

@ -57,7 +57,7 @@ interface IndicesSimulateTemplateRequest extends <<RequestBase>> {
composed_of?: <<Name>>[]
template?: IndicesPutIndexTemplateIndexTemplateMapping
data_stream?: IndicesDataStreamVisibility
priority?: number
priority?: <<_long, long>>
version?: <<VersionNumber>>
_meta?: <<Metadata>>
ignore_missing_component_templates?: string[]

View File

@ -70,12 +70,12 @@ interface KnnSearchRequest extends <<RequestBase>> {
<pre>
++++
interface KnnSearchResponse<TDocument = unknown> {
took: number
took: <<_long, long>>
timed_out: boolean
_shards: <<ShardStatistics>>
hits: SearchHitsMetadata<TDocument>
fields?: Record<string, any>
max_score?: number
max_score?: <<_double, double>>
}
[pass]

View File

@ -1,3 +1,5 @@
[[reference-main-index]]
== API Reference

View File

@ -49,7 +49,7 @@ Delete expired ML data. Deletes all job results, model snapshots and forecast da
++++
interface MlDeleteExpiredDataRequest extends <<RequestBase>> {
job_id?: <<Id>>
requests_per_second?: number
requests_per_second?: <<_float, float>>
timeout?: <<Duration>>
}

View File

@ -49,8 +49,8 @@ Estimate job model memory usage. Makes an estimation of the memory usage for an
++++
interface MlEstimateModelMemoryRequest extends <<RequestBase>> {
analysis_config?: MlAnalysisConfig
max_bucket_cardinality?: Record<<<Field>>, number>
overall_cardinality?: Record<<<Field>>, number>
max_bucket_cardinality?: Record<<<Field>>, <<_long, long>>>
overall_cardinality?: Record<<<Field>>, <<_long, long>>>
}
[pass]

View File

@ -54,7 +54,7 @@ interface MlExplainDataFrameAnalyticsRequest extends <<RequestBase>> {
analysis?: MlDataframeAnalysisContainer
description?: string
model_memory_limit?: string
max_num_threads?: number
max_num_threads?: <<_integer, integer>>
analyzed_fields?: MlDataframeAnalysisAnalyzedFields | string[]
allow_lazy_start?: boolean
}

View File

@ -69,7 +69,7 @@ interface MlFlushJobRequest extends <<RequestBase>> {
++++
interface MlFlushJobResponse {
flushed: boolean
last_finalized_bucket_end?: number
last_finalized_bucket_end?: <<_integer, integer>>
}
[pass]

View File

@ -50,9 +50,9 @@ Get anomaly detection job results for buckets. The API presents a chronological
interface MlGetBucketsRequest extends <<RequestBase>> {
job_id: <<Id>>
timestamp?: <<DateTime>>
from?: number
size?: number
anomaly_score?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
anomaly_score?: <<_double, double>>
desc?: boolean
end?: <<DateTime>>
exclude_interim?: boolean
@ -75,7 +75,7 @@ interface MlGetBucketsRequest extends <<RequestBase>> {
++++
interface MlGetBucketsResponse {
buckets: MlBucketSummary[]
count: number
count: <<_long, long>>
}
[pass]

View File

@ -50,9 +50,9 @@ Get info about events in calendars.
interface MlGetCalendarEventsRequest extends <<RequestBase>> {
calendar_id: <<Id>>
end?: <<DateTime>>
from?: number
from?: <<_integer, integer>>
job_id?: <<Id>>
size?: number
size?: <<_integer, integer>>
start?: <<DateTime>>
}
@ -68,7 +68,7 @@ interface MlGetCalendarEventsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetCalendarEventsResponse {
count: number
count: <<_long, long>>
events: MlCalendarEvent[]
}

View File

@ -49,8 +49,8 @@ Get calendar configuration info.
++++
interface MlGetCalendarsRequest extends <<RequestBase>> {
calendar_id?: <<Id>>
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
page?: MlPage
}
@ -67,7 +67,7 @@ interface MlGetCalendarsRequest extends <<RequestBase>> {
++++
interface MlGetCalendarsResponse {
calendars: MlGetCalendarsCalendar[]
count: number
count: <<_long, long>>
}
[pass]

View File

@ -50,9 +50,9 @@ Get anomaly detection job results for categories.
interface MlGetCategoriesRequest extends <<RequestBase>> {
job_id: <<Id>>
category_id?: <<CategoryId>>
from?: number
from?: <<_integer, integer>>
partition_field_value?: string
size?: number
size?: <<_integer, integer>>
page?: MlPage
}
@ -69,7 +69,7 @@ interface MlGetCategoriesRequest extends <<RequestBase>> {
++++
interface MlGetCategoriesResponse {
categories: MlCategory[]
count: number
count: <<_long, long>>
}
[pass]

View File

@ -50,8 +50,8 @@ Get data frame analytics job configuration info. You can get information for mul
interface MlGetDataFrameAnalyticsRequest extends <<RequestBase>> {
id?: <<Id>>
allow_no_match?: boolean
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
exclude_generated?: boolean
}
@ -67,7 +67,7 @@ interface MlGetDataFrameAnalyticsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetDataFrameAnalyticsResponse {
count: number
count: <<_integer, integer>>
data_frame_analytics: MlDataframeAnalyticsSummary[]
}

View File

@ -50,8 +50,8 @@ Get data frame analytics jobs usage info.
interface MlGetDataFrameAnalyticsStatsRequest extends <<RequestBase>> {
id?: <<Id>>
allow_no_match?: boolean
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
verbose?: boolean
}
@ -67,7 +67,7 @@ interface MlGetDataFrameAnalyticsStatsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetDataFrameAnalyticsStatsResponse {
count: number
count: <<_long, long>>
data_frame_analytics: MlDataframeAnalytics[]
}

View File

@ -64,7 +64,7 @@ interface MlGetDatafeedStatsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetDatafeedStatsResponse {
count: number
count: <<_long, long>>
datafeeds: MlDatafeedStats[]
}

View File

@ -65,7 +65,7 @@ interface MlGetDatafeedsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetDatafeedsResponse {
count: number
count: <<_long, long>>
datafeeds: MlDatafeed[]
}

View File

@ -49,8 +49,8 @@ Get filters. You can get a single filter or all filters.
++++
interface MlGetFiltersRequest extends <<RequestBase>> {
filter_id?: <<Ids>>
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
}
[pass]
@ -65,7 +65,7 @@ interface MlGetFiltersRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetFiltersResponse {
count: number
count: <<_long, long>>
filters: MlFilter[]
}

View File

@ -52,9 +52,9 @@ interface MlGetInfluencersRequest extends <<RequestBase>> {
desc?: boolean
end?: <<DateTime>>
exclude_interim?: boolean
influencer_score?: number
from?: number
size?: number
influencer_score?: <<_double, double>>
from?: <<_integer, integer>>
size?: <<_integer, integer>>
sort?: <<Field>>
start?: <<DateTime>>
page?: MlPage
@ -72,7 +72,7 @@ interface MlGetInfluencersRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetInfluencersResponse {
count: number
count: <<_long, long>>
influencers: MlInfluencer[]
}

View File

@ -64,7 +64,7 @@ interface MlGetJobStatsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetJobStatsResponse {
count: number
count: <<_long, long>>
jobs: MlJobStats[]
}

View File

@ -65,7 +65,7 @@ interface MlGetJobsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetJobsResponse {
count: number
count: <<_long, long>>
jobs: MlJob[]
}

View File

@ -65,7 +65,7 @@ interface MlGetModelSnapshotUpgradeStatsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetModelSnapshotUpgradeStatsResponse {
count: number
count: <<_long, long>>
model_snapshot_upgrades: MlModelSnapshotUpgrade[]
}

View File

@ -50,8 +50,8 @@ Get model snapshots info.
interface MlGetModelSnapshotsRequest extends <<RequestBase>> {
job_id: <<Id>>
snapshot_id?: <<Id>>
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
desc?: boolean
end?: <<DateTime>>
page?: MlPage
@ -71,7 +71,7 @@ interface MlGetModelSnapshotsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetModelSnapshotsResponse {
count: number
count: <<_long, long>>
model_snapshots: MlModelSnapshot[]
}

View File

@ -53,9 +53,9 @@ interface MlGetOverallBucketsRequest extends <<RequestBase>> {
bucket_span?: <<Duration>>
end?: <<DateTime>>
exclude_interim?: boolean
overall_score?: number | string
overall_score?: <<_double, double>> | string
start?: <<DateTime>>
top_n?: number
top_n?: <<_integer, integer>>
}
[pass]
@ -70,7 +70,7 @@ interface MlGetOverallBucketsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetOverallBucketsResponse {
count: number
count: <<_long, long>>
overall_buckets: MlOverallBucket[]
}

View File

@ -49,13 +49,13 @@ Get anomaly records for an anomaly detection job. Records contain the detailed a
++++
interface MlGetRecordsRequest extends <<RequestBase>> {
job_id: <<Id>>
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
desc?: boolean
end?: <<DateTime>>
exclude_interim?: boolean
page?: MlPage
record_score?: number
record_score?: <<_double, double>>
sort?: <<Field>>
start?: <<DateTime>>
}
@ -72,7 +72,7 @@ interface MlGetRecordsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetRecordsResponse {
count: number
count: <<_long, long>>
records: MlAnomaly[]
}

View File

@ -52,9 +52,9 @@ interface MlGetTrainedModelsRequest extends <<RequestBase>> {
allow_no_match?: boolean
decompress_definition?: boolean
exclude_generated?: boolean
from?: number
from?: <<_integer, integer>>
include?: MlInclude
size?: number
size?: <<_integer, integer>>
tags?: string | string[]
}
@ -70,7 +70,7 @@ interface MlGetTrainedModelsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetTrainedModelsResponse {
count: number
count: <<_integer, integer>>
trained_model_configs: MlTrainedModelConfig[]
}

View File

@ -50,8 +50,8 @@ Get trained models usage info. You can get usage information for multiple traine
interface MlGetTrainedModelsStatsRequest extends <<RequestBase>> {
model_id?: <<Ids>>
allow_no_match?: boolean
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
}
[pass]
@ -66,7 +66,7 @@ interface MlGetTrainedModelsStatsRequest extends <<RequestBase>> {
<pre>
++++
interface MlGetTrainedModelsStatsResponse {
count: number
count: <<_integer, integer>>
trained_model_stats: MlTrainedModelStats[]
}

View File

@ -66,21 +66,21 @@ interface MlPostDataRequest<TData = unknown> extends <<RequestBase>> {
<pre>
++++
interface MlPostDataResponse {
bucket_count: number
earliest_record_timestamp: number
empty_bucket_count: number
input_bytes: number
input_field_count: number
input_record_count: number
invalid_date_count: number
bucket_count: <<_long, long>>
earliest_record_timestamp: <<_long, long>>
empty_bucket_count: <<_long, long>>
input_bytes: <<_long, long>>
input_field_count: <<_long, long>>
input_record_count: <<_long, long>>
invalid_date_count: <<_long, long>>
job_id: <<Id>>
last_data_time: number
latest_record_timestamp: number
missing_field_count: number
out_of_order_timestamp_count: number
processed_field_count: number
processed_record_count: number
sparse_bucket_count: number
last_data_time: <<_integer, integer>>
latest_record_timestamp: <<_long, long>>
missing_field_count: <<_long, long>>
out_of_order_timestamp_count: <<_long, long>>
processed_field_count: <<_long, long>>
processed_record_count: <<_long, long>>
sparse_bucket_count: <<_long, long>>
}
[pass]

View File

@ -54,7 +54,7 @@ interface MlPutDataFrameAnalyticsRequest extends <<RequestBase>> {
analyzed_fields?: MlDataframeAnalysisAnalyzedFields | string[]
description?: string
dest: MlDataframeAnalyticsDestination
max_num_threads?: number
max_num_threads?: <<_integer, integer>>
model_memory_limit?: string
source: MlDataframeAnalyticsSource
headers?: <<HttpHeaders>>
@ -81,7 +81,7 @@ interface MlPutDataFrameAnalyticsResponse {
description?: string
dest: MlDataframeAnalyticsDestination
id: <<Id>>
max_num_threads: number
max_num_threads: <<_integer, integer>>
model_memory_limit: string
source: MlDataframeAnalyticsSource
version: <<VersionString>>

View File

@ -58,16 +58,16 @@ interface MlPutDatafeedRequest extends <<RequestBase>> {
delayed_data_check_config?: MlDelayedDataCheckConfig
frequency?: <<Duration>>
indices?: <<Indices>>
/** @alias indices */
pass:[/**] @alias indices */
indexes?: <<Indices>>
indices_options?: <<IndicesOptions>>
job_id?: <<Id>>
max_empty_searches?: number
max_empty_searches?: <<_integer, integer>>
query?: QueryDslQueryContainer
query_delay?: <<Duration>>
runtime_mappings?: MappingRuntimeFields
script_fields?: Record<string, <<ScriptField>>>
scroll_size?: number
scroll_size?: <<_integer, integer>>
headers?: <<HttpHeaders>>
}
@ -92,12 +92,12 @@ interface MlPutDatafeedResponse {
indices: string[]
job_id: <<Id>>
indices_options?: <<IndicesOptions>>
max_empty_searches?: number
max_empty_searches?: <<_integer, integer>>
query: QueryDslQueryContainer
query_delay: <<Duration>>
runtime_mappings?: MappingRuntimeFields
script_fields?: Record<string, <<ScriptField>>>
scroll_size: number
scroll_size: <<_integer, integer>>
}
[pass]

View File

@ -54,16 +54,16 @@ interface MlPutJobRequest extends <<RequestBase>> {
analysis_limits?: MlAnalysisLimits
background_persist_interval?: <<Duration>>
custom_settings?: MlCustomSettings
daily_model_snapshot_retention_after_days?: number
daily_model_snapshot_retention_after_days?: <<_long, long>>
data_description: MlDataDescription
datafeed_config?: MlDatafeedConfig
description?: string
groups?: string[]
model_plot_config?: MlModelPlotConfig
model_snapshot_retention_days?: number
renormalization_window_days?: number
model_snapshot_retention_days?: <<_long, long>>
renormalization_window_days?: <<_long, long>>
results_index_name?: <<IndexName>>
results_retention_days?: number
results_retention_days?: <<_long, long>>
}
[pass]
@ -84,7 +84,7 @@ interface MlPutJobResponse {
background_persist_interval?: <<Duration>>
create_time: <<DateTime>>
custom_settings?: MlCustomSettings
daily_model_snapshot_retention_after_days: number
daily_model_snapshot_retention_after_days: <<_long, long>>
data_description: MlDataDescription
datafeed_config?: MlDatafeed
description?: string
@ -94,10 +94,10 @@ interface MlPutJobResponse {
job_version: string
model_plot_config?: MlModelPlotConfig
model_snapshot_id?: <<Id>>
model_snapshot_retention_days: number
renormalization_window_days?: number
model_snapshot_retention_days: <<_long, long>>
renormalization_window_days?: <<_long, long>>
results_index_name: string
results_retention_days?: number
results_retention_days?: <<_long, long>>
}
[pass]

View File

@ -58,7 +58,7 @@ interface MlPutTrainedModelRequest extends <<RequestBase>> {
input?: MlPutTrainedModelInput
metadata?: any
model_type?: MlTrainedModelType
model_size_bytes?: number
model_size_bytes?: <<_long, long>>
platform_architecture?: string
tags?: string[]
prefix_strings?: MlTrainedModelPrefixStrings

View File

@ -49,10 +49,10 @@ Create part of a trained model definition.
++++
interface MlPutTrainedModelDefinitionPartRequest extends <<RequestBase>> {
model_id: <<Id>>
part: number
part: <<_integer, integer>>
definition: string
total_definition_length: number
total_parts: number
total_definition_length: <<_long, long>>
total_parts: <<_integer, integer>>
}
[pass]

View File

@ -51,7 +51,7 @@ interface MlPutTrainedModelVocabularyRequest extends <<RequestBase>> {
model_id: <<Id>>
vocabulary: string[]
merges?: string[]
scores?: number[]
scores?: <<_double, double>>[]
}
[pass]

View File

@ -51,10 +51,10 @@ interface MlStartTrainedModelDeploymentRequest extends <<RequestBase>> {
model_id: <<Id>>
cache_size?: <<ByteSize>>
deployment_id?: string
number_of_allocations?: number
number_of_allocations?: <<_integer, integer>>
priority?: MlTrainingPriority
queue_capacity?: number
threads_per_allocation?: number
queue_capacity?: <<_integer, integer>>
threads_per_allocation?: <<_integer, integer>>
timeout?: <<Duration>>
wait_for?: MlDeploymentAllocationState
}

View File

@ -51,7 +51,7 @@ interface MlUpdateDataFrameAnalyticsRequest extends <<RequestBase>> {
id: <<Id>>
description?: string
model_memory_limit?: string
max_num_threads?: number
max_num_threads?: <<_integer, integer>>
allow_lazy_start?: boolean
}
@ -71,11 +71,11 @@ interface MlUpdateDataFrameAnalyticsResponse {
allow_lazy_start: boolean
analysis: MlDataframeAnalysisContainer
analyzed_fields?: MlDataframeAnalysisAnalyzedFields | string[]
create_time: number
create_time: <<_long, long>>
description?: string
dest: MlDataframeAnalyticsDestination
id: <<Id>>
max_num_threads: number
max_num_threads: <<_integer, integer>>
model_memory_limit: string
source: MlDataframeAnalyticsSource
version: <<VersionString>>

View File

@ -58,16 +58,16 @@ interface MlUpdateDatafeedRequest extends <<RequestBase>> {
delayed_data_check_config?: MlDelayedDataCheckConfig
frequency?: <<Duration>>
indices?: string[]
/** @alias indices */
pass:[/**] @alias indices */
indexes?: string[]
indices_options?: <<IndicesOptions>>
job_id?: <<Id>>
max_empty_searches?: number
max_empty_searches?: <<_integer, integer>>
query?: QueryDslQueryContainer
query_delay?: <<Duration>>
runtime_mappings?: MappingRuntimeFields
script_fields?: Record<string, <<ScriptField>>>
scroll_size?: number
scroll_size?: <<_integer, integer>>
}
[pass]
@ -91,12 +91,12 @@ interface MlUpdateDatafeedResponse {
indices: string[]
indices_options?: <<IndicesOptions>>
job_id: <<Id>>
max_empty_searches?: number
max_empty_searches?: <<_integer, integer>>
query: QueryDslQueryContainer
query_delay: <<Duration>>
runtime_mappings?: MappingRuntimeFields
script_fields?: Record<string, <<ScriptField>>>
scroll_size: number
scroll_size: <<_integer, integer>>
}
[pass]

View File

@ -57,10 +57,10 @@ interface MlUpdateJobRequest extends <<RequestBase>> {
description?: string
model_plot_config?: MlModelPlotConfig
model_prune_window?: <<Duration>>
daily_model_snapshot_retention_after_days?: number
model_snapshot_retention_days?: number
renormalization_window_days?: number
results_retention_days?: number
daily_model_snapshot_retention_after_days?: <<_long, long>>
model_snapshot_retention_days?: <<_long, long>>
renormalization_window_days?: <<_long, long>>
results_retention_days?: <<_long, long>>
groups?: string[]
detectors?: MlDetector[]
per_partition_categorization?: MlPerPartitionCategorization
@ -85,7 +85,7 @@ interface MlUpdateJobResponse {
create_time: <<EpochTime>><<<UnitMillis>>>
finished_time?: <<EpochTime>><<<UnitMillis>>>
custom_settings?: Record<string, string>
daily_model_snapshot_retention_after_days: number
daily_model_snapshot_retention_after_days: <<_long, long>>
data_description: MlDataDescription
datafeed_config?: MlDatafeed
description?: string
@ -95,10 +95,10 @@ interface MlUpdateJobResponse {
job_version: <<VersionString>>
model_plot_config?: MlModelPlotConfig
model_snapshot_id?: <<Id>>
model_snapshot_retention_days: number
renormalization_window_days?: number
model_snapshot_retention_days: <<_long, long>>
renormalization_window_days?: <<_long, long>>
results_index_name: <<IndexName>>
results_retention_days?: number
results_retention_days?: <<_long, long>>
}
[pass]

View File

@ -49,7 +49,7 @@ Update a trained model deployment.
++++
interface MlUpdateTrainedModelDeploymentRequest extends <<RequestBase>> {
model_id: <<Id>>
number_of_allocations?: number
number_of_allocations?: <<_integer, integer>>
}
[pass]

View File

@ -70,7 +70,7 @@ interface MonitoringBulkResponse {
error?: <<ErrorCause>>
errors: boolean
ignored: boolean
took: number
took: <<_long, long>>
}
[pass]

View File

@ -55,9 +55,9 @@ interface MsearchRequest extends <<RequestBase>> {
ignore_throttled?: boolean
ignore_unavailable?: boolean
include_named_queries_score?: boolean
max_concurrent_searches?: number
max_concurrent_shard_requests?: number
pre_filter_shard_size?: number
max_concurrent_searches?: <<_long, long>>
max_concurrent_shard_requests?: <<_long, long>>
pre_filter_shard_size?: <<_long, long>>
rest_total_hits_as_int?: boolean
routing?: <<Routing>>
search_type?: <<SearchType>>

View File

@ -50,7 +50,7 @@ Run multiple templated searches.
interface MsearchTemplateRequest extends <<RequestBase>> {
index?: <<Indices>>
ccs_minimize_roundtrips?: boolean
max_concurrent_searches?: number
max_concurrent_searches?: <<_long, long>>
search_type?: <<SearchType>>
rest_total_hits_as_int?: boolean
typed_keys?: boolean

View File

@ -49,7 +49,7 @@ You can use this API to clear the archived repositories metering information in
++++
interface NodesClearRepositoriesMeteringArchiveRequest extends <<RequestBase>> {
node_id: <<NodeIds>>
max_archive_version: number
max_archive_version: <<_long, long>>
}
[pass]

View File

@ -51,9 +51,9 @@ interface NodesHotThreadsRequest extends <<RequestBase>> {
node_id?: <<NodeIds>>
ignore_idle_threads?: boolean
interval?: <<Duration>>
snapshots?: number
snapshots?: <<_long, long>>
master_timeout?: <<Duration>>
threads?: number
threads?: <<_long, long>>
timeout?: <<Duration>>
type?: <<ThreadType>>
sort?: <<ThreadType>>

View File

@ -48,8 +48,8 @@ Returns summarized information about existing query rulesets.
<pre>
++++
interface QueryRulesListRulesetsRequest extends <<RequestBase>> {
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
}
[pass]
@ -64,7 +64,7 @@ interface QueryRulesListRulesetsRequest extends <<RequestBase>> {
<pre>
++++
interface QueryRulesListRulesetsResponse {
count: number
count: <<_long, long>>
results: QueryRulesListRulesetsQueryRulesetListItem[]
}

View File

@ -53,7 +53,7 @@ interface QueryRulesPutRuleRequest extends <<RequestBase>> {
type: QueryRulesQueryRuleType
criteria: QueryRulesQueryRuleCriteria | QueryRulesQueryRuleCriteria[]
actions: QueryRulesQueryRuleActions
priority?: number
priority?: <<_integer, integer>>
}
[pass]

View File

@ -64,7 +64,7 @@ interface QueryRulesTestRequest extends <<RequestBase>> {
<pre>
++++
interface QueryRulesTestResponse {
total_matched_rules: number
total_matched_rules: <<_integer, integer>>
matched_rules: QueryRulesTestQueryRulesetMatchedRule[]
}

View File

@ -69,7 +69,7 @@ interface RankEvalRequest extends <<RequestBase>> {
<pre>
++++
interface RankEvalResponse {
metric_score: number
metric_score: <<_double, double>>
details: Record<<<Id>>, RankEvalRankEvalMetricDetail>
failures: Record<string, any>
}

View File

@ -49,7 +49,7 @@ Reindex documents. Copies documents from a source to a destination. The source c
++++
interface ReindexRequest extends <<RequestBase>> {
refresh?: boolean
requests_per_second?: number
requests_per_second?: <<_float, float>>
scroll?: <<Duration>>
slices?: <<Slices>>
timeout?: <<Duration>>
@ -58,9 +58,9 @@ interface ReindexRequest extends <<RequestBase>> {
require_alias?: boolean
conflicts?: <<Conflicts>>
dest: ReindexDestination
max_docs?: number
max_docs?: <<_long, long>>
script?: <<Script>> | string
size?: number
size?: <<_long, long>>
source: ReindexSource
}
@ -76,22 +76,22 @@ interface ReindexRequest extends <<RequestBase>> {
<pre>
++++
interface ReindexResponse {
batches?: number
created?: number
deleted?: number
batches?: <<_long, long>>
created?: <<_long, long>>
deleted?: <<_long, long>>
failures?: <<BulkIndexByScrollFailure>>[]
noops?: number
noops?: <<_long, long>>
retries?: <<Retries>>
requests_per_second?: number
slice_id?: number
requests_per_second?: <<_float, float>>
slice_id?: <<_integer, integer>>
task?: <<TaskId>>
throttled_millis?: <<EpochTime>><<<UnitMillis>>>
throttled_until_millis?: <<EpochTime>><<<UnitMillis>>>
timed_out?: boolean
took?: <<DurationValue>><<<UnitMillis>>>
total?: number
updated?: number
version_conflicts?: number
total?: <<_long, long>>
updated?: <<_long, long>>
version_conflicts?: <<_long, long>>
}
[pass]

View File

@ -49,7 +49,7 @@ Throttle a reindex operation. Change the number of requests per second for a par
++++
interface ReindexRethrottleRequest extends <<RequestBase>> {
task_id: <<Id>>
requests_per_second?: number
requests_per_second?: <<_float, float>>
}
[pass]

View File

@ -53,7 +53,7 @@ interface RollupPutJobRequest extends <<RequestBase>> {
groups: RollupGroupings
index_pattern: string
metrics?: RollupFieldMetric[]
page_size: number
page_size: <<_integer, integer>>
rollup_index: <<IndexName>>
timeout?: <<Duration>>
headers?: <<HttpHeaders>>

View File

@ -52,10 +52,10 @@ interface RollupRollupSearchRequest extends <<RequestBase>> {
rest_total_hits_as_int?: boolean
typed_keys?: boolean
aggregations?: Record<string, AggregationsAggregationContainer>
/** @alias aggregations */
pass:[/**] @alias aggregations */
aggs?: Record<string, AggregationsAggregationContainer>
query?: QueryDslQueryContainer
size?: number
size?: <<_integer, integer>>
}
[pass]
@ -70,7 +70,7 @@ interface RollupRollupSearchRequest extends <<RequestBase>> {
<pre>
++++
interface RollupRollupSearchResponse<TDocument = unknown, TAggregations = Record<<<AggregateName>>, AggregationsAggregate>> {
took: number
took: <<_long, long>>
timed_out: boolean
terminated_early?: boolean
_shards: <<ShardStatistics>>

View File

@ -53,7 +53,7 @@ interface SearchRequest extends <<RequestBase>> {
allow_partial_search_results?: boolean
analyzer?: string
analyze_wildcard?: boolean
batched_reduce_size?: number
batched_reduce_size?: <<_long, long>>
ccs_minimize_roundtrips?: boolean
default_operator?: QueryDslOperator
df?: string
@ -62,17 +62,17 @@ interface SearchRequest extends <<RequestBase>> {
ignore_unavailable?: boolean
include_named_queries_score?: boolean
lenient?: boolean
max_concurrent_shard_requests?: number
max_concurrent_shard_requests?: <<_long, long>>
min_compatible_shard_node?: <<VersionString>>
preference?: string
pre_filter_shard_size?: number
pre_filter_shard_size?: <<_long, long>>
request_cache?: boolean
routing?: <<Routing>>
scroll?: <<Duration>>
search_type?: <<SearchType>>
suggest_field?: <<Field>>
suggest_mode?: <<SuggestMode>>
suggest_size?: number
suggest_size?: <<_long, long>>
suggest_text?: string
typed_keys?: boolean
rest_total_hits_as_int?: boolean
@ -81,19 +81,19 @@ interface SearchRequest extends <<RequestBase>> {
q?: string
force_synthetic_source?: boolean
aggregations?: Record<string, AggregationsAggregationContainer>
/** @alias aggregations */
pass:[/**] @alias aggregations */
aggs?: Record<string, AggregationsAggregationContainer>
collapse?: SearchFieldCollapse
explain?: boolean
ext?: Record<string, any>
from?: number
from?: <<_integer, integer>>
highlight?: SearchHighlight
track_total_hits?: SearchTrackHits
indices_boost?: Record<<<IndexName>>, number>[]
indices_boost?: Record<<<IndexName>>, <<_double, double>>>[]
docvalue_fields?: (QueryDslFieldAndFormat | <<Field>>)[]
knn?: <<KnnSearch>> | <<KnnSearch>>[]
rank?: <<RankContainer>>
min_score?: number
min_score?: <<_double, double>>
post_filter?: QueryDslQueryContainer
profile?: boolean
query?: QueryDslQueryContainer
@ -101,13 +101,13 @@ interface SearchRequest extends <<RequestBase>> {
retriever?: <<RetrieverContainer>>
script_fields?: Record<string, <<ScriptField>>>
search_after?: <<SortResults>>
size?: number
size?: <<_integer, integer>>
slice?: <<SlicedScroll>>
sort?: <<Sort>>
_source?: SearchSourceConfig
fields?: (QueryDslFieldAndFormat | <<Field>>)[]
suggest?: SearchSuggester
terminate_after?: number
terminate_after?: <<_long, long>>
timeout?: string
track_scores?: boolean
version?: boolean

View File

@ -49,8 +49,8 @@ Returns the existing search applications.
++++
interface SearchApplicationListRequest extends <<RequestBase>> {
q?: string
from?: number
size?: number
from?: <<_integer, integer>>
size?: <<_integer, integer>>
}
[pass]
@ -65,7 +65,7 @@ interface SearchApplicationListRequest extends <<RequestBase>> {
<pre>
++++
interface SearchApplicationListResponse {
count: number
count: <<_long, long>>
results: SearchApplicationListSearchApplicationListItem[]
}

View File

@ -54,16 +54,16 @@ interface SearchMvtRequest extends <<RequestBase>> {
x: SearchMvtCoordinate
y: SearchMvtCoordinate
aggs?: Record<string, AggregationsAggregationContainer>
buffer?: number
buffer?: <<_integer, integer>>
exact_bounds?: boolean
extent?: number
extent?: <<_integer, integer>>
fields?: <<Fields>>
grid_agg?: SearchMvtGridAggregationType
grid_precision?: number
grid_precision?: <<_integer, integer>>
grid_type?: SearchMvtGridType
query?: QueryDslQueryContainer
runtime_mappings?: MappingRuntimeFields
size?: number
size?: <<_integer, integer>>
sort?: <<Sort>>
track_total_hits?: SearchTrackHits
with_labels?: boolean

View File

@ -79,15 +79,15 @@ interface SearchTemplateRequest extends <<RequestBase>> {
<pre>
++++
interface SearchTemplateResponse<TDocument = unknown> {
took: number
took: <<_long, long>>
timed_out: boolean
_shards: <<ShardStatistics>>
hits: SearchHitsMetadata<TDocument>
aggregations?: Record<<<AggregateName>>, AggregationsAggregate>
_clusters?: <<ClusterStatistics>>
fields?: Record<string, any>
max_score?: number
num_reduce_phases?: number
max_score?: <<_double, double>>
num_reduce_phases?: <<_long, long>>
profile?: SearchProfile
pit_id?: <<Id>>
_scroll_id?: <<ScrollId>>

View File

@ -68,7 +68,7 @@ interface SecurityCreateApiKeyRequest extends <<RequestBase>> {
++++
interface SecurityCreateApiKeyResponse {
api_key: string
expiration?: number
expiration?: <<_long, long>>
id: <<Id>>
name: <<Name>>
encoded: string

View File

@ -65,7 +65,7 @@ interface SecurityGetServiceCredentialsRequest extends <<RequestBase>> {
++++
interface SecurityGetServiceCredentialsResponse {
service_account: string
count: number
count: <<_integer, integer>>
tokens: Record<string, <<Metadata>>>
nodes_credentials: SecurityGetServiceCredentialsNodesCredentials
}

View File

@ -69,7 +69,7 @@ interface SecurityGetTokenRequest extends <<RequestBase>> {
++++
interface SecurityGetTokenResponse {
access_token: string
expires_in: number
expires_in: <<_long, long>>
scope?: string
type: string
refresh_token?: string

View File

@ -68,7 +68,7 @@ interface SecurityInvalidateApiKeyRequest extends <<RequestBase>> {
<pre>
++++
interface SecurityInvalidateApiKeyResponse {
error_count: number
error_count: <<_integer, integer>>
error_details?: <<ErrorCause>>[]
invalidated_api_keys: string[]
previously_invalidated_api_keys: string[]

View File

@ -66,10 +66,10 @@ interface SecurityInvalidateTokenRequest extends <<RequestBase>> {
<pre>
++++
interface SecurityInvalidateTokenResponse {
error_count: number
error_count: <<_long, long>>
error_details?: <<ErrorCause>>[]
invalidated_tokens: number
previously_invalidated_tokens: number
invalidated_tokens: <<_long, long>>
previously_invalidated_tokens: <<_long, long>>
}
[pass]

View File

@ -52,12 +52,12 @@ interface SecurityQueryApiKeysRequest extends <<RequestBase>> {
with_profile_uid?: boolean
typed_keys?: boolean
aggregations?: Record<string, SecurityQueryApiKeysApiKeyAggregationContainer>
/** @alias aggregations */
pass:[/**] @alias aggregations */
aggs?: Record<string, SecurityQueryApiKeysApiKeyAggregationContainer>
query?: SecurityQueryApiKeysApiKeyQueryContainer
from?: number
from?: <<_integer, integer>>
sort?: <<Sort>>
size?: number
size?: <<_integer, integer>>
search_after?: <<SortResults>>
}
@ -73,8 +73,8 @@ interface SecurityQueryApiKeysRequest extends <<RequestBase>> {
<pre>
++++
interface SecurityQueryApiKeysResponse {
total: number
count: number
total: <<_integer, integer>>
count: <<_integer, integer>>
api_keys: SecurityApiKey[]
aggregations?: Record<<<AggregateName>>, SecurityQueryApiKeysApiKeyAggregate>
}

View File

@ -49,9 +49,9 @@ Find roles with a query. Get roles in a paginated manner. You can optionally fil
++++
interface SecurityQueryRoleRequest extends <<RequestBase>> {
query?: SecurityQueryRoleRoleQueryContainer
from?: number
from?: <<_integer, integer>>
sort?: <<Sort>>
size?: number
size?: <<_integer, integer>>
search_after?: <<SortResults>>
}
@ -67,8 +67,8 @@ interface SecurityQueryRoleRequest extends <<RequestBase>> {
<pre>
++++
interface SecurityQueryRoleResponse {
total: number
count: number
total: <<_integer, integer>>
count: <<_integer, integer>>
roles: SecurityQueryRoleQueryRole[]
}

View File

@ -50,9 +50,9 @@ Find users with a query. Get information for users in a paginated manner. You ca
interface SecurityQueryUserRequest extends <<RequestBase>> {
with_profile_uid?: boolean
query?: SecurityQueryUserUserQueryContainer
from?: number
from?: <<_integer, integer>>
sort?: <<Sort>>
size?: number
size?: <<_integer, integer>>
search_after?: <<SortResults>>
}
@ -68,8 +68,8 @@ interface SecurityQueryUserRequest extends <<RequestBase>> {
<pre>
++++
interface SecurityQueryUserResponse {
total: number
count: number
total: <<_integer, integer>>
count: <<_integer, integer>>
users: SecurityQueryUserQueryUser[]
}

View File

@ -67,7 +67,7 @@ interface SecuritySamlAuthenticateRequest extends <<RequestBase>> {
interface SecuritySamlAuthenticateResponse {
access_token: string
username: string
expires_in: number
expires_in: <<_integer, integer>>
refresh_token: string
realm: string
}

View File

@ -65,7 +65,7 @@ interface SecuritySamlInvalidateRequest extends <<RequestBase>> {
<pre>
++++
interface SecuritySamlInvalidateResponse {
invalidated: number
invalidated: <<_integer, integer>>
realm: string
redirect: string
}

View File

@ -49,7 +49,7 @@ Suggest a user profile. Get suggestions for user profiles that match specified s
++++
interface SecuritySuggestUserProfilesRequest extends <<RequestBase>> {
name?: string
size?: number
size?: <<_long, long>>
data?: string | string[]
hint?: SecuritySuggestUserProfilesHint
}
@ -67,7 +67,7 @@ interface SecuritySuggestUserProfilesRequest extends <<RequestBase>> {
++++
interface SecuritySuggestUserProfilesResponse {
total: SecuritySuggestUserProfilesTotalUserProfiles
took: number
took: <<_long, long>>
profiles: SecurityUserProfile[]
}

View File

@ -50,7 +50,7 @@ Update user profile data. Update specific data for the user profile that is asso
interface SecurityUpdateUserProfileDataRequest extends <<RequestBase>> {
uid: SecurityUserProfileId
if_seq_no?: <<SequenceNumber>>
if_primary_term?: number
if_primary_term?: <<_long, long>>
refresh?: <<Refresh>>
labels?: Record<string, any>
data?: Record<string, any>

File diff suppressed because it is too large Load Diff

View File

@ -64,13 +64,13 @@ interface SlmGetStatsRequest extends <<RequestBase>> {
interface SlmGetStatsResponse {
retention_deletion_time: <<Duration>>
retention_deletion_time_millis: <<DurationValue>><<<UnitMillis>>>
retention_failed: number
retention_runs: number
retention_timed_out: number
total_snapshots_deleted: number
total_snapshot_deletion_failures: number
total_snapshots_failed: number
total_snapshots_taken: number
retention_failed: <<_long, long>>
retention_runs: <<_long, long>>
retention_timed_out: <<_long, long>>
total_snapshots_deleted: <<_long, long>>
total_snapshot_deletion_failures: <<_long, long>>
total_snapshots_failed: <<_long, long>>
total_snapshots_taken: <<_long, long>>
policy_stats: string[]
}

View File

@ -57,10 +57,10 @@ interface SnapshotGetRequest extends <<RequestBase>> {
index_names?: boolean
include_repository?: boolean
sort?: SnapshotSnapshotSort
size?: number
size?: <<_integer, integer>>
order?: <<SortOrder>>
after?: string
offset?: number
offset?: <<_integer, integer>>
from_sort_value?: string
slm_policy_filter?: <<Name>>
}
@ -79,8 +79,8 @@ interface SnapshotGetRequest extends <<RequestBase>> {
interface SnapshotGetResponse {
responses?: SnapshotGetSnapshotResponseItem[]
snapshots?: SnapshotSnapshotInfo[]
total: number
remaining: number
total: <<_integer, integer>>
remaining: <<_integer, integer>>
}
[pass]

View File

@ -68,7 +68,7 @@ interface SqlGetAsyncStatusResponse {
is_partial: boolean
start_time_in_millis: <<EpochTime>><<<UnitMillis>>>
expiration_time_in_millis: <<EpochTime>><<<UnitMillis>>>
completion_status?: number
completion_status?: <<_uint, uint>>
}
[pass]

View File

@ -52,7 +52,7 @@ interface SqlQueryRequest extends <<RequestBase>> {
catalog?: string
columnar?: boolean
cursor?: string
fetch_size?: number
fetch_size?: <<_integer, integer>>
filter?: QueryDslQueryContainer
query?: string
request_timeout?: <<Duration>>

View File

@ -48,7 +48,7 @@ Translates SQL into Elasticsearch queries
<pre>
++++
interface SqlTranslateRequest extends <<RequestBase>> {
fetch_size?: number
fetch_size?: <<_integer, integer>>
filter?: QueryDslQueryContainer
query: string
time_zone?: <<TimeZone>>
@ -67,7 +67,7 @@ interface SqlTranslateRequest extends <<RequestBase>> {
++++
interface SqlTranslateResponse {
aggregations?: Record<string, AggregationsAggregationContainer>
size?: number
size?: <<_long, long>>
_source?: SearchSourceConfig
fields?: (QueryDslFieldAndFormat | <<Field>>)[]
query?: QueryDslQueryContainer

Some files were not shown because too many files have changed in this diff Show More