Return One Atlas Search Index
Returns one Atlas Search index in the specified project. You identify this index using its unique ID. Atlas Search index contains the indexed fields and the analyzers used to create the index. To use this resource, the requesting Service Account or API Key must have the Project Data Access Read Write role.
Path parameters
-
Unique 24-hexadecimal digit string that identifies your project. Use the /groups endpoint to retrieve all projects to which the authenticated user has access.
NOTE: Groups and projects are synonymous terms. Your group id is the same as your project id. For existing groups, your group/project id remains the same. The resource and corresponding endpoints use the term groups.
Format should match the following pattern:
^([a-f0-9]{24})$
. -
Name of the cluster that contains the collection with one or more Atlas Search indexes.
Format should match the following pattern:
^[a-zA-Z0-9][a-zA-Z0-9-]*$
. -
Unique 24-hexadecimal digit string that identifies the Application Search index. Use the Get All Application Search Indexes for a Collection API endpoint to find the IDs of all Application Search indexes.
Format should match the following pattern:
^([a-f0-9]{24})$
.
Query parameters
-
Flag that indicates whether Application wraps the response in an
envelope
JSON object. Some API clients cannot access the HTTP response headers or status code. To remediate this, set envelope=true in the query. Endpoints that return a list of results use the results object as an envelope. Application adds the status parameter to the response body.Default value is
false
. -
Flag that indicates whether the response body should be in the prettyprint format.
Default value is
false
.Prettyprint
Responses
-
OK
One of: Hide attributes Show attributes
-
Human-readable label that identifies the collection that contains one or more Atlas Search indexes.
-
Human-readable label that identifies the database that contains the collection with one or more Atlas Search indexes.
-
Human-readable label that identifies this index. Within each namespace, names of all indexes in the namespace must be unique.
-
Number of index partitions. Allowed values are [1, 2, 4].
Default value is
1
. -
Type of the index. Default type is search.
Value is
search
. -
Specific pre-defined method chosen to convert database field text into searchable words. This conversion reduces the text of fields into the smallest units of text. These units are called a term or token. This process, known as tokenization, involves a variety of changes made to the text in fields:
- extracting words
- removing punctuation
- removing accents
- changing to lowercase
- removing common words
- reducing words to their root form (stemming)
- changing words to their base form (lemmatization) MongoDB Cloud uses the selected process to build the Atlas Search index.
Values are
lucene.standard
,lucene.simple
,lucene.whitespace
,lucene.keyword
,lucene.arabic
,lucene.armenian
,lucene.basque
,lucene.bengali
,lucene.brazilian
,lucene.bulgarian
,lucene.catalan
,lucene.chinese
,lucene.cjk
,lucene.czech
,lucene.danish
,lucene.dutch
,lucene.english
,lucene.finnish
,lucene.french
,lucene.galician
,lucene.german
,lucene.greek
,lucene.hindi
,lucene.hungarian
,lucene.indonesian
,lucene.irish
,lucene.italian
,lucene.japanese
,lucene.korean
,lucene.kuromoji
,lucene.latvian
,lucene.lithuanian
,lucene.morfologik
,lucene.nori
,lucene.norwegian
,lucene.persian
,lucene.portuguese
,lucene.romanian
,lucene.russian
,lucene.smartcn
,lucene.sorani
,lucene.spanish
,lucene.swedish
,lucene.thai
,lucene.turkish
, orlucene.ukrainian
. Default value islucene.standard
.Atlas Search Analyzers -
List of user-defined methods to convert database field text into searchable words.
Settings that describe one Atlas Search custom analyzer.
Custom Atlas Search Analyzers Hide analyzers attributes Show analyzers attributes object
-
Human-readable name that identifies the custom analyzer. Names must be unique within an index, and must not start with any of the following strings:
lucene.
builtin.
mongodb.
-
Filters that examine text one character at a time and perform filtering operations.
One of: Filter that strips out HTML constructs.
Filter that processes normalized text with the ICU Normalizer. It is based on Lucene's ICUNormalizer2CharFilter.
Filter that applies normalization mappings that you specify to characters.
Hide attributes Show attributes
-
Human-readable label that identifies this character filter type.
Value is
mapping
. -
Comma-separated list of mappings. A mapping indicates that one character or group of characters should be substituted for another, using the following format:
<original> : <replacement>
.
-
tokenizer
object Required Tokenizer that you want to use to create tokens. Tokens determine how Atlas Search splits up text into discrete chunks for indexing.
One of: Tokenizer that splits input from the left side, or "edge", of a text input into n-grams of given sizes. You can't use the edgeGram tokenizer in synonym or autocomplete mapping definitions.
Hide attributes Show attributes
-
Human-readable label that identifies this tokenizer type.
Value is
edgeGram
. -
Characters to include in the shortest token that Atlas Search creates.
-
Characters to include in the longest token that Atlas Search creates.
Tokenizer that combines the entire input as a single token.
Tokenizer that splits input into text chunks, or "n-grams", of into given sizes. You can't use the nGram tokenizer in synonym or autocomplete mapping definitions.
Hide attributes Show attributes
Tokenizer that uses a regular expression pattern to extract tokens.
Hide attributes Show attributes
-
Human-readable label that identifies this tokenizer type.
Value is
regexCaptureGroup
. -
Regular expression to match against.
-
Index of the character group within the matching expression to extract into tokens. Use
0
to extract all character groups.
Tokenizer that splits tokens using a regular-expression based delimiter.
Tokenizer that splits tokens based on word break rules from the Unicode Text Segmentation algorithm.
Hide attributes Show attributes
Tokenizer that creates tokens from URLs and email addresses. Although this tokenizer uses word break rules from the Unicode Text Segmentation algorithm, we recommend using it only when the indexed field value includes URLs and email addresses. For fields that don't include URLs or email addresses, use the standard tokenizer to create tokens based on word break rules.
Hide attributes Show attributes
Tokenizer that creates tokens based on occurrences of whitespace between words.
Hide attributes Show attributes
-
-
Filter that performs operations such as:
Stemming, which reduces related words, such as "talking", "talked", and "talks" to their root word "talk".
Redaction, the removal of sensitive information from public documents.
Any of: tokenFilterasciiFoldingobject tokenFilterdaitchMokotoffSoundexobject tokenFilteredgeGramobject TokenFilterEnglishPossessiveobject TokenFilterFlattenGraphobject tokenFiltericuFoldingobject tokenFiltericuNormalizerobject TokenFilterkStemmingobject tokenFilterlengthobject tokenFilterlowercaseobject tokenFilternGramobject TokenFilterPorterStemmingobject tokenFilterregexobject tokenFilterreverseobject tokenFiltershingleobject tokenFiltersnowballStemmingobject TokenFilterSpanishPluralStemmingobject TokenFilterStempelobject tokenFilterstopwordobject tokenFiltertrimobject TokenFilterWordDelimiterGraphobject Filter that converts alphabetic, numeric, and symbolic Unicode characters that are not in the Basic Latin Unicode block to their ASCII equivalents, if available.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
asciiFolding
. -
Value that indicates whether to include or omit the original tokens in the output of the token filter.
Choose
include
if you want to support queries on both the original tokens as well as the converted forms.Choose
omit
if you want to query only on the converted forms of the original tokens.Values are
omit
orinclude
. Default value isomit
.
Filter that creates tokens for words that sound the same based on the Daitch-Mokotoff Soundex phonetic algorithm. This filter can generate multiple encodings for each input, where each encoded token is a 6 digit number.
NOTE: Don't use the daitchMokotoffSoundex token filter in:
-Synonym or autocomplete mapping definitions
- Operators where fuzzy is enabled. Atlas Search supports the fuzzy option only for the autocomplete, term, and text operators.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
daitchMokotoffSoundex
. -
Value that indicates whether to include or omit the original tokens in the output of the token filter.
Choose
include
if you want to support queries on both the original tokens as well as the converted forms.Choose
omit
if you want to query only on the converted forms of the original tokens.Values are
omit
orinclude
. Default value isinclude
.
Filter that tokenizes input from the left side, or "edge", of a text input into n-grams of configured sizes. You can't use this token filter in synonym or autocomplete mapping definitions.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
edgeGram
. -
Value that specifies the minimum length of generated n-grams. This value must be less than or equal to maxGram.
-
Value that specifies the maximum length of generated n-grams. This value must be greater than or equal to minGram.
-
Value that indicates whether to index tokens shorter than minGram or longer than maxGram.
Values are
omit
orinclude
. Default value isomit
.
Filter that removes possessives (trailing 's) from words.
Filter that transforms a token filter graph, such as the token filter graph that the wordDelimiterGraph token filter produces, into a flat form suitable for indexing.
Filter that applies character folding from Unicode Technical Report #30.
Filter that normalizes tokens using a standard Unicode Normalization Mode.
Filter that combines algorithmic stemming with a built-in dictionary for the English language to stem words.
Filter that removes tokens that are too short or too long.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
length
. -
Number that specifies the minimum length of a token. This value must be less than or equal to max.
Default value is
0
. -
Number that specifies the maximum length of a token. Value must be greater than or equal to min.
Default value is
255
.
Filter that normalizes token text to lowercase.
Filter that tokenizes input into n-grams of configured sizes. You can't use this token filter in synonym or autocomplete mapping definitions.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
nGram
. -
Value that specifies the minimum length of generated n-grams. This value must be less than or equal to maxGram.
-
Value that specifies the maximum length of generated n-grams. This value must be greater than or equal to minGram.
-
Value that indicates whether to index tokens shorter than minGram or longer than maxGram.
Values are
omit
orinclude
. Default value isomit
.
Filter that uses the porter stemming algorithm to remove the common morphological and inflectional suffixes from words in English. It expects lowercase text and doesn't work as expected for uppercase text.
Filter that applies a regular expression to each token, replacing matches with a specified string.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
regex
. -
Regular expression pattern to apply to each token.
-
Replacement string to substitute wherever a matching pattern occurs.
-
Value that indicates whether to replace only the first matching pattern or all matching patterns.
Values are
all
orfirst
.
Filter that reverses each string token.
Filter that constructs shingles (token n-grams) from a series of tokens. You can't use this token filter in synonym or autocomplete mapping definitions.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
shingle
. -
Value that specifies the minimum number of tokens per shingle. This value must be less than or equal to maxShingleSize.
-
Value that specifies the maximum number of tokens per shingle. This value must be greater than or equal to minShingleSize.
Filter that stems tokens using a Snowball-generated stemmer.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
snowballStemming
. -
Snowball-generated stemmer to use.
Values are
arabic
,armenian
,basque
,catalan
,danish
,dutch
,english
,finnish
,french
,german
,german2
,hungarian
,irish
,italian
,kp
,lithuanian
,lovins
,norwegian
,porter
,portuguese
,romanian
,russian
,spanish
,swedish
, orturkish
.
Filter that stems Spanish plural words. It expects lowercase text.
Filter that uses Lucene's default Polish stemmer table to stem words in the Polish language. It expects lowercase text.
Filter that removes tokens that correspond to the specified stop words. This token filter doesn't analyze the stop words that you specify.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
stopword
. -
The stop words that correspond to the tokens to remove. Value must be one or more stop words.
-
Flag that indicates whether to ignore the case of stop words when filtering the tokens to remove.
Default value is
true
.
Filter that trims leading and trailing whitespace from tokens.
Filter that splits tokens into sub-tokens based on configured rules.
Hide attributes Show attributes
-
Human-readable label that identifies this token filter type.
Value is
wordDelimiterGraph
. -
Object that contains the rules that determine how to split words into sub-words.
Hide delimiterOptions attributes Show delimiterOptions attributes object
-
Flag that indicates whether to split tokens based on sub-words.
Default value is
true
. -
Flag that indicates whether to split tokens based on sub-numbers. For example, if
true
, this option splits100-2
into100
and2
.Default value is
true
. -
Flag that indicates whether to concatenate runs of sub-words.
Default value is
false
. -
Flag that indicates whether to concatenate runs of sub-numbers.
Default value is
false
. -
Flag that indicates whether to concatenate runs.
Default value is
false
. -
Flag that indicates whether to generate tokens of the original words.
Default value is
true
. -
Flag that indicates whether to split tokens based on letter-case transitions.
Default value is
true
. -
Flag that indicates whether to split tokens based on letter-number transitions.
Default value is
true
. -
Flag that indicates whether to remove trailing possessives from each sub-word.
Default value is
true
. -
Flag that indicates whether to skip tokens with the
keyword
attribute set totrue
.Default value is
false
.
-
-
Object that contains options for protected words.
-
-
Index specifications for the collection's fields.
Hide mappings attributes Show mappings attributes object
-
Flag that indicates whether the index uses dynamic or static mappings. Required if mappings.fields is omitted.
Default value is
false
.Dynamic or Static Mappings -
One or more field specifications for the Atlas Search index. Required if mappings.dynamic is omitted or set to false.
Atlas Search Index Hide fields attribute Show fields attribute object
-
-
Method applied to identify words when searching this index.
Values are
lucene.standard
,lucene.simple
,lucene.whitespace
,lucene.keyword
,lucene.arabic
,lucene.armenian
,lucene.basque
,lucene.bengali
,lucene.brazilian
,lucene.bulgarian
,lucene.catalan
,lucene.chinese
,lucene.cjk
,lucene.czech
,lucene.danish
,lucene.dutch
,lucene.english
,lucene.finnish
,lucene.french
,lucene.galician
,lucene.german
,lucene.greek
,lucene.hindi
,lucene.hungarian
,lucene.indonesian
,lucene.irish
,lucene.italian
,lucene.japanese
,lucene.korean
,lucene.kuromoji
,lucene.latvian
,lucene.lithuanian
,lucene.morfologik
,lucene.nori
,lucene.norwegian
,lucene.persian
,lucene.portuguese
,lucene.romanian
,lucene.russian
,lucene.smartcn
,lucene.sorani
,lucene.spanish
,lucene.swedish
,lucene.thai
,lucene.turkish
, orlucene.ukrainian
. Default value islucene.standard
. -
Flag that indicates whether to store all fields (true) on Atlas Search. By default, Atlas doesn't store (false) the fields on Atlas Search. Alternatively, you can specify an object that only contains the list of fields to store (include) or not store (exclude) on Atlas Search. To learn more, see documentation.
Stored Source Fields -
Rule sets that map words to their synonyms in this index.
Synonyms used for this full text index.
Synonym Mapping Hide synonyms attributes Show synonyms attributes object
-
Specific pre-defined method chosen to apply to the synonyms to be searched.
Values are
lucene.standard
,lucene.simple
,lucene.whitespace
,lucene.keyword
,lucene.arabic
,lucene.armenian
,lucene.basque
,lucene.bengali
,lucene.brazilian
,lucene.bulgarian
,lucene.catalan
,lucene.chinese
,lucene.cjk
,lucene.czech
,lucene.danish
,lucene.dutch
,lucene.english
,lucene.finnish
,lucene.french
,lucene.galician
,lucene.german
,lucene.greek
,lucene.hindi
,lucene.hungarian
,lucene.indonesian
,lucene.irish
,lucene.italian
,lucene.japanese
,lucene.korean
,lucene.kuromoji
,lucene.latvian
,lucene.lithuanian
,lucene.morfologik
,lucene.nori
,lucene.norwegian
,lucene.persian
,lucene.portuguese
,lucene.romanian
,lucene.russian
,lucene.smartcn
,lucene.sorani
,lucene.spanish
,lucene.swedish
,lucene.thai
,lucene.turkish
, orlucene.ukrainian
. -
Human-readable label that identifies the synonym definition. Each synonym.name must be unique within the same index definition.
-
Data set that stores the mapping one or more words map to one or more synonyms of those words.
-
Hide attributes Show attributes
-
Human-readable label that identifies the collection that contains one or more Atlas Search indexes.
-
Human-readable label that identifies the database that contains the collection with one or more Atlas Search indexes.
-
Human-readable label that identifies this index. Within each namespace, names of all indexes in the namespace must be unique.
-
Number of index partitions. Allowed values are [1, 2, 4].
Default value is
1
. -
Type of the index. Default type is search.
Value is
vectorSearch
. -
Settings that configure the fields, one per object, to index. You must define at least one "vector" type field. You can optionally define "filter" type fields also.
Fields to index for vector search.
Vector Search Fields
-
-
Unauthorized.
Hide response attributes Show response attributes object
-
Bad request detail.
Hide badRequestDetail attribute Show badRequestDetail attribute object
-
Describes the specific conditions or reasons that cause each type of error.
-
HTTP status code returned with this error.
External documentation -
Application error code returned with this error.
-
Parameters used to give more information about the error.
-
Application error message returned with this error.
-
-
Forbidden.
Hide response attributes Show response attributes object
-
Bad request detail.
Hide badRequestDetail attribute Show badRequestDetail attribute object
-
Describes the specific conditions or reasons that cause each type of error.
-
HTTP status code returned with this error.
External documentation -
Application error code returned with this error.
-
Parameters used to give more information about the error.
-
Application error message returned with this error.
-
-
Not Found.
Hide response attributes Show response attributes object
-
Bad request detail.
Hide badRequestDetail attribute Show badRequestDetail attribute object
-
Describes the specific conditions or reasons that cause each type of error.
-
HTTP status code returned with this error.
External documentation -
Application error code returned with this error.
-
Parameters used to give more information about the error.
-
Application error message returned with this error.
-
-
Internal Server Error.
Hide response attributes Show response attributes object
-
Bad request detail.
Hide badRequestDetail attribute Show badRequestDetail attribute object
-
Describes the specific conditions or reasons that cause each type of error.
-
HTTP status code returned with this error.
External documentation -
Application error code returned with this error.
-
Parameters used to give more information about the error.
-
Application error message returned with this error.
-
curl \ --request GET 'https://cloud.mongodb.com/api/atlas/v1.0/groups/32b6e34b3d91647abb20e7b8/clusters/{clusterName}/fts/indexes/{indexId}' \ --header "Authorization: Bearer $ACCESS_TOKEN"
{ "collectionName": "string", "database": "string", "name": "string", "numPartitions": 1, "type": "search", "analyzer": "lucene.standard", "analyzers": [ { "name": "string", "charFilters": [ { "type": "htmlStrip", "ignoredTags": [ "string" ] } ], "tokenizer": { "type": "edgeGram", "minGram": 42, "maxGram": 42 }, "tokenFilters": [ { "type": "asciiFolding", "originalTokens": "omit" } ] } ], "mappings": { "dynamic": false, "fields": { "additionalProperty1": {}, "additionalProperty2": {} } }, "searchAnalyzer": "lucene.standard", "storedSource": { "include | exclude": [ "field1", "field2" ] }, "synonyms": [ { "analyzer": "lucene.standard", "name": "string", "source": { "collection": "string" } } ] }
{ "collectionName": "string", "database": "string", "name": "string", "numPartitions": 1, "type": "vectorSearch", "fields": [ { "additionalProperty1": {}, "additionalProperty2": {} } ] }
{ "error": 401, "detail": "(This is just an example, the exception may not be related to this endpoint)", "reason": "Unauthorized", "errorCode": "NOT_ORG_GROUP_CREATOR" }
{ "error": 403, "detail": "(This is just an example, the exception may not be related to this endpoint)", "reason": "Forbidden", "errorCode": "CANNOT_CHANGE_GROUP_NAME" }
{ "error": 404, "detail": "(This is just an example, the exception may not be related to this endpoint) Cannot find resource AWS", "reason": "Not Found", "errorCode": "RESOURCE_NOT_FOUND" }
{ "error": 500, "detail": "(This is just an example, the exception may not be related to this endpoint)", "reason": "Internal Server Error", "errorCode": "UNEXPECTED_ERROR" }