allowDiskUse lets the server know if it can use disk to store temporary results for the aggregation (requires mongodb 2.6 >).
If awaitData is set to true, when the cursor reaches the end of the capped collection, MongoDB blocks the query thread for a period of time waiting for new data to arrive. When new data is inserted into the capped collection, the blocked thread is signaled to wake up and return the next batch to the client.
Specifies the number of documents to return in each response from MongoDB
return BSON regular expressions as BSONRegExp instances.
Allow driver to bypass schema validation in MongoDB 3.2 or higher.
the serializer will check if keys are valid.
Comment to apply to the operation.
In server versions pre-4.4, 'comment' must be string. A server error will be thrown if any other type is provided.
In server versions 4.4 and above, 'comment' can be any valid BSON type.
Return the query as cursor, on 2.6 > it returns as a real cursor on pre 2.6 it returns as an emulated cursor.
Enable utf8 validation when deserializing BSON documents. Defaults to true.
Specifies the verbosity mode for the explain output.
allow to specify if there what fields we wish to return as unserialized raw buffer.
Add an index selection hint to an aggregation command
serialize will not emit undefined fields (default:true)
Map of parameter names and values that can be accessed using $$var (requires MongoDB 5.0).
maxAwaitTimeMS controls the amount of time subsequent getMores
that a cursor uses to fetch more data should take. (ex. cursor.next())
maxTimeMS controls the amount of time the initial command
that constructs a cursor should take. (ex. find, aggregate, listCollections)
when deserializing a Binary will return it as a node.js Buffer instance.
when deserializing a Long will fit it into a Number if it's smaller than 53 bits
when deserializing will promote BSON values to their Node.js closest equivalent types.
Enabling the raw option will return a Node.js Buffer which is allocated using allocUnsafe API. See this section from the Node.js Docs here for more detail about what "unsafe" refers to in this context. If you need to maintain your own editable clone of the bytes returned for an extended life time of the process, it is recommended you allocate your own buffer and clone the contents:
Specify a read concern and level for the collection. (only MongoDB 3.2 or higher supported)
The preferred read preference (ReadPreference.primary, ReadPreference.primary_preferred, ReadPreference.secondary, ReadPreference.secondary_preferred, ReadPreference.nearest).
Should retry failed writes
Specify ClientSession for this command
By default, MongoDB will automatically close a cursor when the client has exhausted all results in the cursor. However, for capped collections you may use a Tailable Cursor that remains open after the client exhausts the results in the initial cursor.
Write Concern as an object
Generated using TypeDoc