Mongo ETL Target
In this article
Target Configuration
Connection Name: Connections are the service identifiers. A connection name can be selected from the list if you have created and saved connection details for MongoDB earlier. Or create one as explained in the topic - MongoDB Connection →
Database Name: The database name should be selected to write data.
Collection Name: The name of the database collection should be selected to write data.
Output Fields: The fields to be included in the output data should be selected from the drop-down list.
Extended BSON Types: When it is checked, this option enables the extended BSON types while writing the data to Mongo DB emitter.
Replace Document: Replace the whole document when saving datasets that contain an _id field. If unchecked, it will only update the fields in the document that match the fields in the dataset.
Local Threshold: Provide the threshold value (in milliseconds) for choosing a server from multiple Mongo DB servers.
Max Batch Size: The maximum batch size for bulk operations when saving data. The default value provided is 512.
Write Concern W: The w option requests for an acknowledgment that the write operation has propagated to a specified number of mongod instances or to mongod instances with specified tags.
Write Concern Journal: The j option requests acknowledgement from MongoDB that the write operation has been written to the on-disk journal.
Write Concern Timeout: This option specifies a time limit, in milliseconds, for the write concern. wtimeout is only applicable for w values greatet than 1.
Shard Key: MongoDB partitions data in the collection using ranges of shard key values. The field should be indexed and contain unique values.
Force Insert: Check the option to enable Force Insert to save inserts even if the datasets contain _IDs.
Ordered: This option can be enabled to allow setting the bulk operations ordered property.
Save Mode: Save Mode is used to specify the expected behavior of saving data to a data sink.
ErrorifExist: When persisting data, if the data already exists, an exception is expected to be thrown.
Append: When persisting data, if data/table already exists, contents of the Schema are expected to be appended to existing data.
Overwrite: When persisting data, if data/table already exists, existing data is expected to be overwritten by the contents of the source data.
Ignore: When persisting data, if data/table already exists, the save operation is expected to not save the contents of the source data and to not change the existing data.
This is similar to a CREATE TABLE IF NOT EXISTS in SQL.
Add Configuration: Additional properties can be added using this option as key-value pairs.
Post Action
To understand how to provide SQL queries or Stored Procedures that will be executed during pipeline run, see Post-Actions →
Notes
Optionally, enter notes in the Notes → tab and save the configuration.
If you have any feedback on Gathr documentation, please email us!