Salesforce Data Source
In this article
Parse Salesforce data from the source or import a file in either data type format except Parquet. Salesforce channel allows to read data from a Salesforce account.
Salesforce is a top-notch CRM application built on the Force.com platform. It can manage all the customer interactions of an organization through different media, like phone calls, site email inquiries, communities, as well as social media. This is done by reading Salesforce object specified by Salesforce Object Query Language.
First is to create a Salesforce connection and for that you would require the following prerequisites:
A valid Salesforce account.
User name of Salesforce account.
Password of Salesforce account.
Security token of Salesforce account.
Configuring Salesforce Data Source
To add a Salesforce Data Source into your pipeline, select the Data Source and click on it to configure.
Under the Schema Type tab, select Fetch From Source or Upload Data File.
Now, select a Salesforce connection and write a query to fetch any Salesforce object. Then provide an API version. Below are the configuration details:
Field | Description |
---|---|
Connection Name | Select the Salesforce connection. |
Override Credentials | Unchecked by default, check the override credentials option for user specific actions. |
Username | Once the Override Credentials option is checked, provide the user name through which the Salesforce service is running. |
Password | Provide the password for Salesforce account. |
Security Token | Provide the security token for Salesforce account. |
Table Name | Select the source table name for detecting schema. |
Query | Use the Salesforce Object Query Language (SOQL) to search your organization’s Salesforce data for specific information. SOQL is similar to the SELECT statement in the widely used Structured Query Language (SQL) but is designed specifically for Salesforce data. |
Infer Schema | Inferschema from the query results to find the data type of the field specified in SOQL.This will work if number of record is 5 or greater. (Optional) |
Bulk | Flag to enable bulk query. This is the preferred method when loading large sets of data. Bulk API is based on REST principles and is optimized for loading large sets of data. You can use it to query many records asynchronously by submitting batches. Salesforce will process batches in the background. The Default value is false. (Optional) |
Pk Chunking | Flag to enable automatic primary key chunking for bulk query job. (Optional) This splits bulk queries into separate batches of the size defined by chunkSize option. By default, false and the default chunk size is 100,000. PK Chunking feature can automatically make large queries manageable when using the Bulk API. PK stands for Primary Key — the object’s record ID — which is always indexed. This feature is supported for all custom objects, many standard objects, and their sharing tables. |
ChunkSize | The size of the number of records to include in each batch. Default value is 100,000. This option can only be used when Pk Chunking is true. Maximum size is 250,000. |
Timeout | The maximum time in milliseconds spent polling for the completion of bulk query job. This option can only be used when bulk is true. |
Version | Select Salesforce version from the drop-down list. |
Configure Pre-Action in Source →
If you have any feedback on Gathr documentation, please email us!