You can use a dataflow to create one or more datasets based on data from Salesforce objects or existing datasets.
Design the Dataflow Before you start creating the dataflow definition file in the .json format, think about the dataflow design. Consider what data to make available for queries, where to extract the data from, and whether you need to transform the extracted data to get the data you want.
Configure the Dataflow Configure the dataflow based on your dataflow design. You can configure the dataflow to extract data, transform datasets based on your business requirements, and register datasets that you want to make available for queries. To configure the dataflow, add transformations to the dataflow definition file.
Start and Stop a Dataflow You can manually start a dataflow job to load the data into datasets immediately. You can also stop the job while it’s running. You can run a maximum of 24 dataflow jobs during a rolling 24-hour period.
Monitor a Dataflow Job Use the data monitor to monitor dataflow jobs to ensure that they complete successfully or to troubleshoot them if they fail.
Reschedule and Unschedule a Dataflow After you run a dataflow job for the first time, it runs on a daily schedule, by default. You can change the time of the daily schedule or schedule it to run on an hourly interval. You might change the schedule to ensure that the data is available by a particular time or to run the job during non-business hours. You can also unschedule a dataflow.