Knowledge base

Importing Your Historical Data

When you first [connect your cloud software](/kb/gettingyourdata?api=) to SyncHub, the first thing it will do is automatically begin importing your historical data. ### Changing how far back SyncHub will go On the [connection dashboard](/kb/connectiondashboardsexplained?api=), you can see the currently selected date for each table below the progress bar on the left-hand side. To change this, follow these steps: 1. Click the table name. 2. The text box will say “The next update is due at … and will check the [run size] from [date].” Click [date]. 3. Change the Sync From field to whatever date you would like SyncHub to start pulling data from. 4. Click Update, then Save. 5. Assuming you set the date earlier than what was previously displayed below the progress bar, that date will have now changed. Note: Whenever you change the Sync From field, data already pulled from after the new date you set will be re-synced. So to avoid needlessly re-syncing data it’s a good idea to decide the date as soon as you've connected. ### Importing your historical data quicker NB: This also applies to any [Segments](/kb/segments?api=) you create. Importing your historical data will take some time. Exactly how long it will take depends on how much historical data you are pulling in. Regardless though, you can speed it up by changing the [run size](/kb/howsynchubworks?api=). From the [connections page](/kb/quickstartguide?api=), head to the dashboard of the connection you have just added. For each table shown, follow these steps: 1. Click the table name. 2. The text box will say “The next update is due at … and will check the [run size] from [date].” Click [run size]. 3. Increase the Run Size field. We recommend making each run about 2-4 weeks – read below to learn why. 4. Click Update, then Save. This means SyncHub will now pull larger chunks of data each time it completes a run. But don't get carried away: making the run size too big may cause the cloud software’s API to timeout mid-run, forcing SyncHub to start that run again and therefore slowing down your sync. Also, while SyncHub is pulling your historical data it will automatically start a new run every minute – provided the previous run is complete. This means there is no need to make the run size very large since a new run will begin within one minute anyway.