Skip to main content
All CollectionsGetting Started
Lesson 6: Schedule regular runs

Lesson 6: Schedule regular runs

Updated this week

By now, you've finished building your first scraping task and know how to run the task to get your target data. Let's take it to the next level and find out how you can make your daily scraping routines more effective and efficient with the below features:


1. Task scheduling

If you are planning on getting data extracted regularly, task scheduling is exactly what you need and can save you a lot of time. You can schedule your task to run once, on a recurring schedule, or even run repeatedly, such as every 1 min, 5 mins, 10 mins, or 30 mins.

STEP 1.

Find your task on the Dashboard, and click Not Set under the Next Run column. The first one is for Cloud schedules and the second one is for local schedules.

STEP 2.

Choose how often you would like to run the task based on your needs.

STEP 3.

You can also save the settings for later use. Give the settings a name and click Save. This way, you can always select the saved schedule setting and apply it directly to any other task.

STEP 4.

After everything's done, click Schedule ON to start running the task on the schedule right away.

STEP 5.

When a task is scheduled, you'll see the next run time on the Task List.

You can easily turn it ON or OFF by clicking the next run time in the Task List and selecting either Schedule ON or Schedule OFF.


2. Auto-data export (for Cloud data)

Data export to the database can also be automated and scheduled. If you need to export data to your databases regularly, data export scheduling can save you tons of work.

STEP 1.

View the Cloud data associated with the task.

STEP 2.

Click on Export Data

STEP 3.

Find Export to database and click Set, then select the type of database you have.

STEP 4.

Complete the information to connect to your database. Click Test Connection to test if the database is connected successfully. Then, click Next to proceed to map the data fields, and choose the desired time interval for the export.

STEP 5.

Lastly, click Confirm to finish the process.

STEP 6.

You can find your auto-export tasks in the Database Auto-export Tool

tool.jpg

3. Connect via API

With the Octoparse API, you can run scraping tasks, retrieve extracted data, and even edit your tasks programmatically by coordinating with your application.

Check the API documentation for details.


4. Connect via Zapier

You can connect the data scraped with other applications (e.g., Google Drive, Google Sheet, Dropbox) easily without coding by using Zapier. Check this to find out more details: How to Connect Octoparse with Zapier


Did this answer your question?