BACKGROUND
The scheduler is available on the Desktop and Power Desktop as an optional add-on, and on the server version by default.
If you’re using the Desktop, you need to be logged in as the Analyze user
If you don't see what you need here, check out our other self-service training material:
How to create and use Run Properties
LEARNING OUTCOMES
This article and videos aim to arm you with the skills needed to run dataflow through a schedule, allowing you to increase productivity and batch throughput. Things you'll learn:
- some Analyze scheduling terminology
- how to create and manage a schedule
- the best way to view scheduled runs, how to create chaining and how to send email notifications upon dataflow failure
- how to purge old unwanted temp files for the purposes of keeping a healthy system
Software version: 3.6.3
HOW TO CREATE A SCHEDULE
Before we get started creating a schedule, here are some scheduling terms that my be useful going forward -
Schedule: a customizable set of rules that you create to govern the timing and conditions under which the dataflow is run
Run: a single execution of a dataflow, which can be kicked off automatically via a schedule or manually in the Designer
HOW TO SETUP AN EMAIL NOTIFICATION ON FAILURE
The system has a send email node available by default that will notify you and other users in the event of a failed schedule. This video demonstrates how to configure that notification.
HOW TO MANAGE AND VIEW SCHEDULES
When viewing schedules you will see several different symbols to indicate their status. Here is an explanation of each:
This video takes you through what you can expect to find under the Schedules tab and the Runs tab, as well as the various ways to edit a schedule.
HOW TO PURGE OLD RUNS
Before purging old unwanted data it's important to understand what is being cleaned up -
Temp data - the data that is available at the output pins of nodes within the data flow - i.e. the source for the information that would be displayed in the Data Viewer when you click on an output pin. The temporary data is stored in files and you can choose to delete these files. When the temporary data file has been deleted, you can no longer view the data in the Data Viewer. If you just delete the temporary data you can still see the pass/fail status of each node in the graph and the record counts.
Run State - includes the success/fail status of each node, the record count at output pins and the associated temporary data. Run State is a super-set of the information stored for the run and includes the 'temporary data'. If you choose to delete the run state then all information is deleted (i.e. the data flow would be displayed as if it had never been run)
Scheduled Run - the execution of a dataflow that was initiated using a schedule
Interactive runs - when you have opened the data flow (e.g.for editing) and run some or all of the nodes in the data flow. This does not apply to runs created by schedules, nor the Run Now feature on a schedule.
Note: cleanup is done daily at midnight for interactive and scheduled runs, and the audit log should show two separate entries for them both.
By purging old unwanted temp data, you'll see:
- faster processing times
- reduced disk space
We'd love to hear your thoughts on this article and it's contents, or if you have a use case that you'd like to see demonstrated in a video / article, please let us know by posting in the comments below, or email me directly at ccostello@infogix.com
Comments
0 comments
Please sign in to leave a comment.