Description of Scheduled Task Capabilities

  • Last update: December 06, 2024
  • Overview

    The Scheduled Task module in Data Development supports data extraction, transformation, and loading using various nodes and operators on a visual page. It enables the automatic execution of scheduled tasks through the Timed Scheduling function, helping you build offline data warehouses effortlessly and ensuring efficient and stable data production.

    This article describes the capabilities of the Scheduled Task module.

    Efficient Integration and Processing of Multi-source Heterogeneous Data

    Flexible Selection of Source and Target Terminals 

    Function Description

    For details, see the "Types of Data Sources Supported by Scheduled Task" section of Data Sources Supported by FineDataLink.

    The data sources in the first and second columns in the following table can be combined flexibly. For example, you can read file data and write the data into a database after processing (optional).

    Supported Source End Data SourceSupported Target End Data SourceRemark
    DatabaseDatabase/
    API (API and web service)

    API (API)

    Writing data into web services is not supported currently.

    /
    File (such as CSV, Excel, and TXT files)File (such as CSV, Excel, and TXT files)/
    FineBI Public DataFineBI Public DataFor details, see Dataset Output.
    Server datasetServer datasets are not supported./
    Big data platform (such as SAP HANA, Impala, and Greenplum)Big data platform (such as SAP HANA, Greenplum, and ClickHouse)/
    Connector (Jodoo and SAP RFC)

    Connector (Jodoo)

    Writing data into SAP RFC is not supported currently.

    /

    Others

    TypeDescription
    Read data from multiple tables.

    1. The Union All and Data Association operators can be used to merge two different tables.

    2. The SQL Script node allows for processing multi-table data by writing SQL statements.

    Output data to multiple tables.

    The Data Distribution function can be used to save the processed data to multiple tables.

     

    Create, read data from, and write data into a partition table.Many databases support partition tables to enhance query performance in scenarios with large data volumes. You may want to create, read data from, and write data into partition tables using FineDataLink. For details, see Partition Table Creation and Data Reading/Writing
    Call a database stored procedure.You can call the database stored procedure in the scheduled task. For details, see Scheduled Task Calling Database Stored Procedures
    Call the API related to scheduled tasks.
    • You can execute the scheduled task by calling the API.

    • You can view the status of the scheduled task instance by calling the API.

    • You can terminate the running of the scheduled task by calling the API.

    For details, see Calling Scheduled Task Related APIs

    Write data into the specified directory of Public Data in the remote FineBI project.For details, see Dataset Output.

    Flexible Conversion of Data Structures 

    Scheduled Task enables you to convert data to structured data, semi-structured data, and unstructured data flexibly.

    Example:

    Data Cleaning

    Clean the raw data to output standardized data.

    TypeDescription



    Data Cleaning

    For details about the nodes and operators used to process data, see the "Node Introduction" section of Overview of Scheduled Task.
    For details about the syntax supported by the Spark SQL operator, see Spark SQL Syntax Overview
    For details about data processing functions, see Text Function Overview, Date Function Overview, Logic Function Overview, Numeric Function, and ISNULL - Determining Whether an Object is Null.

    Data Update

    The data update methods include incremental update, full update, and comparison-based update. For details, see Overview of Data Synchronization Method.

    iconNote:
    You can use the Data Comparison operator to check the data consistency by comparing source and target data.

    Data Integration

    Integrating various types of heterogeneous data is supported to achieve data association and dimensional data modeling. For details, see Union All, Data Association, SQL Script, and Spark SQL.

    Data Alert

    Scheduled Task provides instant alerts for key business data when limits are reached. For details, see Notification.

    Notification of scheduled task execution failure is also supported. For details, see Task Control - Result Notification.

    Data Sharing

    iconNote:
    You can use Data Service to output data as APIs for other systems to call.

    Data can also be output as files for storage. For details, see Function Description of File Output Operator and File Transfer

    Others

    1. You can be prompted for DDL changes in the source table of a scheduled task and notify specified users. For details, see Synchronizing DDL Changes Using Scheduled Task.

    2. Data Lineage helps you understand the entire data production process and the impact scope of modifications to data tables. For details, see Lineage Analysis.

    3. The front-end Database Table Management function enables visual management of multi-source data and helps understand data conditions quickly by searching for specific database tables.

    Separated Development and Production

    Task development and editing are separated from task execution. You can edit and modify scheduled tasks in a development environment, where you can also perform task commissioning, and then publish them to a stable production environment, ensuring that task development and execution do not interfere with each other.

    For details, see Development Mode and Production Mode

    Canvas-Based Task Development 

    The canvas-like development page enables efficient task development and easy O&M.

    For details about differences between Data Synchronization and Data Transformation, see Differences Between Data Synchronization and Data Transformation.

    TypeDescription
    Step flow
    • You can define the execution order of data flow nodes and scripts (such as serial and parallel execution) by connecting nodes.

    • Logic nodes for task orchestration, such as Conditional Branch and Loop Container, are provided.

    Examples of the step flow are shown in the following figure.

    Data flow
    • A data flow focuses on data processing, such as data synchronization and transformation and file transfer.

    • Various visual operators are provided for fast data input, transformation, and output.

    The data flow can flexibly meet all data cleaning and transformation needs.

    Single Scheduled Task Management

    Combining Various Scheduling Strategies Flexibly

    FineDataLink provides three scheduling strategies, namely time-based, event dependency-based, and trigger-based scheduling, which can be flexibly combined to meet diverse scheduling needs.

    For details, see Overview of Scheduling Plan.

    Setting Priority for a Single Scheduled Task

    You can set task priority based on the business data priority and set the log level to ensure focused support for critical businesses.

    For details, see Task Control - Task Attribute.

    Ensuring Normal Task Execution with Fault Tolerance Mechanisms

    FineDataLink offers fault tolerance mechanisms, including Timeout LimitRetry After Failure, and Dirty Data Tolerance. When encountering controllable risks such as network fluctuations or a small amount of dirty data, the platform automatically retries the task to ensure smooth task operation.

    For details, see Task Control - Fault Tolerance Mechanism.

    Notification of Task Execution Failure

    With the result notification enabled, the platform will notify the relevant personnel in case of task execution failure.

    • The notification channel includes platform messages, emails, SMSs, DingTalk chatbot, Lark chatbot, and WeCom chatbot.

    • The notification content includes task exceptions, dirty data, and notification of source table structure changes.

    For details, see Task Control - Result Notification.

    Recycle Bin

    You can restore and hard delete the deleted tasks in Scheduled TaskData Pipeline, and Data Service in Recycle Bin.

    For details, see Recycle Bin.

    Batch Scheduled Task O&M 

    1. You can monitor and manage scheduled tasks in real time under O&M Center > Scheduled Task > Running Record, where you can view task running information such as the running status, duration, and triggering method, and retry scheduled tasks (for supplementing data or re-executing tasks after solving the dirty data that causes data writing failure).

    For details, see Running Record.

    2. You can set the execution frequency and the event-based scheduling plan for multiple/single scheduled tasks and modify and delete the execution frequency and the event-based scheduling plan of the task under O&M Center > Scheduled Task > Scheduling Plan.

    For details, see Scheduled Task O&M - Scheduling Plan

    3. You can view task information such as the number of scheduled tasks, task scheduling status, and scheduling plan execution status, and configure result notification for tasks in batches under O&M Center > Scheduled Task > Task Management.

    For details, see Scheduled Task O&M - Task Management.

    4. You can manage the memory resources used by scheduled tasks for an independently deployed FineDataLink project.

    For details, see Load Distribution.

    Enterprise-Level Permission Control

    Permission related to the Scheduled Task module is explained in the following table.

    DocumentDescription
    Overview of Data Connection Permission

    After being granted Use permission on a data connection, the user can use the data connection to create scheduled tasks, pipeline tasks, and API tasks after login.

    After being granted Management permission on a data connection, the user can copy, rename, modify, and delete the data connection under System Management > Data Connection > Data Connection Management after login.

    After being granted Authorization permission on a data connection, the user (subordinate admin) can assign corresponding permission on the data connection to manageable users under System Management > Permission Management after login.

    Data Platform Use Permission

    In FineData Link, you (the admin) want to assign Use permission on Scheduled Task in Data DevelopmentData PipelineData Service, and Database Table Management to other users for them to:

    • Use Scheduled Task and Data Pipeline for data processing.

    • Use Data Service to publish APIs.

    • Use Database Table Management to write/debug SQL statements, view created tables, as well as delete, clear, and copy and paste tables in FineDataLink.

    Data Platform Authorization Permission

    You (the super admin) can assign Authorization permission on Data Platform to a user to make the user a subordinate admin.

    The subordinate admin can assign Use permission on modules on the data platform to subordinate members.

    Scheduled Task Management PermissionYou (the admin) can assign users permission to view and edit specific folders and scheduled tasks in Scheduled Task.
    Scheduled Task Authorization Permission

    You (the super admin) can assign Authorization permission on Scheduled Task to a user to make the user a subordinate admin.

    The subordinate admin can assign Management permission on scheduled tasks to subordinate members to allow them to view and edit the scheduled task.



    附件列表


    主题: Data Development - Scheduled Task
    Previous
    Next
    • Helpful
    • Not helpful
    • Only read

    滑鼠選中內容,快速回饋問題

    滑鼠選中存在疑惑的內容,即可快速回饋問題,我們將會跟進處理。

    不再提示

    10s後關閉

    Get
    Help
    Online Support
    Professional technical support is provided to quickly help you solve problems.
    Online support is available from 9:00-12:00 and 13:30-17:30 on weekdays.
    Page Feedback
    You can provide suggestions and feedback for the current web page.
    Pre-Sales Consultation
    Business Consultation
    Business: international@fanruan.com
    Support: support@fanruan.com
    Page Feedback
    *Problem Type
    Cannot be empty
    Problem Description
    0/1000
    Cannot be empty

    Submitted successfully

    Network busy