Data Source - TROCCO
    • PDF

    Data Source - TROCCO

    • PDF

    Article summary

    summary

    Help page for ETL Configuration of Data Settings held in TROCCO.
    Data Source TROCCO can transfer any of the following data

    • Data Catalog
      • Metadata held in the Data Catalog can be transferred.
    • ETL Configuration, Data Mart Configuration, Workflow Definition
      • Historical data on previously executed ETL Jobs can be ETLed.

    constraints

    About Data Catalog

    In order to transfer data in the Data Catalog, the Data Catalog must be activated.

    Data Catalog is a paid option.
    To request a trial or to sign up, please contact your sales representative or Customer Success.

    Period during which job execution history can be obtained

    ETL Job can transfer historical data of jobs executed up to one year in the past, starting from the date and time when the job execution history is transferred.
    Historical data for jobs executed before that time cannot be ETL Job.

    Setting items

    STEP1 Basic settings

    item nameindispensableContents
    subject (of taxation, etc.)YesSelect the target to be transferred from the following
  • Data Catalog
  • ETL Configuration
  • Data Mart Configuration
  • Workflow
  • data typeYesSelect the data type to be transferred from the following
    • If Data Catalog is selected
      • User-defined metadata (BigQuery table information)
      • User-defined metadata (BigQuery Column Setting information)
      • column reference list
    • When ETL Configuration, Data Mart Configuration, or Workflow Configuration is selected
      • job execution history

    When ETL Configuration, Data Mart Configuration, or Workflow Configuration is selected

    item nameindispensableContents
    Data acquisition periodYesEnter the start and end dates and times of the data you wish to transfer.
    Please enter in YYYYY-MM-DD or YYYYY-MM-DD HH:mm:ss format.
    If HH:mm:ss is not specified, 00:00:00 is specified instead.
    time zoneYesSelect the time zone for the data acquisition period.
    Criteria for data acquisition period

    The data acquisition period is based on the execution date/time ( created_at) of each job.
    Therefore, the historical data of ETL Job whose created_at value is within the data acquisition period is subject to ETL Job.

    transfer item

    Below is a list of items transferred for each target and each data type.

    Data Catalog

    user defined metadata

    There are two types of data types that can be transferred

    • User-defined metadata (BigQuery table information):.
    • User-defined metadata (BigQuery Column Setting information)
    Table for metadata transfer

    Only metadata for table columns for which at least one of the basic or user-defined metadata has already been entered will be transferred.

    column nametype (e.g. of machine, goods, etc.)Description.
    bigquery_project_idstringMetadata projects related to BigQuery
    bigquery_datasetstringDataset of metadata about BigQuery
    bigquery_tablestringBigQuery table name
    bigquery_columnstringBigQuery column name
    Only user-defined metadata (BigQuery Column Setting information) will be transferred if selected as the data type.
    trocco_metadata_logical_namestringLogical name of basic metadata
    trocco_metadata_descriptionstringBasic Metadata Description
    trocco_metadata_last_updated_attimestampLast update date and time of basic metadata
    (Fields defined in the template for user-defined metadata )(see below for details)User-defined metadata values
    user_defined_metadata_last_updated_attimestampLast modified date and time of user-defined metadata
    Field defined in template for user-defined metadata
    • About Data Types
      The data types defined in the template are converted to the following types

      • String -> string
      • Text(Markdown) -> string
      • Integer -> long
      • Boolean -> boolean
    • About column names

      • If non-alphanumeric characters are used in the field names defined in the Template for User-Defined Metadata, some Data Destination Connectors may not be able to forward the data.
      • In the above case, change the column name to one consisting only of alphanumeric characters in Column Setting in ETL Configuration STEP 2.

    column reference list

    Forward the column reference list.

    column nametype (e.g. of machine, goods, etc.)Description.
    bigquery_src_project_idstringMetadata projects related to BigQuery (references)
    bigquery_src_datasetstringDataset of metadata about BigQuery (references)
    bigquery_src_tablestringBigQuery table name (reference source)
    bigquery_src_columnstringBigQuery column name (reference source)
    bigquery_dst_project_idstringMetadata projects related to BigQuery (references)
    bigquery_dst_datasetstringDataset of metadata about BigQuery (references)
    bigquery_dst_tablestringBigQuery table names (references)
    bigquery_dst_columnstringBigQuery column names (references)
    creation_typestringHow to create a column reference
  • For automatic definition: trocco_data_source
  • If user-defined: user
  • created_bystringEmail address of the column reference creator
    last_updated_attimestampLast modified date and time of the column reference

    ETL Configuration

    job execution history

    ETL Job transfers historical data about previously executed ETL Jobs.

    column nametype (e.g. of machine, goods, etc.)Description.
    job_idlongETL Job ID
    job_urlstringETL Job URL
    job_definition_idlongETL Job Source ETL Configuration ID
    job_definition_urlstringETL Job Source ETL Configuration URL
    job_definition_namestringName of Data Source ETL Configuration from which the Job Source was created.
    executor_typestringHow to execute an ETL Job
  • manual:Manual execution
  • Scheduler: Scheduled execution
  • workflow: Executed as a workflow task
  • api: Execution by API
  • job_dependency: Execution by trigger job
  • statusstringStatus of ETL Job
  • queued: waiting for execution
  • setting_up: preparing to run
  • executing: Executing
  • interrupting: execution is interrupted
  • Succeeded: Execution completed (success)
  • error: Execution completed (error)
  • canceled: Execution completed (canceled)
  • skipped: execution completed (skipped)
  • transfer_recordslongNumber of records transferred
    transfer_byteslongnull
    skipped_recordslongnull
    started_attimestampnull
    finished_attimestampnull
    created_attimestampDate and time of execution of ETL Job ( %Y-%m-%d %H:%M:%S %Z)

    Data Mart Configuration

    job execution history

    Transfers historical data on previously executed data mart jobs.

    column nametype (e.g. of machine, goods, etc.)Description.
    datamart_job_idlongDATAMART Job ID
    datamart_job_urlstringData Mart Job URL
    datamart_definition_idlongData Mart Configuration ID from which the job originated
    datamart_definition_urlstringData Mart Configuration URL from which the job was created
    datamart_definition_namestringName of the Data Mart Configuration from which the job originated.
    executor_typestringHow to run a data mart job
  • manual:Manual execution
  • Scheduler: Scheduled execution
  • workflow: Executed as a workflow task
  • job_dependency: Execution by trigger job
  • statusstringData Mart Job Status
  • queued: waiting for execution
  • setting_up: preparing to run
  • executing: Executing
  • interrupting: execution is interrupted
  • Succeeded: Execution completed (success)
  • error: Execution completed (error)
  • canceled: Execution completed (canceled)
  • skipped: execution completed (skipped)
  • transfer_recordslongNumber of records transferred
    started_attimestampnull
    finished_attimestampnull
    created_attimestampDate and time of execution of the data mart job ( %Y-%m-%d %H:%M:%S %Z)

    Workflow

    job execution history

    Transfers historical data about previously executed Workflow Jobs.

    column nametype (e.g. of machine, goods, etc.)Description.
    pipeline_job_item_idlongUnique ID associated with Workflow Job ID
    pipeline_job_idlongWorkflow Job ID
    pipeline_job_urlstringWorkflow Job URL
    pipeline_definition_idlongWorkflow Job's original Workflow Job ID
    pipeline_definition_urlstringWorkflow Job's original Workflow Job URL
    pipeline_definition_namestringThe name of the Workflow Job's original Workflow definition
    executor_typestringHow to Perform a Workflow Job
  • manual:Manual execution
  • Scheduler: Scheduled execution
  • workflow: Executed as a workflow task
  • api: Execution by API
  • retry: Retry execution (reexecution from the stop position or reexecution by retry setting)
  • canceled_typestringnull
    statusstringWorkflow Job Status
  • queued: waiting for execution
  • setting_up: preparing to run
  • executing: Executing
  • interrupting: execution is interrupted
  • Succeeded: Execution completed (success)
  • error: Execution completed (error)
  • canceled: Execution completed (canceled)
  • skipped: execution completed (skipped)
  • retry_waiting: Waiting state from execution completion (error) state to start retry (only if retry count is set)
  • started_attimestampnull
    finished_attimestampnull
    created_attimestampWorkflow Job execution date and time ( %Y-%m-%d %H:%M:%S %Z)

    Was this article helpful?