Metadata
DataFibers define three type of meta data model for data structuring and serializing as follows. There are defined in the package com.datafibers.model
.
JOB
The purpose of Job model is to keep all information for task/job either for connects or transforms type of tasks in DataFibers.
class DFJobPOPJ {
private String id;
private String taskId;
private String name;
private String connector;
private ConstantApp.DF_CONNECT_TYPE connectorType;
private String connectorCategory;
private String description;
private String status;
private String udfUpload;
private HashMap jobConfig;
private HashMap connectorConfig;
}
Attribute | Type | Usage |
id | String | Unique number string which is generated from mongodb, used as internal task id. |
taskId | String | Identify each task in a job. It is manually set now, will deprecated. |
name | String | Name of the job. |
connector | String | Name of the connector/task used. This maps to Kafka Connect name attribute. |
connectorType | Enum | Identify proper connector type from enum. |
connectorCategory | String | Identify the category of task, connect or transform. |
description | String | Task description as free text. |
status | String | Identify the task status, such as FAILED, RUNNING, etc. |
udfUpload | String | Only used for UDF transform to identify the path of UDF Jar. |
jobConfig | HashMap | Key-value pairs of job level configuration in details. |
connectorConfig | HashMap | Key-value pairs of task level configuration in details. |
Meta
Meta view will be a read-only view in DataFibers web console.
Attribute | Type | Usage |
topicId | String | Unique number string which is generated from mongodb, used as job id. |
topic | String | Identify subject of data. It is now Kafka topic. |
schemaID | String | Identify the schema in schema registry. |
schemaName | String | Identify the schema name in schema registry. |
schemaVersion | int | Identify the schema version associated with the topic in schema registry. |
sourceList | HashMap | Identify the list of source as where we ingest to this topic. |
targetList | HashMap | Identify the list of target as where we extract from this topic. |
consumeList | [String] | Identify the list of consumer group of Kafka. |
volumn | String | Identify the volumn of data in this topic in terms of MB/GB/TB. |
namespace | String | java package/maven like namespace for topic category and permission. |
topicCreateDate | String | Datetime when the topic is created. |
dataCreateDate | String | Datetime when the topic is firstly populated data. |
dataUpdateDate | String | Datetime when the latest data is been written. |
topicStatus | String | Identify whether the topic is being used or not, such as ACTIVE, DEACTIVE, DELETED, etc. |
permissions | String | Reserved for future permission control on the row. |
History
History view will be a read-only view in DataFibers web console.
Attribute | Type | Usage |
id | String | Unique number string which is generated from mongodb, used as history id. |
taskId | String | Identify each task in a job. |
taskName | String | Name of the task. |
StartDate | String | Timestamp when the task starts. |
EndDate | String | Timestamp when the task ends. |
rows | String | Rows that task processed sucessfully. |
rowsBad | String | Rows that task cannot processed. |
throughput | String | rows/seconds for the rows processed. |
status | String | Identify the task status, such as RUNNING, FAILED, SUCESSED etc. |
schedule | String | describe the schedule of the task. |
submitBy | String | Identify whom/where the task is submitted. |
sourceList | HashMap | Identify the list of source as that this task ingest. |
targetList | HashMap | Identify the list of target as that this task extract. |
validation | HashMap | Identify the validation for the task. |
Last updated