You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+3-4Lines changed: 3 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -92,6 +92,7 @@ The **unique identifier** of a DataProduct is the concatenation of Domain, Name
92
92
*`FullyQualifiedName: [String]` Human-readable that uniquely identifies an entity
93
93
*`Description: [String]` detailed description about the process, its purpose and characteristics
94
94
*`ResourceType: [String]` explain what type of workload is: Ingestion ETL, Streaming, Internal Process, etc.
95
+
*`Type: [String]` This is an enum `[HouseKeeping|DataPipeline]`, `Housekeeping` is for all the workloads that are acting on internal data without any external dependency. `DataPipeline` instead is for workloads that are reading from outputport of other DP or external systems.
95
96
*`Technology: [String]` this is a list of technologies: Airflow, Spark, Scala. It is a free field but it is useful to understand better how it is behaving
96
97
*`Description: [String]` detailed explaination about the purpose of the workload, what sources is reading, what business logic is apllying, etc
97
98
*`Version: [String]` Specific version of the workload. Displayed as X.Y.Z where X is the major version of the data product, Y is minor feature and Z is patch. Major version(X) is also shown in the component ID and those 2 fields(version and ID) are always aligned with one another.
@@ -109,11 +110,9 @@ The **unique identifier** of a DataProduct is the concatenation of Domain, Name
109
110
*`ID: [String]` the unique identifier of the Storage Area
110
111
*`Name: [String]` the name of the Storage Area
111
112
*`FullyQualifiedName: [String]` Human-readable that uniquely identifies an entity
112
-
*`ResourceType: [String]` explain what type of workload is, at the moment: batch or streaming
113
-
*`Type: [String]` This is an enum `[HouseKeeping|DataPipeline]`, `Housekeeping` is for all the workloads that are acting on internal data without any external dependency. `DataPipeline` instead is for workloads that are reading from outputport of other DP or external systems.
114
-
*`Technology: [String]` this is a list of technologies: S3, ADLS, GFS.
113
+
*`Technology: [String]` this is a list of technologies: S3, ADLS, SQLServer, Kafka.
115
114
*`Description: [String]` detailed explanation about the function and the meaning of this storage area
116
-
*`Tags: [Array[Yaml]]` Free tags at OutputPort level ( please refer to OpenMetadata https://docs.open-metadata.org/openmetadata/schemas/entities/tagcategory )
115
+
*`Tags: [Array[Yaml]]` Free tags at Storage area level ( please refer to OpenMetadata https://docs.open-metadata.org/openmetadata/schemas/entities/tagcategory )
117
116
*`Specific: [Yaml]` this is a custom section where we can put all the information strictly related to a specific technology or dependent from a standard/policy defined in the federated governance.
0 commit comments