Skip to content

Commit 6134bd2

Browse files
authored
Update README.md
1 parent c77f61b commit 6134bd2

File tree

1 file changed

+3
-4
lines changed

1 file changed

+3
-4
lines changed

README.md

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -92,6 +92,7 @@ The **unique identifier** of a DataProduct is the concatenation of Domain, Name
9292
* `FullyQualifiedName: [String]` Human-readable that uniquely identifies an entity
9393
* `Description: [String]` detailed description about the process, its purpose and characteristics
9494
* `ResourceType: [String]` explain what type of workload is: Ingestion ETL, Streaming, Internal Process, etc.
95+
* `Type: [String]` This is an enum `[HouseKeeping|DataPipeline]`, `Housekeeping` is for all the workloads that are acting on internal data without any external dependency. `DataPipeline` instead is for workloads that are reading from outputport of other DP or external systems.
9596
* `Technology: [String]` this is a list of technologies: Airflow, Spark, Scala. It is a free field but it is useful to understand better how it is behaving
9697
* `Description: [String]` detailed explaination about the purpose of the workload, what sources is reading, what business logic is apllying, etc
9798
* `Version: [String]` Specific version of the workload. Displayed as X.Y.Z where X is the major version of the data product, Y is minor feature and Z is patch. Major version(X) is also shown in the component ID and those 2 fields(version and ID) are always aligned with one another.
@@ -109,11 +110,9 @@ The **unique identifier** of a DataProduct is the concatenation of Domain, Name
109110
* `ID: [String]` the unique identifier of the Storage Area
110111
* `Name: [String]` the name of the Storage Area
111112
* `FullyQualifiedName: [String]` Human-readable that uniquely identifies an entity
112-
* `ResourceType: [String]` explain what type of workload is, at the moment: batch or streaming
113-
* `Type: [String]` This is an enum `[HouseKeeping|DataPipeline]`, `Housekeeping` is for all the workloads that are acting on internal data without any external dependency. `DataPipeline` instead is for workloads that are reading from outputport of other DP or external systems.
114-
* `Technology: [String]` this is a list of technologies: S3, ADLS, GFS.
113+
* `Technology: [String]` this is a list of technologies: S3, ADLS, SQLServer, Kafka.
115114
* `Description: [String]` detailed explanation about the function and the meaning of this storage area
116-
* `Tags: [Array[Yaml]]` Free tags at OutputPort level ( please refer to OpenMetadata https://docs.open-metadata.org/openmetadata/schemas/entities/tagcategory )
115+
* `Tags: [Array[Yaml]]` Free tags at Storage area level ( please refer to OpenMetadata https://docs.open-metadata.org/openmetadata/schemas/entities/tagcategory )
117116
* `Specific: [Yaml]` this is a custom section where we can put all the information strictly related to a specific technology or dependent from a standard/policy defined in the federated governance.
118117

119118

0 commit comments

Comments
 (0)