Support Questions

Find answers, ask questions, and share your expertise

Data Lake Architecture

avatar
Rising Star

Hi all,

Can anyone advise me on how to organize data in my data lake? For instance, split data into categories, like Archived Data, that probably won't be used but it's needed, another division for raw data, and the last one for transformed data.

I'm using Hbase and Hive for now.

Thanks

1 ACCEPTED SOLUTION

avatar

Let me put it in simple words.

Basically 4 layers are needed in Datalake.

Landing Zone: It contains all the raw data from all different source systems available. There is no cleansing and any logics applied on this layer. It just a one to one move from outside world into Hadoop. This raw data can be consumed by different application for analysis/predictive analysis as only raw data will give us many insights about the data.

Cleansing Zone: Here data's are properly arranged. For Example: Defining proper data type for the schema and cleansing, trimming works. There is no need for data model as well till this layer. If there are any data's which has to cleansed regularly and consumed by application then this layer would serve that purpose.

Transformed Zone: As the name suggest data modelling, proper schema are applied to build this layer. In short if there are any reports which has to run on a daily basis, on some conformed dimension which can serve for any specific purpose can be built in this layer. Also datamart which serves only for one/two particular needs can be built. For example: Conformed dimension like demographic, geography & data/time dimensions can be built in this layer which can satisfy your reporting as well as act as a source for machine learning algorithms as well.

https://hortonworks.com/blog/heterogeneous-storages-hdfs/

https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.0/bk_hdfs_admin_tools/content/storage_policie...

check for this links for archival storage. Archival can be built in landing zone itself once you have decided to move it to archive you are compress the data and push it to archive layer. Check the above links so that resources are properly used and allocated.

If needed check this book from oreilly. It covers a wide range of uses based data lake architecture.

http://www.oreilly.com/data/free/architecting-data-lakes.csp

View solution in original post

3 REPLIES 3

avatar

@Francisco Pires

Typically, for data warehousing, we recommend logically organizing your data into tiers for processing.

14051-screen-shot-2017-03-26-at-123125-pm.png

The physical organization is a little different for everyone, but here is an example for Hive:

14052-physical.jpg

avatar
Rising Star

thanks, this will help.

avatar

Let me put it in simple words.

Basically 4 layers are needed in Datalake.

Landing Zone: It contains all the raw data from all different source systems available. There is no cleansing and any logics applied on this layer. It just a one to one move from outside world into Hadoop. This raw data can be consumed by different application for analysis/predictive analysis as only raw data will give us many insights about the data.

Cleansing Zone: Here data's are properly arranged. For Example: Defining proper data type for the schema and cleansing, trimming works. There is no need for data model as well till this layer. If there are any data's which has to cleansed regularly and consumed by application then this layer would serve that purpose.

Transformed Zone: As the name suggest data modelling, proper schema are applied to build this layer. In short if there are any reports which has to run on a daily basis, on some conformed dimension which can serve for any specific purpose can be built in this layer. Also datamart which serves only for one/two particular needs can be built. For example: Conformed dimension like demographic, geography & data/time dimensions can be built in this layer which can satisfy your reporting as well as act as a source for machine learning algorithms as well.

https://hortonworks.com/blog/heterogeneous-storages-hdfs/

https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.0/bk_hdfs_admin_tools/content/storage_policie...

check for this links for archival storage. Archival can be built in landing zone itself once you have decided to move it to archive you are compress the data and push it to archive layer. Check the above links so that resources are properly used and allocated.

If needed check this book from oreilly. It covers a wide range of uses based data lake architecture.

http://www.oreilly.com/data/free/architecting-data-lakes.csp