There are a few concepts in the big data world I have been struggling to really understand. One of them is the data lake model. The hardware and physical structure of Hadoop have always made logical sense to me. However coming from a structured database world I have always wondered how the data is organized. There have to be some ways to manage the data and the metadata so when the time comes to retrieve your data, it can be found quickly. I have always pictured a bunch of random data stuffed into a data lake. I think I wasn’t that far off. I found this article proposing standards for data lakes. This article focuses on something called Semantic Data Models to keep big complex data lakes from becoming “data swamps.” The proposals make sense to a structured data guy like me, and they help make the whole thing a little more clear. I hope you enjoy the article, and it brings you a little more insight into the big data world.