When one starts a new project that involves analyzing his company's data especially when the data is stored across functional areas, that person is in trouble. The data lake model helps in this case. To get access to data doesn't require an integration effort, because data is already there in the lake and one can apply MapReduce and other algorithms to use it. In the lake some data are unstructured or not structured by us for a given project. To construct a data lake one needs to learn some of the Hadoop stack such as Sqoop, Oozie and Flume. Next a data scientist should be found who understands Hadoop as well as business and the company’s business data in particular. Then one should start with basic cases and use simple and familiar tools like Tableau to make nice charts, graphics, and reports demonstrating that he can do something useful with the data. Next security up front should be considered, as well as who can access what data. Use of core Hadoop platform is beneficial. Apart from this one should keep in mind that lake security may have business unit implications and one should not have a lot of mini lakes i.e. data ponds that are separate and not equal. Read more at:http://www.infoworld.com/d/application-development/how-create-data-lake-fun-and-profit-246874?page=0,0