Build Big Data Pipelines w/ Hadoop, Flume, Pig, MongoDB
- 4.4
Brief Introduction
Learn how to combine Hadoop, MongoDB, Pig, Sqoop and Flume to Architect and Build Big Data Pipelines and Data lakes.Description
How do you build end-to-end Big Data Pipelines using multiple Big Data Technologies? You have seen courses and books that teach you individual technologies, but how do you combine and apply them to solve your business problems? This course teach you exact that !
Building Big Data Solutions require you to acquire data from multiple sources, transport them, process them and store them in Big Data repositories. You have to do that with scalability and reliability. Big Data Technologies like Hadoop, Sqoop, Pig, Flume etc. solve individual problems, but building an end-to-end solution requires stitching them together. This course teaches you how to do that. You solve complete business problems by building end-to-end pipelines in this course.
Requirements
- Requirements
- Hadoop
- Linux / CDH 5
- Java Programming preferred