You can also subscribe to our channel and follow this playlist on YouTube for the videos.
Quite often we have to build Big Data clusters using plain vanilla distributions rather than using vendor distributions such as Cloudera or Hortonworks. It is not practical to manually set up the cluster, rather we need to use server automation tools like Puppet, Chef or Ansible. We are going to set up a 7 node Hadoop (HDFS + YARN) cluster using Ansible.
On top of free content we also support in case you run into any issues. Please sign up to our community (as we do not have single sign on enabled yet)
In case you want to get notifications about this live session and other live sessions in future, please join systems engineering group.
Live sessions will be conducted on July 21st and 22nd via YouTube live, after which the same videos archive will be available in this course.
Here is the agenda for the course
Provision baremetal server and install CentOS (OVH)
Setup single node hadoop cluster
Setup 7 virtual machines on bare metal server using Vagrant
Is it a user friendly ui tool where you drag and drop modules that you want to use? or do you actually need to write code for it to do what you want? I believe talend is similar where you drag and drop various modules, but i have on my mind that the performance isnt so great. anyway Ive seen more job postings mentioning gobblin recently and Im wondering (in general) is it a regular developer role writing applications using a framework/library, or it figuring out which modules to drag into a chain?
Whitepaper explaining why does an organization need to consider a Big Data solution and does it help address specific challenges. Moreover what factors should be evaluated when deciding whether a Big Data solution makes sense for your data and analytics needs?