Hadoop Administrator: Expected Roles and Responsibilities

Hadoop is gradually becoming part of production environment in several organizations and it has been seamlessly adopted by traditional enterprise IT solutions in recent times. This has increased the need for Hadoop Administrators to take care of the large Hadoop Clusters. Well, that is not only responsibility but definitely the primary one; skills include working knowledge of how to deploy Hadoop cluster, add and remove nodes, keep track of jobs, monitor critical parts of the cluster, configure name-node high availability, schedule and configure it and take backups. In traditional terms, we can think about a role of a system administrator. But there is more!

Apart from having general operational expertise such as good troubleshooting skills, understanding of systems capacity, bottlenecks, basics of memory, CPU, OS, storage, and networks, Hadoop administrator must have good knowledge of Linux, since Hadoop runs on Linux. Hadoop related skills like HBase, Hive, Pig and Mahout would be essential along with familiarity with open source configuration management and deployment tools such as Puppet or Chef and Linux scripting.

Key responsibilities are list as below: * Responsible for implementation and ongoing administration of Hadoop infrastructure. * Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments. * Working with data delivery teams to setup new Hadoop users: this job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users. * Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise, Dell Open Manage and other tools. * Performance tuning of Hadoop clusters and Hadoop MapReduce routines. * Screen Hadoop cluster job performances and capacity planning * Monitor Hadoop cluster connectivity and security * Manage and review Hadoop log files. * File system management and monitoring. * HDFS support and maintenance. * Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability. * Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required. * Point of Contact for Vendor escalation

At Spring People, we are Hortonworks Certified Training delivery partners, and we bring to you various Hortonworks Certified courses to make you a Hadoop Expert with HDP Operations: Hadoop Administration I& HDP Operations Hadoop Adminstration 2: Courses. #BeTheExpert in Hadoop with SpringPeople.

Leave a Reply

Your email address will not be published. Required fields are marked *

CAPTCHA

*

About SpringPeople

Founded in 2009, SpringPeople is a global corporate training provider for high-end and emerging technologies, methodologies and products. As master partner for Pivotal / SpringSource, Elasticsearch, Typesafe, EMC, VMware, MuleSoft and Hortonworks, SpringPeople brings authentic, Certified training, designed and developed by the people who created the technology, to Corporates and Development/IT Professionals community in India. This makes SpringPeople an exclusive master certified training delivery wing, and one of the hand-picked few global partners, of these organizations - delivering their immensely popular, high-quality Certified training courses in India – for a fraction of what it costs globally.

Posts by SpringPeople