Key Job Functions and Duties :
Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users.
Cluster maintenance for Cloudera Management Cluster
Performance tuning of Hadoop clusters, MapReduce, Yarn and Spark
Screen Hadoop cluster job performances and capacity planning
Troubleshooting errors on Infrastructure and Platform level (Network / Connectivity / System) encountered by the team
Monitor Hadoop cluster connectivity and security
Manage and review Hadoop log files.
File system management and monitoring.
HDFS support and maintenance.
Cloud Administration (AWS / Azure)
Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
Point of Contact for Vendor escalation
Desired Experience & Qualities
Broad technical background with at least three year’s work experience in a similar role as a Hadoop Administrator
Knowledge of RHEL7 / CentOS7 preferable 1 to 3 Years Kafka Administrator
Preferred : Experience in secure environments (Kerberized, Encrypted) in Kafka
Preferred : Experience in other apache big data applications such as HBase, Cassandra, Storm, etc.
Qualifications and Accreditations :
RHCSA (Red Hat Certified System Administrator) is a distinct advantage