Hadoop Operations Manager

Palo Alto, CA, US

Job Description / Skills Required

In this management position, you will lead the design, implementation, and administration of large Hadoop clusters and related technologies with the help of Hadoop onsite/offshore engineers. These clusters help our business customers as well as the EDW teams to solve difficult analytics, business intelligence, machine learning and many other problems. The successful candidate will demonstrate leadership, technical and collaboration skills to deliver and maintain the Hadoop infrastructure, enhancing existing clusters, and sustaining operations.

Key Qualifications

Candidate should have several years of experience building and administering large business critical Hadoop clusters. The ideal candidate will bring a lot of energy, initiative, and excitement. Will be ready to learn and explore new ideas, processes, methodologies and leading-edge technologies.

At least 6 years of experience in engineering, system administration and/or Devops
At least 4 years of experience in designing, implementing and administering Hadoop
Managing 24×7 shifts with the onsite/offsite engineers, responding to PagerDuty alerts
Experience working within the requirements of a change management system
Excellent leadership and communication skills
Strong analytical and organizational skills
Proven ability to adapt to a dynamic project environment and manage multiple projects simultaneously
Proven ability to collaborate with application development and other cross functional teams
Ability to coach and provide guidance to junior team members
Key leadership skill set needed in the following areas to manage teams in multiple locations:

Understanding of HDFS
Understanding of YARN
Using Ambari in administering Hadoop
Experience in administering Cluster size greater than 6 PB OR 200 Datanodes
Knowledge in bash shell scripting to automate administration tasks
Onboarding new users to Hadoop
Maintaining SOX compliance
Experience in writing HQL (Hive Queries)
Understanding of Hive metadata store objects
Monitoring Linux host health in Ganglia and responding to Nagios alerts/Pager alerts
Experience in capacity planning the big data infrastructure.
Providing optimization tips to ETL team about efficient methods in performing operations in Hadoop Platform (Hive)
Involvement on Open source products/technologies development is a great plus

Groupon provides a global marketplace where people can buy just about anything, anywhere, anytime. We’re enabling real-time commerce across an expanding range of categories including local businesses, travel destinations, consumer products, and live or lively events. At the same time, we are providing advertising options and tools that merchants can use to grow and manage their businesses. Culturally, we believe that great people make great companies and that starting with the customer and working backward moves us forward. Community matters to us on an internal, local and global scale—it’s fundamental to our company’s growth and to the well-being of the world at large. We also value self-awareness, candor, lunch and WiFi. If we match with you, please apply to join us.