Your browser does not support javascript! Please enable it, otherwise web will not work for you.

Hadoop- Senior/team Lead Software Engineering @ Sabre Travel

Home > Programming & Design

 Hadoop- Senior/team Lead Software Engineering

Job Description


Job Description:
Sabre Enterprise Data & Analytics Big Data Hadoop Platform team is primarily responsible for the Hadoop Big data lake platform and applications stability. We are looking to hire a Team Lead/Senior Hadoop administrator to provide application and system support. Hadoop administrator will also help developing scripts for platform automation. The main role of the Team Lead/Senior Hadoop Administrator is to investigate and diagnose issues with Big Data systems, data ingestion and transformation processes as a part of Hadoop Big data lake team. This work includes researching the problem, identifying root cause, developing and executing workarounds and documenting the work. When appropriate, the Hadoop Big data administrator will help to develop tools to facilitate implementation of workarounds. This includes working closely with Development and Architecture teams and internal Sabre stakeholders.
Responsibilities:
* Responsible for implementation and ongoing administration of Hadoop infrastructure.
* Work with the Dev teams to optimize cluster usage and ensure timely execution of business-critical workloads
* Install Hadoop updates, patches, and version upgrades as required
* Perform routine cluster maintenance, such as provisioning new nodes and performing HDFS backups and restores
* Routine cluster monitoring and troubleshooting
* Configuration and monitoring of job isolation, security, and resource queues
* Develop scripts and tools to automate common Hadoop administration tasks
* Participate in an on-call rotation with 24x7 availability requirements
* Screen Hadoop cluster job performances and capacity planning
* Monitor Hadoop cluster connectivity and security
* Manage and review Hadoop log files.
* File system management and monitoring.
* Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
* Setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users.
* Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise, Dell Open Manage and other tools.
* Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
* HDFS support and maintenance.
* Develop cloud formation, ansible and Cloudera navigator scripts for cloud migration and Hadoop platform automation.
* Clear understanding of security concepts including Kerberos, Key Trustee, TSL, LDAP and others.
* Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
* Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
* Point of Contact for Vendor escalation

EXPERIENCE: 5+ years of experience 
REQUIRED SKILLS:
* Five plus years' experience with at least 4 years as a Hadoop Administrator
* Experience as an Architect building Enterprise scale, production quality Products and Solutions Development
* At least 3 years of Enterprise level Big Data Systems Architecture, Design and Development experience and 5+ years of Enterprise level Application, Database, Data Warehouse, and BI Development. Prefer candidates with hands-on Experience with Big Data Systems Such as Hadoop, Vertica, HBase, and Cassandra.
* Experience in XML, JMS (Message Queues) and Web Services technologies
* Experience with JUnit (any Java Unit Testing Framework), Test Driven development, maven
* Strong scripting skills, preferably in Python and Linux shell scripts and automation using scripting. Demonstrated knowledge of Linux operating systems.
* Demonstrated knowledge of Object-Oriented Analysis and Design
* Functional understanding of relational databases (Teradata, Vertica, Big Data, Oracle) or similar databases
* Proficient knowledge of complex SQL with any major RDBMS and SQL performance tuning
* Good Experience with OLAP concepts and methods
* Knowledge of Web Technologies such as HTML, CSS, JavaScript
* Design and development using Java development stack: Java EE, JSF, Hibernate, Aspect Oriented Programing (AOP), Webservices & Spring Framework.
* Experience in Hadoop Monitoring tools (Nagios, Ganglia, CDH CM)
* Experience in Web/App Server & SOA administration (Tomcat, JBoss, etc.)
* Knowledge of AWS, Azure, EMR, HDInsight's.
* Knowledge of Big Data ecosystem (Hadoop, HDFS, Map Reduce, Yarn, Pig, Hive, Oozie, Sqoop, Flume) to derive insights/analytics from Big Data.
Exp: 5- 9 yrs
CTC: upto 22LPA
Job Type: Full-time
No.of Positions: 05
Interview Mode: F2F(1st Round). 2nd Round Tech & Manager Round- Through Webex
Date of Interview: 14th November, Thursday
Interested candidates who are available for the interview on 14th November, Thursday can share their updated resumes to de**************n@sa**e.com. 


Job Classification

Industry: IT-Software / Software Services,
Functional Area: IT Software - Application Programming, Maintenance,
Role Category: Programming & Design
Role: Programming & Design
Employement Type: Full time

Education

Under Graduation: Any Graduate in Any Specialization, Graduation Not Required
Post Graduation: Any Postgraduate in Any Specialization, Post Graduation Not Required
Doctorate: Any Doctorate in Any Specialization, Doctorate Not Required

Contact Details:

Company: Sabre Travel

+ View Contactajax loader


Keyskills:   Cloudera Hadoop Hadoop Developer Hadoop Administrator

 Job seems aged, it may have been expired!
 Fraud Alert to job seekers!

₹ 7,00,000 - 17,00,000 P.A

Sabre Travel

About Us: Sabre is the global leader in innovative technology that leads the travel industry. We are always looking for bright and driven people who have a penchant for technology, are hackers at heart and want to hone their skills. If you are interested in challenging work, being part of a global ...