Key Responsibilities
- Installation, configuration, and upgrading of Big Data solutions while leading the team for smooth operationalization
- Benchmark systems, analyze system bottlenecks, and propose solutions to eliminate them
- Provide deep subject matter expertise, advice for the program delivery while establishing architectural principles and patterns
- Conduct technical and design reviews while understanding system integration concepts and patterns
- Review the business requirements captured and assess gaps
- Innovation role on Digital Data with latest technologies and identify customer expectations and competitor movements fast and accurate and same time coordination among organizational units to enhance operational backbone
- Carrying out R&Ds to select best suited Big Data technologies
- Design ETL flow using technologies such as Sqoop / Flume / Kafka / Hive / Oozie / PIG / AWS Glue / Airflow
- Responsible for performing monthly / quarterly performance analysis to identify the application bottlenecks, issues and new resources requirements while defining the process and continuous improvement
- Troubleshooting on system issues, attending to faults, analysis of issues, applying corrective action / workarounds, proving RCAs
- Periodically update relevant patches on Big Data platforms while assessment of the Big Data platform to identify gaps for patch and make long-term and short-term plans to bridge the gaps
- Perform Unix / Linux administrative tasks while planning and execution of performance improvement activities essential to Maintain smooth operational environment
- Implementing and managing OS level security and patch upgrades while maintaining minimum baseline security of servers (Periodic VA Scans / Recommending patches)
- Maintaining up to date documentation, connectivity diagrams, and related information for all the application liaised within the responsibility of this position
- Career path development and create Learning opportunities for subordinates
Key Requirements
- More than 5 years’ experience in a relevant position at similar field
- Bachelors’ degree in IT/Computer science from a recognized university
- Having Scalable Machine Learning certification, certified on Data warehousing with MSSQL Server, certified on Cloudera Developer for Apache Hadoop & Spark, certified on Cloudera Designing and Building Applications and AWS solution architect is advantageous
- Strong Knowledge Big Data related technologies (Hadoop, MapReduce, Pig, Hive, HBase, Oozie, Flume, ZooKeeper, Kafka, Spark, Solr, Impala, Hue)
- Knowledge on RDBMS (MySQL, Oracle), PLSQL, Shell Scripting, MySQL Clustering, Impala Load Balancing, NoSQL / column-oriented Databases (Preferred Snowflake / ClickHouse)
- Strong Programming skills with Scala / Python/ Java