Rackspace Technology Logo

Rackspace Technology

Sr Big Data Engineer (GCP)- Airflow and Oozie

Sorry, this job was removed at 08:47 p.m. (IST) on Thursday, Nov 21, 2024
Be an Early Applicant
Remote
Remote

About the Role:

We are seeking a highly skilled and experienced Senior Big Data Engineer to join our dynamic team. The ideal candidate will have a strong background in developing and scaling both stream and batch processing systems, and a solid understanding of public cloud technologies, especially GCP. This role involves working in a remote environment, requiring excellent communication skills and the ability to solve complex problems independently and creatively.


What you will be doing

Build a reusable, and reliable code for stream and batch processing systems at scale. This includes working with technologies like Pub/Sub, Kafka, Kinesis, DataFlow, Flink, Hadoop, Pig, Hive, and Spark. Implementing automation/DevOps best practices for CI/CD, IaC, Containerization, etc.

Requirements

  • About the Role:
  • We are seeking a highly skilled and experienced Senior Big Data Engineer to join our dynamic team. The ideal candidate will have a strong background in developing batch processing systems, with extensive experience in Oozie, the Apache Hadoop ecosystem, Airflow, and a solid understanding of public cloud technologies, especially GCP. This role involves working in a remote environment, requiring excellent communication skills and the ability to solve complex problems independently and creatively.

  • What you will be doing
  • Develop scalable and robust code for batch processing systems. This includes working with technologies like Hadoop, Oozie, Pig, Hive, Map Reduce, Spark (Java), Python, Hbase
  • Develop, Manage and optimize data workflows using Oozie and Airflow within the Apache Hadoop ecosystem
  • Leverage GCP for scalable big data processing and storage solutions
  • Implementing automation/DevOps best practices for CI/CD, IaC, etc.

  • Requirements:
  • Experience with GCP managed services and understanding of cloud-based batch processing systems are critical.
  • Proficiency in Oozie, Airflow, Map Reduce, Java
  • Strong programming skills with Java (specifically Spark), Python, Pig, and SQL
  • Expertise in public cloud services, particularly in GCP.
  • Proficiency in the Apache Hadoop ecosystem with Oozie, Pig, Hive, Map Reduce
  • Familiarity with BigTable and Redis
  • Experienced in Infrastructure and Applied DevOps principles in daily work. Utilize tools for continuous integration and continuous deployment (CI/CD), and Infrastructure as Code (IaC) like Terraform to automate and improve development and release processes.
  • Ability to tackle complex challenges and devise effective solutions. Use critical thinking to approach problems from various angles and propose innovative solutions.
  • Worked effectively in a remote setting, maintaining strong written and verbal communication skills. Collaborate with team members and stakeholders, ensuring clear understanding of technical requirements and project goals.
  • Proven experience in engineering batch processing systems at scale.
  • Hands-on experience in public cloud platforms, particularly GCP. Additional experience with other cloud technologies is advantageous.

Similar Jobs

17 Days Ago
Remote
Chennai, Tamil Nadu, IND
Junior
Junior
Insurance • Software • Energy • Financial Services
As a Data Platform Engineer, you will monitor production processes, manage incident tickets, and ensure smooth operations in data platforms. Responsibilities include production monitoring, incident management, and platform support, focusing on improving workflows and system reliability.
An Hour Ago
Remote
Hybrid
Pune, Maharashtra, IND
Senior level
Senior level
Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
The Senior Data Engineer will design and maintain scalable data pipelines, develop API integrations, and collaborate with teams to enhance data models for business intelligence. Responsibilities include building algorithms, researching data acquisition opportunities, and providing technical consulting to peers.
13 Hours Ago
Easy Apply
Remote
3 Locations
Easy Apply
Senior level
Senior level
Cloud • Information Technology • Security • Software
As a Senior Data Engineer, you'll join the Data Enablement team at JumpCloud, focusing on enhancing data capabilities, modernizing the Data Warehouse/Lakehouse, and developing an Event Based Architecture while engaging with various stakeholders to improve data practices.
Top Skills: Data LakehouseData WarehouseEvent Based Architecture

What you need to know about the Bengaluru Tech Scene

Dubbed the "Silicon Valley of India," Bengaluru has emerged as the nation's leading hub for information technology and a go-to destination for startups. Home to tech giants like ISRO, Infosys, Wipro and HAL, the city attracts and cultivates a rich pool of tech talent, supported by numerous educational and research institutions including the Indian Institute of Science, Bangalore Institute of Technology, and the International Institute of Information Technology.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account