Synechron Logo

Synechron

Data Engineer – AWS + Hadoop

Reposted 21 Days Ago
Be an Early Applicant
In-Office
Bellandur, Bangalore, Karnataka
Senior level
In-Office
Bellandur, Bangalore, Karnataka
Senior level
The Data Engineer will design and implement scalable ETL/ELT pipelines, build data ingestion frameworks, and manage data governance using AWS and Hadoop technologies, while optimizing for performance and reliability.
The summary above was generated by AI
Job Title: Data Engineer – AWS + Hadoop

Location: Bangalore
Experience: 7+ Years
 

About the Role

We’re looking for a seasoned Data Engineer with hands-on expertise in AWS data services and the Hadoop ecosystem. You will design, build, and optimize batch/streaming data pipelines, enable reliable data ingestion/processing, and support analytics, ML, and BI use cases at scale.

Key Responsibilities
  • Design and implement scalable ETL/ELT pipelines for batch and streaming workloads.
  • Build data ingestion frameworks using Kafka/Kinesis, and process data with Spark (PySpark/Scala).
  • Develop and optimize data lakes and data warehouses on AWS (S3, Glue, EMR, Athena, Redshift).
  • Manage and tune Hadoop ecosystem components (HDFS, Hive, Spark, Oozie/Airflow, Sqoop).
  • Model data (star/snowflake), manage schemas, partitioning, and metadata; ensure data quality (DQ checks).
  • Implement data governance, security, and access controls (IAM, Lake Formation, encryption, key management).
  • Set up orchestrations and CI/CD for data jobs (Airflow/AWS Step Functions, Jenkins/GitHub Actions).
  • Monitor pipelines and optimize cost, performance, and reliability (CloudWatch, logs, metrics).
  • Collaborate with Analytics/ML/BI teams; provide high-quality curated datasets and APIs/Views.
  • Document solutions, conduct code reviews, and enforce engineering best practices.
Required Skills & Qualifications
  • 7+ years in Data Engineering with large-scale distributed data systems.
  • Strong experience with AWS data stack: S3, Glue, EMR, Athena, Lambda, Redshift, IAM, CloudWatch.
  • Hands-on with Hadoop ecosystem: HDFS, Hive, Spark (PySpark/Scala), Kafka, Oozie/Airflow.
  • Expertise in SQL (complex queries, performance tuning) and data modeling.
  • Practical knowledge of streaming (Kafka/Kinesis, Spark Streaming/Structured Streaming).
  • Experience with Python or Scala for data pipelines; Shell scripting.
  • Familiarity with Orchestration (Airflow/AWS Step Functions) and CI/CD for data jobs.
  • Strong understanding of security & governance (encryption, PII handling, RBAC, Lake Formation).
  • Proficient with version control (Git) and containers (Docker) for reproducible jobs.
  • Excellent problem-solving, communication, and collaboration skills.

S​YNECHRON’S DIVERSITY & INCLUSION STATEMENT
 

Diversity & Inclusion are fundamental to our culture, and Synechron is proud to be an equal opportunity workplace and is an affirmative action employer. Our Diversity, Equity, and Inclusion (DEI) initiative ‘Same Difference’ is committed to fostering an inclusive culture – promoting equality, diversity and an environment that is respectful to all. We strongly believe that a diverse workforce helps build stronger, successful businesses as a global company. We encourage applicants from across diverse backgrounds, race, ethnicities, religion, age, marital status, gender, sexual orientations, or disabilities to apply. We empower our global workforce by offering flexible workplace arrangements, mentoring, internal mobility, learning and development programs, and more.

All employment decisions at Synechron are based on business needs, job requirements and individual qualifications, without regard to the applicant’s gender, gender identity, sexual orientation, race, ethnicity, disabled or veteran status, or any other characteristic protected by law.

Candidate Application Notice

Top Skills

Aws,S3,Glue,Emr,Athena,Lambda,Redshift,Iam,Cloudwatch,Hadoop,Hdfs,Hive,Spark,Pyspark,Scala,Kafka,Oozie,Airflow,Sql,Python,Shell Scripting,Git,Docker

Similar Jobs

8 Days Ago
In-Office
Thanisandra Nagavara, Bangalore, Karnataka, IND
Senior level
Senior level
Fintech • Financial Services
The position involves designing and maintaining scalable ETL/ELT data pipelines on AWS, optimizing data ingestion frameworks, and ensuring data quality. The engineer will collaborate with teams on data modeling and deployment practices while documenting processes clearly.
Top Skills: AWSAws EmrAws GlueAws LambdaAws S3Aws Step FunctionsBig DataCi/CdDataopsEltETLPysparkPythonSpark
A Minute Ago
Easy Apply
Remote or Hybrid
India
Easy Apply
Senior level
Senior level
Big Data • Cloud • Software • Database
The Solutions Architect guides customers in designing and building systems using the data platform, ensuring technical alignment and customer success through collaboration and expertise in software development and database systems.
Top Skills: Apache KafkaC#C/C++JavaKubernetesMongoDBNode.jsPythonSQL
4 Minutes Ago
Easy Apply
Remote or Hybrid
India
Easy Apply
Senior level
Senior level
Big Data • Cloud • Software • Database
The Senior Solutions Architect supports customer success by designing scalable systems using MongoDB, collaborating with sales teams, and providing technical guidance to clients.
Top Skills: C#C/C++JavaMongoDBNode.jsPythonSQL

What you need to know about the Bengaluru Tech Scene

Dubbed the "Silicon Valley of India," Bengaluru has emerged as the nation's leading hub for information technology and a go-to destination for startups. Home to tech giants like ISRO, Infosys, Wipro and HAL, the city attracts and cultivates a rich pool of tech talent, supported by numerous educational and research institutions including the Indian Institute of Science, Bangalore Institute of Technology, and the International Institute of Information Technology.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account