# Senior Data Engineer

**Company:** [Weekday AI](http://jobs.workable.com/companies/pxG9rDgnvZm2c86JUchT1j.md)
**Location:** Chennai, India
**Workplace:** on site
**Employment type:** Full-time
**Department:** Weekday's Client via platform

[Apply for this job](http://jobs.workable.com/view/26887d4d-90d0-43b9-a4ad-bb1073edb5c4)

## Description

**This role is for one of the Weekday's clients**

**Salary range: Rs 1500000 - Rs 3000000 (ie INR 15-30 LPA)**

Experience: 6+ yrs

Location: Chennai

Job type: full-time

We are looking for a highly experienced Data Engineer to design and build scalable, high-performance data systems that power enterprise-grade analytics and real-time data-driven applications. The role involves working on complex data architectures, integrating multiple data sources, and building reliable pipelines that support both batch and streaming use cases.

You will be responsible for developing robust data engineering solutions across cloud and distributed environments, ensuring seamless data flow, high performance, and strong data quality standards. The role requires strong hands-on expertise in Python, Snowflake, and modern data engineering ecosystems.

This position is ideal for someone who enjoys solving large-scale data challenges, working with modern data warehouses, and building systems that support real-time decision-making across the organization.

## Requirements

### Key Responsibilities

-   Design, develop, and maintain scalable data pipelines for ingestion, transformation, and integration across multiple data sources
-   Build and optimize ETL/ELT workflows to support batch and real-time data processing requirements
-   Work with relational databases such as MySQL and PostgreSQL as well as NoSQL systems like MongoDB and Cassandra
-   Develop and manage data pipelines using modern data warehouse platforms such as Snowflake and Redshift
-   Process large-scale structured and unstructured data using AWS services such as S3, Athena, and Glue
-   Implement real-time streaming pipelines using Kafka for data ingestion and event-driven architecture
-   Utilize Apache Spark for distributed data processing and large-scale analytics (batch and streaming)
-   Collaborate with cross-functional teams including product, analytics, and engineering to translate business requirements into technical data solutions
-   Ensure data reliability, accuracy, consistency, and governance across all systems
-   Monitor, troubleshoot, and optimize pipeline performance for scalability and cost efficiency
-   Build reusable data engineering components and improve existing data infrastructure
-   Implement best practices for data modeling, storage optimization, and pipeline architecture
-   Work closely with cloud and DevOps teams to manage data infrastructure on AWS, GCP, or Azure environments
-   Continuously evaluate and adopt new technologies and frameworks in the data engineering ecosystem

### What Makes You a Great Fit

-   6–9 years of hands-on experience in data engineering and large-scale data pipeline development
-   Strong proficiency in Python for data processing and backend engineering tasks
-   Solid experience working with Snowflake and modern data warehouse platforms
-   Strong understanding of SQL and relational databases such as MySQL and PostgreSQL
-   Experience working with NoSQL databases like MongoDB or Cassandra
-   Hands-on experience with Kafka for real-time streaming and event-driven data systems
-   Strong understanding of ETL/ELT processes, data modeling, and pipeline optimization
-   Experience with big data frameworks such as Apache Spark is highly preferred
-   Familiarity with cloud platforms such as AWS, GCP, or Azure is an added advantage
-   Experience working with AWS services like S3, Athena, and Glue for data processing
-   Strong analytical and problem-solving skills with attention to detail
-   Ability to collaborate effectively with cross-functional teams in fast-paced environments
-   Strong communication skills with the ability to explain technical concepts clearly
-   A mindset focused on building scalable, reliable, and production-grade data systems
