Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
7.0 - 10.0 years
7 - 10 Lacs
Gurgaon, Haryana, India
On-site
Design, develop, and maintain scalable data pipelines and data assets using modern data engineering techniques Strong experience in code optimisation using Spark SQL and PySpark Apply AWS architecture knowledge, especially S3, EC2, Lambda, Redshift, CloudFormation Refactor legacy codebase to improve readability, maintainability, and performance Write tests before code to ensure functionality and catch bugs early Debug complex code and resolve performance, concurrency, or logic issues Role Requirements and Qualifications: Minimum 7+ years of strong hands-on programming experience with PySpark / Python / Boto3 Experience using Python frameworks and libraries in line with Python best practices Understanding of code versioning tools (Git), repositories (e.g., JFrog Artifactory) Strong commitment to TDD (Test-Driven Development), unit testing, and participating in code reviews Excellent problem-solving skills, analytical thinking, and ability to work independently
Posted 4 days ago
6.0 - 12.0 years
2 - 11 Lacs
Hyderabad / Secunderabad, Telangana, Telangana, India
On-site
Develop and implement efficient data pipelines using Apache Spark (PySpark preferred) to process and analyze large-scale data. Design, build, and optimize complex SQL queries to extract, transform, and load (ETL) data from multiple sources. Orchestrate data workflows using Apache Airflow , ensuring smooth execution and error-free pipelines. Design, implement, and maintain scalable and cost-effective data storage and processing solutions on AWS using S3, Glue, EMR, and Athena . Leverage AWS Lambda and Step Functions for serverless compute and task orchestration in data pipelines. Work with AWS databases like RDS and DynamoDB to ensure efficient data storage and retrieval. Monitor data processing and pipeline health using AWS CloudWatch and ensure smooth operation in production environments. Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and deliver solutions. Perform performance tuning, optimize distributed data processing tasks, and handle scalability issues. Provide troubleshooting and support for data pipeline failures and ensure high availability and reliability. Contribute to the setup and maintenance of CI/CD pipelines for automated deployment and testing of data workflows. Required Skills & Experience : Experience: Minimum of 6+ years of hands-on experience in data engineering or big data development roles, with a focus on designing and building data pipelines and processing systems. Technical Skills: Strong programming skills in Python with hands-on experience in Apache Spark (PySpark preferred). Proficient in writing and optimizing complex SQL queries for data extraction, transformation, and loading. Hands-on experience with Apache Airflow for orchestration of data workflows and pipeline management. In-depth understanding and practical experience with AWS services : Data Storage & Processing: S3, Glue, EMR, Athena Compute & Execution: Lambda, Step Functions Databases: RDS, DynamoDB Monitoring: CloudWatch Experience with distributed data processing, parallel computing, and performance tuning techniques. Strong analytical and problem-solving skills to troubleshoot and optimize data workflows and pipelines. Familiarity with CI/CD pipelines and DevOps practices for continuous integration and automated deployments is a plus. Preferred Qualifications: Familiarity with other cloud platforms (Azure, Google Cloud) and services related to data engineering. Experience in handling unstructured and semi-structured data and working with data lakes. Knowledge of containerization technologies such as Docker or orchestration systems like Kubernetes . Experience with NoSQL databases or data warehouses like Redshift or BigQuery is a plus. Qualifications: Education: Bachelor's or Master's degree in Computer Science, Data Engineering, or a related field. Experience: Minimum of 6+ years in a data engineering role with strong expertise in AWS and big data processing frameworks.
Posted 3 weeks ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
23962 Jobs | Dublin
Wipro
12595 Jobs | Bengaluru
EY
8867 Jobs | London
Accenture in India
7480 Jobs | Dublin 2
Uplers
7207 Jobs | Ahmedabad
Amazon
6884 Jobs | Seattle,WA
IBM
6543 Jobs | Armonk
Oracle
6473 Jobs | Redwood City
Muthoot FinCorp (MFL)
6161 Jobs | New Delhi
Capgemini
5121 Jobs | Paris,France