Position- Data EngineerExperience- 3+ yearsLocation : Trivandrum, HybridSalary : Upto 8 LPA
Job Summary
We are seeking a highly motivated and skilled Data Engineer with 3+ years of experience tojoin our growing data team. In this role, you will be instrumental in designing, building, andmaintaining robust, scalable, and efficient data pipelines and infrastructure. You will work closelywith data scientists, analysts, and other engineering teams to ensure data availability, quality,and accessibility for various analytical and machine learning initiatives.
Key Responsibilities
○ Design, develop, and optimize scalable ETL/ELT pipelines to ingest, transform,
and load data from diverse sources into data warehouses/lakes.○ Implement data models and schemas that support analytical and reportingrequirements.○ Build and maintain robust data APIs for data consumption by various applicationsand services.
○ Contribute to the architecture and evolution of our data platform, leveraging cloud
services (AWS, Azure, GCP) or on-premise solutions.○ Ensure data security, privacy, and compliance with relevant regulations.○ Monitor data pipelines for performance, reliability, and data quality, implementingalerting and anomaly detection.
- Collaboration & Optimization:
○ Collaborate with data scientists, business analysts, and product managers to
understand data requirements and translate them into technical solutions.○ Optimize existing data processes for efficiency, cost-effectiveness, andperformance.○ Participate in code reviews, contribute to documentation, and uphold bestpractices in data engineering.
- Troubleshooting & Support:
○ Diagnose and resolve data-related issues, ensuring minimal disruption to dataconsumers.○ Provide support and expertise to teams consuming data from the data platform.
Required Qualifications
- Bachelor's degree in Computer Science, Engineering, or a related quantitative field.
- 3+ years of hands-on experience as a Data Engineer or in a similar role.
- Strong proficiency in at least one programming language commonly used for data
engineering (e.g., Python, Java, Scala).
- Extensive experience with SQL and relational databases (e.g., PostgreSQL, MySQL,
SQL Server).
- Proven experience with ETL/ELT tools and concepts.
- Experience with data warehousing concepts and technologies (e.g., Snowflake,
Redshift, BigQuery, Azure Synapse, Data Bricks).
- Familiarity with cloud platforms (AWS, Azure, or GCP) and their data services (e.g.,
S3, EC2, Lambda, Glue, Data Factory, Blob Storage, BigQuery, Dataflow).
- Understanding of data modeling techniques (e.g., dimensional modeling, Kimball,
Inmon).
- Experience with version control systems (e.g., Git).
- Excellent problem-solving, analytical, and communication skills.
Preferred Qualifications
- Master's degree in a relevant field.
- Experience with Apache Spark (PySpark, Scala Spark) or other big data processing
frameworks.
- Familiarity with NoSQL databases (e.g., MongoDB, Cassandra).
- Experience with data streaming technologies (e.g., Kafka, Kinesis).
- Knowledge of containerization technologies (e.g., Docker, Kubernetes).
- Experience with workflow orchestration tools (e.g., Apache Airflow, Azure Data
Factory, AWS Step Functions).
- Understanding of DevOps principles as applied to data pipelines.
- Prior experience in Telecom is a plus.
Skills: data streaming technologies (kafka, kinesis),azure,data modeling,apache spark,workflow orchestration tools (apache airflow, azure data factory, aws step functions),pipelines,apache,data engineering,kubernetes,cloud,programming languages (python, java, scala),docker,data apis,data warehousing,aws,version control systems (git),python,,cloud services (aws, azure, gcp),sql,nosql databases (mongodb, cassandra),etl/elt pipelines