Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
5.0 - 10.0 years
10 - 17 Lacs
Pune, Chennai, Bengaluru
Hybrid
Job Opportunity from Hexaware Technologies ! We are hiring Azure Databricks consultant with immediate joiner required, interested please reply to manojkumark2@hexaware.com with below details Shortlisted candidates will get interview call on Saturday 7th June Total IT Exp: Exp in Azure Databricks: Exp in Pyspark: Exp in Synapse: CCTC & ECTC: Immediate joiner: Yes /No Location:
Posted 3 weeks ago
10.0 - 14.0 years
10 - 14 Lacs
Hyderabad / Secunderabad, Telangana, Telangana, India
On-site
What you will do In this vital role you will for designing, developing, and maintaining software solutions for Research scientists. Additionally, it involves automating operations, monitoring system health, and responding to incidents to minimize downtime. You will join a multi-functional team of scientists and software professionals that enables technology and data capabilities to evaluate drug candidates and assess their abilities to affect the biology of drug targets. This team implements scientific software platforms that enable the capture, analysis, storage, and report of in vitro assays and in vivo / pre-clinical studies as well as those that manage compound inventories / biological sample banks. The ideal candidate possesses experience in the pharmaceutical or biotech industry, strong technical skills, and full stack software engineering experience (spanning SQL, back-end, front-end web technologies, automated testing). Roles & Responsibilities: Design, develop, and implement applications and modules, including custom reports, interfaces, and enhancements Analyze and understand the functional and technical requirements of applications, solutions and systems and translate them into software architecture and design specifications Develop and implement unit tests, integration tests, and other testing strategies to ensure the quality of the software Identify and resolve software bugs and performance issues Work closely with multi-functional teams, including product management, design, and QA, to deliver high-quality software on time Maintain detailed documentation of software designs, code, and development processes Customize modules to meet specific business requirements Work on integrating with other systems and platforms to ensure seamless data flow and functionality Provide ongoing support and maintenance for applications, ensuring that they operate smoothly and efficiently Possesses strong rapid prototyping skills and can quickly translate concepts into working code Contribute to both front-end and back-end development using cloud technology Develop innovative solution using generative AI technologies Create and maintain documentation on software architecture, design, deployment, disaster recovery, and operations Identify and resolve technical challenges effectively Stay updated with the latest trends and advancements Work closely with product team, business team including scientists, and other collaborators Roles & Responsibilities: Project & Portfolio Delivery Lead the execution of initiatives across the data platforms portfolio, ensuring projects are delivered on time, within scope, and to expected quality standards. Coordinate cross-functional teams (Business, engineering, architecture, operations, governance) to deliver tools, technologies and platforms. Lead the initiatives for evaluating latest market technologies in the area of data Engineering & Management & Governance Financial Management Own and manage project and portfolio budgets, including tracking actuals vs forecasts, accruals, and reporting on financial performance to stakeholders. Partner with Finance, Procurement, and Vendor Management teams to support contract reviews, Platform costs. Proactively monitor financial risks and ensure alignment of project spend with approved business cases and funding models. Prepare financial summaries and variance reports for leadership and program steering committees. Planning & Governance Maintain integrated plans and roadmaps across projects within the data platforms portfolio. Run governance forums, manage stakeholder expectations, and ensure project artifacts, status reports, and RAID logs are consistently maintained. Stakeholder & Communication Management Serve as the central point of contact between technical teams, business stakeholders, and vendors. Lead project steering committee meetings and provide clear and concise updates to senior leadership. Agile & Hybrid Delivery Apply agile, SAFe or hybrid delivery methods based on project needs, support backlog grooming, sprint planning, and release planning. Promote continuous improvement in delivery through retrospectives and feedback loops. Must Have skills: Demonstrated experience managing project financials (budgeting, forecasting, variance analysis, cost optimization) Experience working in large, complex enterprise environments with cross-functional stakeholders Familiarity with modern data platforms such as Azure Data Lake, Databricks, Snowflake, Synapse, Kafka, Delta Lake, etc. Strong understanding of data management lifecycle, data architecture, and platform components (ingestion, processing, governance, access) Excellent interpersonal, presentation, and negotiation skills PMP, PMI-ACP, SAFe, or equivalent certifications are a plus Basic Qualifications and Experience: Masters degree with 8-10+ years of experience in Business, Engineering, IT or related field OR Bachelors degree with 10-14+ years of experience in Business, Engineering, IT or related field OR Diploma with 14+ years of experience in Business, Engineering, IT or related field Good-to-Have Skills: Strong understanding of Cloud Infrastructure, Data & Analytics tools like Databricks, Informatica, PowerBI, Tableau and Data Governance technologies Experience with cloud (e.g. AWS) and on-premises compute infrastructure Experience with Databricks platform. Professional Certifications : Project Managerment Certifications Agile Certified Practitioner (preferred) AWS certification Soft Skills: Excellent interpersonal, presentation, and negotiation skills Strong analytical abilities to assess and improve data processes and solutions. Excellent verbal and written communication skills, with the ability to convey complex data concepts clearly to technical and non-technical stakeholders. Effective problem-solving skills to address data-related issues and implement scalable solutions. Ability to work effectively with global, virtual teams
Posted 4 weeks ago
5.0 - 10.0 years
18 - 33 Lacs
Kochi, Hyderabad, Bengaluru
Hybrid
Neudesic, an IBM Company is home to some very smart, talented and motivated people. People who want to work for an innovative company that values their skills and keeps their passions alive with new challenges and opportunities. We have created a culture of innovation that makes Neudesic not only an industry leader, but also a career destination for todays brightest technologists. You can see it in our year-over-year growth, made possible by satisfied employees dedicated to delivering the right solutions to our clients Must Have Skills: Prior experience in ETL, data pipelines, data flow techniques using Azure Data Services Working experience in Python, Scala, PySpark, Azure Data Factory, Azure Data Lake Gen2, Databricks, Azure Synapse and file formats like JSON & Parquet Experience in creating ADF Pipelines to source and process data sets. Experience in creating Databricks notebooks to cleanse, transform and enrich data sets. Good understanding about SQL, Databases, NO-SQL DBs, Data Warehouse, Hadoop and various data storage options on the cloud. Development experience in orchestration of pipelines Experience in deployment and monitoring techniques Working experience with Azure DevOps CI/CD pipelines to deploy Azure resources. Experience in handling operations/Integration with source repository Must have good knowledge on Datawarehouse concepts and Datawarehouse modelling Good to Have Skills: Familiarity with DevOps, Agile Scrum methodologies and CI/CD Domain-driven development exposure Analytical / problem solving skills Strong communication skills Good experience with unit, integration and UAT support Able to design and code reusable components and functions Should able to review design, code & provide review comments with justification Zeal to learn new tool/technologies and adoption Power BI and Data Catalog experience
Posted 1 month ago
7.0 - 12.0 years
0 - 2 Lacs
Pune, Ahmedabad, Gurugram
Work from Office
Urgent Hiring: Azure Data Engineer (Strong PySpark + SCD II/III Expert) Work Mode: Remote Client-Focused Interview on PySpark + SCD II/III Key Must-Haves: Very Strong hands-on PySpark coding Practical experience implementing Slowly Changing Dimensions (SCD) Type II and Type III Strong expertise in Azure Data Engineering (ADF, Databricks, Data Lake, Synapse) Proficiency in SQL and Python for scripting and transformation Strong understanding of data warehousing concepts and ETL pipelines Good to Have: Experience with Microsoft Fabric Familiarity with Power BI Domain knowledge in Finance, Procurement, and Human Capital Note: This role is highly technical. The client will focus interviews on PySpark coding and SCD Type II/III implementation . Only share profiles that are hands-on and experienced in these areas. Share strong, relevant profiles to: b.simrana@ekloudservices.com
Posted 1 month ago
8.0 - 10.0 years
10 - 12 Lacs
Hyderabad
Work from Office
Overview Data Analyst will be responsible to partner closely with business and S&T teams in preparing final analysis reports for the stakeholders enabling them to make important decisions based on various facts and trends and lead data requirement, source analysis, data analysis, data transformation and reconciliation activities. This role will be interacting with DG, DPM, EA, DE, EDF, PO and D &Ai teams for historical data requirement and sourcing the data for Mosaic AI program to scale solution to new markets. Responsibilities Lead data requirement, source analysis, data analysis, data transformation and reconciliation activities. Partners with FP&A Product Owner and associated business SMEs to understand & document business requirements and associated needs Performs the analysis of business data requirements and translates into a data design that satisfies local, sector and global requirements Using automated tools to extract data from primary and secondary sources. Using statistical tools to identify, analyse, and interpret patterns and trends in complex data sets could be helpful for the diagnosis and prediction. Working with engineers, and business teams to identify process improvement opportunities, propose system modifications. Proactively identifies impediments and looks for pragmatic and constructive solutions to mitigate risk. Be a champion for continuous improvement and drive efficiency. Preference will be given to candidate having functional understanding of financial concepts (P&L, Balance Sheet, Cash Flow, Operating Expense) and has experience modelling data & designing data flows Qualifications Bachelor of Technology from a reputed college Minimum 8-10 years of relevant work experience on data modelling / analytics, preferably Minimum 5-6year experience of navigating data in Azure Databricks, Synapse, Teradata or similar database technologies Expertise in Azure (Databricks, Data Factory, Date Lake Store Gen2) Proficient in SQL, Pyspark to analyse data for both development validation and operational support is critical Exposure to GenAI Good Communication & Presentation skill is must for this role.
Posted 1 month ago
6.0 - 11.0 years
10 - 18 Lacs
Noida
Remote
Job Title: Senior BI & Analytics Consultant Location: Remote (India-based, serving European clients) Type: Full-time Travel: Occasional client visits within Europe About Us We are a new, fast-scaling consultancy with a sharp focus on delivering high-quality Business Intelligence & Analytics solutions to mid-sized and enterprise customers across Europe. Our mission is simple: bring superior quality at a lower price point , enabled by smart delivery and modern GenAI technologies. We're building a next-generation BI service line from the ground up and we're looking for entrepreneurial consultants ready to shape our foundation. What Youll Do As a Senior BI & Analytics Consultant, youll lead client engagements from discovery through delivery. Your focus will be on shaping data strategies, designing robust architectures, and delivering scalable analytics solutions — all while mentoring future team members and leveraging GenAI to drive delivery efficiency. You’ll be expected to: Lead workshops to translate business needs into BI requirements and data models Design end-to-end data architectures using Azure Data Lake, Data Fabric, Synapse, Databricks, and other components Lead or support data acquisition, transformation, and modeling (SQL, Python, Spark) Build and optimize Power BI reports, including complex DAX development and semantic modeling Use tools like GitHub Copilot and Power BI Copilot to accelerate high-quality delivery Define delivery best practices and coach junior team members over time Act as a trusted advisor to clients and represent our values of quality, efficiency, and innovation Your Profile You are a seasoned BI consultant with a hands-on mindset and the ability to lead clients and projects independently. Must-Haves: 6+ years of experience in BI & analytics, preferably in consulting Strong expertise in Power BI (including DAX and data modeling) Proficiency in SQL and Python for data manipulation and transformation Solid knowledge of modern data platforms such as Azure Data Lake, Synapse, Fabric, and Databricks Ability to architect and deliver end-to-end solutions Comfortable working remotely and engaging with European clients in English Experience with or curiosity about GenAI tools like Microsoft Copilots, GitHub Copilot, etc. Strong communication and leadership skills Nice-to-Haves: Experience with dbt, Azure Data Factory, or similar pipeline/orchestration tools Familiarity with CI/CD practices for BI deployments (Git, DevOps) Exposure to enterprise data governance and security best practices Additional language skills (Dutch, German, French) are a bonus Why join Boundaryless? Enjoy the flexibility of working from anywhere Mediclaim policy to ensure the well-being of you and your family Referral bonus when help bring in great talent We actively promote and support your professional development A structured reward and recognition program that recognizes your contributions and achievements Job security and fair employment policies We’re committed to building a workplace that puts people first where your growth, well-being, and satisfaction matter just as much as business goals. Salary Offered: Best as per market standards.
Posted 1 month ago
10.0 - 16.0 years
25 - 27 Lacs
Chennai
Work from Office
We at Dexian India, are looking to hire a Cloud Data PM with over 10 years of hands-on experience in AWS/Azure, DWH, and ETL. The role is based in Chennai with a shift from 2.00pm to 11.00pm IST. Key qualifications we seek in candidates include: - Solid understanding of SQL and data modeling - Proficiency in DWH architecture, including EDW/DM concepts and Star/Snowflake schema - Experience in designing and building data pipelines on Azure Cloud stack - Familiarity with Azure Data Explorer, Data Factory, Data Bricks, Synapse Analytics, Azure Fabric, Azure Analysis Services, and Azure SQL Datawarehouse - Knowledge of Azure DevOps and CI/CD Pipelines - Previous experience managing scrum teams and working as a Scrum Master or Project Manager on at least 2 projects - Exposure to on-premise transactional database environments like Oracle, SQL Server, Snowflake, MySQL, and/or Postgres - Ability to lead enterprise data strategies, including data lake delivery - Proficiency in data visualization tools such as Power BI or Tableau, and statistical analysis using R or Python - Strong problem-solving skills with a track record of deriving business insights from large datasets - Excellent communication skills and the ability to provide strategic direction to technical and business teams - Prior experience in presales, RFP and RFI responses, and proposal writing is mandatory - Capability to explain complex data solutions clearly to senior management - Experience in implementing, managing, and supporting data warehouse projects or applications - Track record of leading full-cycle implementation projects related to Business Intelligence - Strong team and stakeholder management skills - Attention to detail, accuracy, and ability to meet tight deadlines - Knowledge of application development, APIs, Microservices, and Integration components Tools & Technology Experience Required: - Strong hands-on experience in SQL or PLSQL - Proficiency in Python - SSIS or Informatica (Mandatory one of the tools) - BI: Power BI, or Tableau (Mandatory one of the tools)
Posted 1 month ago
5.0 - 9.0 years
15 - 30 Lacs
Hyderabad
Hybrid
Hi! Greetings of the day!! We have openings for one of our product based company. Location : Hyderabad Notice Period: Only Immediate - 30 Days Work Mode - Hybrid Key Purpose Statement Core mission The core purpose of a Senior Data Engineer will play a key role in designing, building, and optimizing our data infrastructure and pipelines. This individual will leverage their deep expertise in Azure Synapse , Databricks cloud platforms, and Python programming to deliver high-quality data solutions. RESPONSIBILITIES Data Infrastructure and Pipeline Development: - Develop and maintain complex ETL/ELT pipelines using Databricks and Azure Synapse. - Optimize data pipelines for performance, scalability, and cost-efficiency. - Implement best practices for data governance, quality, and security. Cloud Platform Management: - Design and manage cloud-based data infrastructure on platforms such as Azure - Utilize cloud-native tools and services to enhance data processing and storage capabilities. - understanding and designing CI/CD pipelines for data engineering projects. Programming: - Develop and maintain high-quality, reusable Code on Databricks, and Synapse environment for data processing and automation. - Collaborate with data scientists and analysts to design solutions into data workflows. - Conduct code reviews and mentor junior engineers in Python , PySpark & SQL environments best practices. If interested, please share resume to aparna.ch@v3staffing.in
Posted 1 month ago
4.0 - 9.0 years
6 - 14 Lacs
Hyderabad
Remote
Job description Job Location : Hyderabad / Bangalore / Chennai / Kolkata / Noida/ Gurgaon / Pune / Indore / Mumbai Preferred: Hyderabad At least 4+ years of relevant hands on development experience as Azure Data Engineering role Proficient in Azure technologies like ADB, ADF, SQL(capability of writing complex SQL queries), ADB, PySpark, Python, Synapse, Delta Tables, Unity Catalog Hands on in Python, PySpark or Spark SQL Hands on in Azure Analytics and DevOps Taking part in Proof of Concepts (POCs) and pilot solutions preparation Ability to conduct data profiling, cataloguing, and mapping for technical design and construction of technical data flow Experience in business processing mapping of data and analytics solutions
Posted 1 month ago
9.0 - 14.0 years
30 - 45 Lacs
Bengaluru
Work from Office
Design, deploy, and optimize Azure-based data pipelines and architectures. Ensure scalability, data integrity, and CI/CD automation. Collaborate with analytics teams and lead data engineering initiatives across hybrid data platforms Required Candidate profile Bachelor’s in CS/IT with 7–12 years of experience in Azure data engineering. Strong in ADF, Synapse, Databricks, and CI/CD. Able to mentor junior engineers, optimize large-scale data systems
Posted 1 month ago
4.0 - 6.0 years
4 - 7 Lacs
Pune
Work from Office
Job Summary We are looking for a Data Quality Engineer who will safeguard the integrity of our cloud-native data assets. You will design and execute automated and manual data-quality checks across structured and semi-structured sources on Azure and GCP, validating that our data pipelines deliver accurate, complete, and consistent datasets for analytics, reporting, and AI initiatives. Key Responsibilities Define, build, and maintain data-quality frameworks that measure accuracy, completeness, timeliness, consistency, and validity of data ingested through ETL/ELT pipelines. Develop automated tests using SQL, Python, or similar tools; supplement with targeted manual validation where required. Collaborate with data engineers to embed data-quality gates into CI/CD pipelines on Azure Data Factory / Synapse / Fabric and GCP Dataflow / Cloud Composer. Profile new data sources (structured and semi-structuredJSON, Parquet, Avro) to establish baselines, detect anomalies, and recommend cleansing or transformation rules. Monitor data-quality KPIs and publish dashboards/alerts that surface issues to stakeholders in near-real time. Conduct root-cause analysis for data-quality defects, propose remediation strategies, and track resolution to closure. Maintain comprehensive documentation of test cases, data-quality rules, lineage, and issue logs for audit and governance purposes. Partner with data governance, security, and compliance teams to ensure adherence to regulatory requirements Must-Have Skills 4-6 years of experience in data quality, data testing, or data engineering roles within cloud environments. Hands-on expertise with at least one major cloud data stackAzure (Data Factory, Synapse, Databricks/Fabric) or GCP (BigQuery, Dataflow, Cloud Composer). Strong SQL skills and proficiency in a scripting language such as Python for building automated validation routines. Solid understanding of data-modeling concepts (dimensional, 3NF, data vault) and how they impact data-quality rules. Experience testing semi-structured data formats (JSON, XML, Avro, Parquet) and streaming/near-real-time pipelines. Excellent analytical and communication skills; able to translate complex data issues into clear, actionable insights for technical and business stakeholders. Nice-to-Have Skills Familiarity with BI/reporting tools (Power BI, Looker, Tableau) for surfacing data-quality metrics. Preferred Certifications Google Professional Data Engineer or Associate Cloud Engineer (GCP track) - OR - Microsoft Certified: Azure Data Engineer Associate Education Bachelors or Masters degree in Computer Science, Information Systems, Engineering, Mathematics, or a related field. Comparable professional experience will also be considered. Why Join Us? You will be the guardian of our datas trustworthiness, enabling decision-makers to rely on insights with confidence. If you are passionate about building automated, scalable data-quality solutions in a modern cloud environment, we’d love to meet you.
Posted 1 month ago
6.0 - 8.0 years
8 - 10 Lacs
Pune
Work from Office
Job Summary We are looking for a seasoned Data Modeler / Data Analyst to design and implement scalable, reusable logical and physical data models on Google Cloud Platformprimarily BigQuery. You will partner closely with data engineers, analytics teams, and business stakeholders to translate complex business requirements into performant data models that power reporting, self-service analytics, and advanced data science workloads. Key Responsibilities Gather and analyze business requirements to translate them into conceptual, logical, and physical data models on GCP (BigQuery, Cloud SQL, Cloud Spanner, etc.). Design star/snowflake schemas, data vaults, and other modeling patterns that balance performance, flexibility, and cost. Implement partitioning, clustering, and materialized views in BigQuery to optimize query performance and cost efficiency. Establish and maintain data modelling standards, naming conventions, and metadata documentation to ensure consistency across analytic and reporting layers. Collaborate with data engineers to define ETL/ELT pipelines and ensure data models align with ingestion and transformation strategies (Dataflow, Cloud Composer, Dataproc, dbt). Validate data quality and lineage; work with BI developers and analysts to troubleshoot performance issues or data anomalies. Conduct impact assessments for schema changes and guide version-control processes for data models. Mentor junior analysts/engineers on data modeling best practices and participate in code/design reviews. Contribute to capacity planning and cost-optimization recommendations for BigQuery datasets and reservations. Must-Have Skills 6-8 years of hands-on experience in data modeling, data warehousing, or database design, including at least 2 years on GCP BigQuery. Proficiency in dimensional modeling, 3NF, and modern patterns such as data vault. Expert SQL skills with demonstrable ability to optimize complex analytical queries on BigQuery (partitioning, clustering, sharding strategies). Strong understanding of ETL/ELT concepts and experience working with tools such as Dataflow, Cloud Composer, or dbt. Familiarity with BI/reporting tools (Looker, Tableau, Power BI, or similar) and how model design impacts dashboard performance. Experience with data governance practices—data cataloging, lineage, and metadata management (e.g., Data Catalog). Excellent communication skills to translate technical concepts into business-friendly language and collaborate across functions. Good to Have Experience of working on Azure Cloud (Fabric, Synapse, Delta Lake) Education Bachelor’s or Master’s degree in Computer Science, Information Systems, Engineering, Statistics, or a related field. Equivalent experience will be considered.
Posted 1 month ago
2.0 - 7.0 years
20 - 30 Lacs
Pune
Work from Office
Work mode – Currently this is remote but it’s not permanent WFH , once business ask the candidate to come to office, they must relocate. Mandatory:- DE , Azure , synapse , SQL python , Pyspark, ETL,Fabric, • Exp.in Python for scripting or data tasks. Required Candidate profile • Hands-on exp in SQL& relational databases (SQL Server,PostgreSQL). • data warehousing concepts (ETL, • Hands-on exp in Azure data integration tools like DF, Synapse, Data Lake and Blob Storage.
Posted 1 month ago
10.0 - 15.0 years
12 - 17 Lacs
Indore, Hyderabad, Ahmedabad
Work from Office
Job Title: Technical Architect / Solution Architect / Data Architect (Data Analytics) ?? Notice Period: Immediate to 15 Days ?? Experience: 9+ Years Job Summary: We are looking for a highly technical and experienced Data Architect / Solution Architect / Technical Architect with expertise in Data Analytics. The candidate should have strong hands-on experience in solutioning, architecture, and cloud technologies to drive data-driven decisions. Key Responsibilities: ? Design, develop, and implement end-to-end data architecture solutions. ? Provide technical leadership in Azure, Databricks, Snowflake, and Microsoft Fabric. ? Architect scalable, secure, and high-performing data solutions. ? Work on data strategy, governance, and optimization. ? Implement and optimize Power BI dashboards and SQL-based analytics. ? Collaborate with cross-functional teams to deliver robust data solutions. Primary Skills Required: ? Data Architecture & Solutioning ? Azure Cloud (Data Services, Storage, Synapse, etc.) ? Databricks & Snowflake (Data Engineering & Warehousing) ? Power BI (Visualization & Reporting) ? Microsoft Fabric (Data & AI Integration) ? SQL (Advanced Querying & Optimization) ?? Looking for immediate to 15-day joiners!
Posted 1 month ago
9 - 14 years
30 - 37 Lacs
Bengaluru
Work from Office
Hi, Greetings from Decision Minds! Mandatory Skill: Lead Azure Data Engg (Data Lake, Synapse, Data Bricks, Spark) Databricks Certified Data Engg Exp: 9 - 14 years B'LORE, HYD, CHE If interested, please share your profile to barsas@decisionminds.com
Posted 1 month ago
12 - 15 years
15 - 17 Lacs
Bengaluru
Work from Office
About The Role Overview Technology for today and tomorrow The Boeing India Engineering & Technology Center (BIETC) is a 5500+ engineering workforce that contributes to global aerospace growth. Our engineers deliver cutting-edge R&D, innovation, and high-quality engineering work in global markets, and leverage new-age technologies such as AI/ML, IIoT, Cloud, Model-Based Engineering, and Additive Manufacturing, shaping the future of aerospace. People-driven culture At Boeing, we believe creativity and innovation thrives when every employee is trusted, empowered, and has the flexibility to choose, grow, learn, and explore. We offer variable arrangements depending upon business and customer needs, and professional pursuits that offer greater flexibility in the way our people work. We also believe that collaboration, frequent team engagements, and face-to-face meetings bring together different perspectives and thoughts enabling every voice to be heard and every perspective to be respected. No matter where or how our teammates work, we are committed to positively shaping peoples careers and being thoughtful about employee wellbeing. Boeing India Software Engineering team is currently looking for one Lead Software Engineer Developer to join their team in Bengaluru, KA. As a ETL Developer , you will be part of the Application Solutions team, which develops software applications and Digital products that create direct value to its customers. We provide re-vamped work environments focused on delivering data-driven solutions at a rapidly increased pace over traditional development. Be a part of our passionate and motivated team who are excited to use the latest in software technologies for modern web and mobile application development. Through our products we deliver innovative solutions to our global customer base at an accelerated pace. Position Responsibilities: Perform data mining and collection procedures. Ensure data quality and integrity, Interpret and analyze data problems. Visualize data and create reports. Experiments with new models and techniques Determines how data can be used to achieve customer / user goals. Designs data modeling processes Create algorithms and predictive models to for analysis. Enables development of prediction engines, pattern detection analysis, and optimization algorithms, etc. Develops guidance for analytics-based wireframes. Organizes and conducts data assessments. Discovers insights from structured and unstructured data. Estimate user stories/features (story point estimation) and tasks in hours with the required level of accuracy and commit them as part of Sprint Planning. Contributes to the backlog grooming meetings by promptly asking relevant questions to ensure requirements achieve the right level of DOR. Raise any impediments/risks (technical/operational/personal) they come across and approaches Scrum Master/Technical Architect/PO accordingly to arrive at a solution. Update the status and the remaining efforts for their tasks on a daily basis. Ensures change requests are treated correctly and tracked in the system, impact analysis done, and risks/timelines are appropriately communicated. Hands-on experience in understanding aerospace domain specific data Must coordinate with data scientists in data preparation, exploration and making data ready. Must have clear understanding of defining data products and monetizing. Must have experience in building self-service capabilities to users. Build quality checks across the data lineage and responsible in designing and implementing different data patterns. Can influence different stakeholders for funding and building the vision of the product in terms of usage, productivity, and scalability of the solutions. Build impactful or outcome-based solutions/products. Basic Qualifications (Required Skills/Experience): Bachelors or masters degree as BASIC QUALIFICATION 12-15 years of experience as a data engineer. Expertise in SQL, Python, Knowledge of Java, Oracle, R, Data modeling, Power BI. Experience in understanding and interacting with multiple data formats. Ability to rapidly learn and understand software from source code. Expertise in understanding, analyzing & optimizing large, complicated SQL statements Strong knowledge and experience in SQL Server, database design and ETL queries. Develop software models to simulate real world problems to help operational leaders understand on which variables to focus. Candidate should have proficiency to streamline and optimize databases for efficient and consistent data consumption. Strong understanding of Datawarehouse concepts, data lake, data mesh Familiar with ETL tools and Data ingestion patterns Hands on experience in building data pipelines using GCP. Hands on experience in writing complex SQL (No- SQL is a big plus) Hands on experience with data pipeline orchestration tools such as Airflow/GCP Composer Hands on experience on Data Modelling Experience in leading teams with diversity Experience in performance tuning of large datawarehouse/datalakes. Exposure to prompt engineering, LLMs, and vector DB. Python, SQL and Pyspark Spark Ecosystem (Spark Core, Spark Streaming, Spark SQL) / Databricks Azure (ADF, ADB, Logic Apps, Azure SQL database, Azure Key Vaults, ADLS, Synapse) Preferred Qualifications [Required Skills/Experience] PubSUB, Terraform Deep Learning - Tensor flow Time series, BI/Visualization Tools - Power BI and Tablaeu, Languages - R/Phython Deep Learning - Tensor flow Machine Learning NLP Typical Education & Experience Education/experience typically acquired through advanced education (e.g. Bachelor) and typically 12 to 15 years' related work experience or an equivalent combination of education and experience (e.g. Master+11 years of related work experience etc.) Relocation This position does offer relocation within INDIA. Export Control Requirements This is not an Export Control position. Education Bachelor's Degree or Equivalent Required Relocation This position offers relocation based on candidate eligibility. Visa Sponsorship Employer will not sponsor applicants for employment visa status. Shift Not a Shift Worker (India)
Posted 1 month ago
10 - 18 years
35 - 55 Lacs
Hyderabad, Bengaluru, Mumbai (All Areas)
Hybrid
Warm Greetings from SP Staffing Services Private Limited!! We have an urgent opening with our CMMI Level 5 client for the below position. Please send your update profile if you are interested. Relevant Experience: 8 Yrs - 18 Yrs Location- Pan India Job Description : - Experience in Synapase with pyspark Knowledge of Big Data pipelinesData Engineering Working Knowledge on MSBI stack on Azure Working Knowledge on Azure Data factory Azure Data Lake and Azure Data lake storage Handson in Visualization like PowerBI Implement endend data pipelines using cosmosAzure Data factory Should have good analytical thinking and Problem solving Good communication and coordination skills Able to work as Individual contributor Requirement Analysis CreateMaintain and Enhance Big Data Pipeline Daily status reporting interacting with Leads Version controlADOGIT CICD Marketing Campaign experiences Data Platform Product telemetry Analytical thinking Data Validation of the new streams Data quality check of the new streams Monitoring of data pipeline created in Azure Data factory updating the Tech spec and wiki page for each implementation of pipeline Updating ADO on daily basis If interested please forward your updated resume to sankarspstaffings@gmail.com / Sankar@spstaffing.in With Regards, Sankar G Sr. Executive - IT Recruitment
Posted 1 month ago
10 - 20 years
35 - 55 Lacs
Hyderabad, Bengaluru, Mumbai (All Areas)
Hybrid
Warm Greetings from SP Staffing Services Private Limited!! We have an urgent opening with our CMMI Level 5 client for the below position. Please send your update profile if you are interested. Relevant Experience: 8 Yrs - 18 Yrs Location- Pan India Job Description : - Mandatory Skill: Azure ADB with Azure Data Lake Lead the architecture design and implementation of advanced analytics solutions using Azure Databricks Fabric The ideal candidate will have a deep understanding of big data technologies data engineering and cloud computing with a strong focus on Azure Databricks along with Strong SQL Work closely with business stakeholders and other IT teams to understand requirements and deliver effective solutions Oversee the endtoend implementation of data solutions ensuring alignment with business requirements and best practices Lead the development of data pipelines and ETL processes using Azure Databricks PySpark and other relevant tools Integrate Azure Databricks with other Azure services eg Azure Data Lake Azure Synapse Azure Data Factory and onpremise systems Provide technical leadership and mentorship to the data engineering team fostering a culture of continuous learning and improvement Ensure proper documentation of architecture processes and data flows while ensuring compliance with security and governance standards Ensure best practices are followed in terms of code quality data security and scalability Stay updated with the latest developments in Databricks and associated technologies to drive innovation Essential Skills Strong experience with Azure Databricks including cluster management notebook development and Delta Lake Proficiency in big data technologies eg Hadoop Spark and data processing frameworks eg PySpark Deep understanding of Azure services like Azure Data Lake Azure Synapse and Azure Data Factory Experience with ETLELT processes data warehousing and building data lakes Strong SQL skills and familiarity with NoSQL databases Experience with CICD pipelines and version control systems like Git Knowledge of cloud security best practices Soft Skills Excellent communication skills with the ability to explain complex technical concepts to nontechnical stakeholders Strong problemsolving skills and a proactive approach to identifying and resolving issues Leadership skills with the ability to manage and mentor a team of data engineers Experience Demonstrated expertise of 8 years in developing data ingestion and transformation pipelines using DatabricksSynapse notebooks and Azure Data Factory Solid understanding and handson experience with Delta tables Delta Lake and Azure Data Lake Storage Gen2 Experience in efficiently using Auto Loader and Delta Live tables for seamless data ingestion and transformation Proficiency in building and optimizing query layers using Databricks SQL Demonstrated experience integrating Databricks with Azure Synapse ADLS Gen2 and Power BI for endtoend analytics solutions Prior experience in developing optimizing and deploying Power BI reports Familiarity with modern CICD practices especially in the context of Databricks and cloudnative solutions If interested please forward your updated resume to sankarspstaffings@gmail.com / Sankar@spstaffing.in With Regards, Sankar G Sr. Executive - IT Recruitment
Posted 1 month ago
8 - 13 years
15 - 30 Lacs
Bengaluru
Work from Office
Design, develop, and maintain scalable ETL pipelines, data lakes, and hosting solutions using Azure tools. Ensure data quality, performance optimization, and compliance across hybrid and cloud environments. Required Candidate profile Data engineer with experience in Azure data services, ETL workflows, scripting, and data modeling. Strong collaboration with analytics teams and hands-on pipeline deployment using best practices
Posted 1 month ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
22558 Jobs | Dublin
Wipro
12294 Jobs | Bengaluru
EY
8435 Jobs | London
Accenture in India
7026 Jobs | Dublin 2
Uplers
6784 Jobs | Ahmedabad
Amazon
6588 Jobs | Seattle,WA
IBM
6430 Jobs | Armonk
Oracle
6230 Jobs | Redwood City
Virtusa
4470 Jobs | Southborough
Capgemini
4309 Jobs | Paris,France