Home
Jobs

3662 Databricks Jobs - Page 3

Filter Interviews
Min: 0 years
Max: 25 years
Min: ₹0
Max: ₹10000000
Setup a job Alert
Filter
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

0 years

0 Lacs

India

On-site

Linkedin logo

Shape the Future of AI At Labelbox, we're building the critical infrastructure that powers breakthrough AI models at leading research labs and enterprises. Since 2018, we've been pioneering data-centric approaches that are fundamental to AI development, and our work becomes even more essential as AI capabilities expand exponentially. About Labelbox We're the only company offering three integrated solutions for frontier AI development: Enterprise Platform & Tools: Advanced annotation tools, workflow automation, and quality control systems that enable teams to produce high-quality training data at scale Frontier Data Labeling Service: Specialized data labeling through Aligner, leveraging subject matter experts for next-generation AI models Expert Marketplace: Connecting AI teams with highly skilled annotators and domain experts for flexible scaling Why Join Us High-Impact Environment: We operate like an early-stage startup, focusing on impact over process. You'll take on expanded responsibilities quickly, with career growth directly tied to your contributions. Technical Excellence: Work at the cutting edge of AI development, collaborating with industry leaders and shaping the future of artificial intelligence. Innovation at Speed: We celebrate those who take ownership, move fast, and deliver impact. Our environment rewards high agency and rapid execution. Continuous Growth: Every role requires continuous learning and evolution. You'll be surrounded by curious minds solving complex problems at the frontier of AI. Clear Ownership: You'll know exactly what you're responsible for and have the autonomy to execute. We empower people to drive results through clear ownership and metrics. Role Overview Labelbox is seeking Subject Matter Experts (SMEs) to support high-impact AI data projects across a range of specialized domains, including Math, STEM, Programming, and Internationalization (i18n). As an SME, you’ll bring deep domain knowledge to help shape data labeling workflows, ensure quality, and guide contributor success. This is a contract-based role, deployed per project depending on expertise needs. Employment Type: Project-based, paid hourly Your Impact Design project structures and labeling workflows tailored to domain-specific goals Develop clear contributor guidelines and quality assurance frameworks Define ideal contributor profiles and task acceptance criteria Collaborate with internal teams to ensure subject matter accuracy and relevance What You Bring Proven expertise in one or more relevant domains (e.g., Math, Programming, i18n, etc.) Experience designing or reviewing data workflows, educational content, or technical documentation Strong communication and organizational skills Bonus Points Prior experience with data labeling, ML/AI, or evaluation projects Alignerr Services at Labelbox As part of the Alignerr Services team, you'll lead implementation of customer projects and manage our elite network of AI experts who deliver high-quality human feedback crucial for AI advancement. Your team will oversee 250,000+ monthly hours of specialized work across RLHF, complex reasoning, and multimodal AI projects, resulting in quality improvements for Frontier AI Labs. You'll leverage our AI-powered talent acquisition system and exclusive access to 16M+ specialized professionals to rapidly build and deploy expert teams that help customers like Google and ElevenLabs achieve breakthrough AI capabilities through precisely aligned human data—directly contributing to the critical human element in advancing artificial intelligence. Life at Labelbox Location: Join our dedicated tech hubs in San Francisco or Wrocław, Poland Work Style: Hybrid model with 2 days per week in office, combining collaboration and flexibility Environment: Fast-paced and high-intensity, perfect for ambitious individuals who thrive on ownership and quick decision-making Growth: Career advancement opportunities directly tied to your impact Vision: Be part of building the foundation for humanity's most transformative technology Our Vision We believe data will remain crucial in achieving artificial general intelligence. As AI models become more sophisticated, the need for high-quality, specialized training data will only grow. Join us in developing new products and services that enable the next generation of AI breakthroughs. Labelbox is backed by leading investors including SoftBank, Andreessen Horowitz, B Capital, Gradient Ventures, Databricks Ventures, and Kleiner Perkins. Our customers include Fortune 500 enterprises and leading AI labs. Your Personal Data Privacy : Any personal information you provide Labelbox as a part of your application will be processed in accordance with Labelbox’s Job Applicant Privacy notice. Any emails from Labelbox team members will originate from a @labelbox.com email address. If you encounter anything that raises suspicions during your interactions, we encourage you to exercise caution and suspend or discontinue communications.

Posted 1 day ago

Apply

0 years

0 Lacs

India

On-site

Linkedin logo

Shape the Future of AI At Labelbox, we're building the critical infrastructure that powers breakthrough AI models at leading research labs and enterprises. Since 2018, we've been pioneering data-centric approaches that are fundamental to AI development, and our work becomes even more essential as AI capabilities expand exponentially. About Labelbox We're the only company offering three integrated solutions for frontier AI development: Enterprise Platform & Tools: Advanced annotation tools, workflow automation, and quality control systems that enable teams to produce high-quality training data at scale Frontier Data Labeling Service: Specialized data labeling through Aligner, leveraging subject matter experts for next-generation AI models Expert Marketplace: Connecting AI teams with highly skilled annotators and domain experts for flexible scaling Why Join Us High-Impact Environment: We operate like an early-stage startup, focusing on impact over process. You'll take on expanded responsibilities quickly, with career growth directly tied to your contributions. Technical Excellence: Work at the cutting edge of AI development, collaborating with industry leaders and shaping the future of artificial intelligence. Innovation at Speed: We celebrate those who take ownership, move fast, and deliver impact. Our environment rewards high agency and rapid execution. Continuous Growth: Every role requires continuous learning and evolution. You'll be surrounded by curious minds solving complex problems at the frontier of AI. Clear Ownership: You'll know exactly what you're responsible for and have the autonomy to execute. We empower people to drive results through clear ownership and metrics. Role Overview Labelbox is seeking Subject Matter Experts (SMEs) to support high-impact AI data projects across a range of specialized domains, including Math, STEM, Programming, and Internationalization (i18n). As an SME, you’ll bring deep domain knowledge to help shape data labeling workflows, ensure quality, and guide contributor success. This is a contract-based role, deployed per project depending on expertise needs. Employment Type: Project-based, paid hourly Your Impact Design project structures and labeling workflows tailored to domain-specific goals Develop clear contributor guidelines and quality assurance frameworks Define ideal contributor profiles and task acceptance criteria Collaborate with internal teams to ensure subject matter accuracy and relevance What You Bring Proven expertise in one or more relevant domains (e.g., Math, Programming, i18n, etc.) Experience designing or reviewing data workflows, educational content, or technical documentation Strong communication and organizational skills Bonus Points Prior experience with data labeling, ML/AI, or evaluation projects Alignerr Services at Labelbox As part of the Alignerr Services team, you'll lead implementation of customer projects and manage our elite network of AI experts who deliver high-quality human feedback crucial for AI advancement. Your team will oversee 250,000+ monthly hours of specialized work across RLHF, complex reasoning, and multimodal AI projects, resulting in quality improvements for Frontier AI Labs. You'll leverage our AI-powered talent acquisition system and exclusive access to 16M+ specialized professionals to rapidly build and deploy expert teams that help customers like Google and ElevenLabs achieve breakthrough AI capabilities through precisely aligned human data—directly contributing to the critical human element in advancing artificial intelligence. Life at Labelbox Location: Join our dedicated tech hubs in San Francisco or Wrocław, Poland Work Style: Hybrid model with 2 days per week in office, combining collaboration and flexibility Environment: Fast-paced and high-intensity, perfect for ambitious individuals who thrive on ownership and quick decision-making Growth: Career advancement opportunities directly tied to your impact Vision: Be part of building the foundation for humanity's most transformative technology Our Vision We believe data will remain crucial in achieving artificial general intelligence. As AI models become more sophisticated, the need for high-quality, specialized training data will only grow. Join us in developing new products and services that enable the next generation of AI breakthroughs. Labelbox is backed by leading investors including SoftBank, Andreessen Horowitz, B Capital, Gradient Ventures, Databricks Ventures, and Kleiner Perkins. Our customers include Fortune 500 enterprises and leading AI labs. Your Personal Data Privacy : Any personal information you provide Labelbox as a part of your application will be processed in accordance with Labelbox’s Job Applicant Privacy notice. Any emails from Labelbox team members will originate from a @labelbox.com email address. If you encounter anything that raises suspicions during your interactions, we encourage you to exercise caution and suspend or discontinue communications.

Posted 1 day ago

Apply

0 years

0 Lacs

India

On-site

Linkedin logo

Shape the Future of AI At Labelbox, we're building the critical infrastructure that powers breakthrough AI models at leading research labs and enterprises. Since 2018, we've been pioneering data-centric approaches that are fundamental to AI development, and our work becomes even more essential as AI capabilities expand exponentially. About Labelbox We're the only company offering three integrated solutions for frontier AI development: Enterprise Platform & Tools: Advanced annotation tools, workflow automation, and quality control systems that enable teams to produce high-quality training data at scale Frontier Data Labeling Service: Specialized data labeling through Aligner, leveraging subject matter experts for next-generation AI models Expert Marketplace: Connecting AI teams with highly skilled annotators and domain experts for flexible scaling Why Join Us High-Impact Environment: We operate like an early-stage startup, focusing on impact over process. You'll take on expanded responsibilities quickly, with career growth directly tied to your contributions. Technical Excellence: Work at the cutting edge of AI development, collaborating with industry leaders and shaping the future of artificial intelligence. Innovation at Speed: We celebrate those who take ownership, move fast, and deliver impact. Our environment rewards high agency and rapid execution. Continuous Growth: Every role requires continuous learning and evolution. You'll be surrounded by curious minds solving complex problems at the frontier of AI. Clear Ownership: You'll know exactly what you're responsible for and have the autonomy to execute. We empower people to drive results through clear ownership and metrics. Role Overview Labelbox is seeking Subject Matter Experts (SMEs) to support high-impact AI data projects across a range of specialized domains, including Math, STEM, Programming, and Internationalization (i18n). As an SME, you’ll bring deep domain knowledge to help shape data labeling workflows, ensure quality, and guide contributor success. This is a contract-based role, deployed per project depending on expertise needs. Employment Type: Project-based, paid hourly Your Impact Design project structures and labeling workflows tailored to domain-specific goals Develop clear contributor guidelines and quality assurance frameworks Define ideal contributor profiles and task acceptance criteria Collaborate with internal teams to ensure subject matter accuracy and relevance What You Bring Proven expertise in one or more relevant domains (e.g., Math, Programming, i18n, etc.) Experience designing or reviewing data workflows, educational content, or technical documentation Strong communication and organizational skills Bonus Points Prior experience with data labeling, ML/AI, or evaluation projects Alignerr Services at Labelbox As part of the Alignerr Services team, you'll lead implementation of customer projects and manage our elite network of AI experts who deliver high-quality human feedback crucial for AI advancement. Your team will oversee 250,000+ monthly hours of specialized work across RLHF, complex reasoning, and multimodal AI projects, resulting in quality improvements for Frontier AI Labs. You'll leverage our AI-powered talent acquisition system and exclusive access to 16M+ specialized professionals to rapidly build and deploy expert teams that help customers like Google and ElevenLabs achieve breakthrough AI capabilities through precisely aligned human data—directly contributing to the critical human element in advancing artificial intelligence. Life at Labelbox Location: Join our dedicated tech hubs in San Francisco or Wrocław, Poland Work Style: Hybrid model with 2 days per week in office, combining collaboration and flexibility Environment: Fast-paced and high-intensity, perfect for ambitious individuals who thrive on ownership and quick decision-making Growth: Career advancement opportunities directly tied to your impact Vision: Be part of building the foundation for humanity's most transformative technology Our Vision We believe data will remain crucial in achieving artificial general intelligence. As AI models become more sophisticated, the need for high-quality, specialized training data will only grow. Join us in developing new products and services that enable the next generation of AI breakthroughs. Labelbox is backed by leading investors including SoftBank, Andreessen Horowitz, B Capital, Gradient Ventures, Databricks Ventures, and Kleiner Perkins. Our customers include Fortune 500 enterprises and leading AI labs. Your Personal Data Privacy : Any personal information you provide Labelbox as a part of your application will be processed in accordance with Labelbox’s Job Applicant Privacy notice. Any emails from Labelbox team members will originate from a @labelbox.com email address. If you encounter anything that raises suspicions during your interactions, we encourage you to exercise caution and suspend or discontinue communications.

Posted 1 day ago

Apply

0 years

0 Lacs

India

On-site

Linkedin logo

Shape the Future of AI At Labelbox, we're building the critical infrastructure that powers breakthrough AI models at leading research labs and enterprises. Since 2018, we've been pioneering data-centric approaches that are fundamental to AI development, and our work becomes even more essential as AI capabilities expand exponentially. About Labelbox We're the only company offering three integrated solutions for frontier AI development: Enterprise Platform & Tools: Advanced annotation tools, workflow automation, and quality control systems that enable teams to produce high-quality training data at scale Frontier Data Labeling Service: Specialized data labeling through Aligner, leveraging subject matter experts for next-generation AI models Expert Marketplace: Connecting AI teams with highly skilled annotators and domain experts for flexible scaling Why Join Us High-Impact Environment: We operate like an early-stage startup, focusing on impact over process. You'll take on expanded responsibilities quickly, with career growth directly tied to your contributions. Technical Excellence: Work at the cutting edge of AI development, collaborating with industry leaders and shaping the future of artificial intelligence. Innovation at Speed: We celebrate those who take ownership, move fast, and deliver impact. Our environment rewards high agency and rapid execution. Continuous Growth: Every role requires continuous learning and evolution. You'll be surrounded by curious minds solving complex problems at the frontier of AI. Clear Ownership: You'll know exactly what you're responsible for and have the autonomy to execute. We empower people to drive results through clear ownership and metrics. Role Overview Labelbox is seeking Subject Matter Experts (SMEs) to support high-impact AI data projects across a range of specialized domains, including Math, STEM, Programming, and Internationalization (i18n). As an SME, you’ll bring deep domain knowledge to help shape data labeling workflows, ensure quality, and guide contributor success. This is a contract-based role, deployed per project depending on expertise needs. Employment Type: Project-based, paid hourly Your Impact Design project structures and labeling workflows tailored to domain-specific goals Develop clear contributor guidelines and quality assurance frameworks Define ideal contributor profiles and task acceptance criteria Collaborate with internal teams to ensure subject matter accuracy and relevance What You Bring Proven expertise in one or more relevant domains (e.g., Math, Programming, i18n, etc.) Experience designing or reviewing data workflows, educational content, or technical documentation Strong communication and organizational skills Bonus Points Prior experience with data labeling, ML/AI, or evaluation projects Alignerr Services at Labelbox As part of the Alignerr Services team, you'll lead implementation of customer projects and manage our elite network of AI experts who deliver high-quality human feedback crucial for AI advancement. Your team will oversee 250,000+ monthly hours of specialized work across RLHF, complex reasoning, and multimodal AI projects, resulting in quality improvements for Frontier AI Labs. You'll leverage our AI-powered talent acquisition system and exclusive access to 16M+ specialized professionals to rapidly build and deploy expert teams that help customers like Google and ElevenLabs achieve breakthrough AI capabilities through precisely aligned human data—directly contributing to the critical human element in advancing artificial intelligence. Life at Labelbox Location: Join our dedicated tech hubs in San Francisco or Wrocław, Poland Work Style: Hybrid model with 2 days per week in office, combining collaboration and flexibility Environment: Fast-paced and high-intensity, perfect for ambitious individuals who thrive on ownership and quick decision-making Growth: Career advancement opportunities directly tied to your impact Vision: Be part of building the foundation for humanity's most transformative technology Our Vision We believe data will remain crucial in achieving artificial general intelligence. As AI models become more sophisticated, the need for high-quality, specialized training data will only grow. Join us in developing new products and services that enable the next generation of AI breakthroughs. Labelbox is backed by leading investors including SoftBank, Andreessen Horowitz, B Capital, Gradient Ventures, Databricks Ventures, and Kleiner Perkins. Our customers include Fortune 500 enterprises and leading AI labs. Your Personal Data Privacy : Any personal information you provide Labelbox as a part of your application will be processed in accordance with Labelbox’s Job Applicant Privacy notice. Any emails from Labelbox team members will originate from a @labelbox.com email address. If you encounter anything that raises suspicions during your interactions, we encourage you to exercise caution and suspend or discontinue communications.

Posted 1 day ago

Apply

0 years

0 Lacs

India

On-site

Linkedin logo

Shape the Future of AI At Labelbox, we're building the critical infrastructure that powers breakthrough AI models at leading research labs and enterprises. Since 2018, we've been pioneering data-centric approaches that are fundamental to AI development, and our work becomes even more essential as AI capabilities expand exponentially. About Labelbox We're the only company offering three integrated solutions for frontier AI development: Enterprise Platform & Tools: Advanced annotation tools, workflow automation, and quality control systems that enable teams to produce high-quality training data at scale Frontier Data Labeling Service: Specialized data labeling through Aligner, leveraging subject matter experts for next-generation AI models Expert Marketplace: Connecting AI teams with highly skilled annotators and domain experts for flexible scaling Why Join Us High-Impact Environment: We operate like an early-stage startup, focusing on impact over process. You'll take on expanded responsibilities quickly, with career growth directly tied to your contributions. Technical Excellence: Work at the cutting edge of AI development, collaborating with industry leaders and shaping the future of artificial intelligence. Innovation at Speed: We celebrate those who take ownership, move fast, and deliver impact. Our environment rewards high agency and rapid execution. Continuous Growth: Every role requires continuous learning and evolution. You'll be surrounded by curious minds solving complex problems at the frontier of AI. Clear Ownership: You'll know exactly what you're responsible for and have the autonomy to execute. We empower people to drive results through clear ownership and metrics. Role Overview Labelbox is seeking Subject Matter Experts (SMEs) to support high-impact AI data projects across a range of specialized domains, including Math, STEM, Programming, and Internationalization (i18n). As an SME, you’ll bring deep domain knowledge to help shape data labeling workflows, ensure quality, and guide contributor success. This is a contract-based role, deployed per project depending on expertise needs. Employment Type: Project-based, paid hourly Your Impact Design project structures and labeling workflows tailored to domain-specific goals Develop clear contributor guidelines and quality assurance frameworks Define ideal contributor profiles and task acceptance criteria Collaborate with internal teams to ensure subject matter accuracy and relevance What You Bring Proven expertise in one or more relevant domains (e.g., Math, Programming, i18n, etc.) Experience designing or reviewing data workflows, educational content, or technical documentation Strong communication and organizational skills Bonus Points Prior experience with data labeling, ML/AI, or evaluation projects Alignerr Services at Labelbox As part of the Alignerr Services team, you'll lead implementation of customer projects and manage our elite network of AI experts who deliver high-quality human feedback crucial for AI advancement. Your team will oversee 250,000+ monthly hours of specialized work across RLHF, complex reasoning, and multimodal AI projects, resulting in quality improvements for Frontier AI Labs. You'll leverage our AI-powered talent acquisition system and exclusive access to 16M+ specialized professionals to rapidly build and deploy expert teams that help customers like Google and ElevenLabs achieve breakthrough AI capabilities through precisely aligned human data—directly contributing to the critical human element in advancing artificial intelligence. Life at Labelbox Location: Join our dedicated tech hubs in San Francisco or Wrocław, Poland Work Style: Hybrid model with 2 days per week in office, combining collaboration and flexibility Environment: Fast-paced and high-intensity, perfect for ambitious individuals who thrive on ownership and quick decision-making Growth: Career advancement opportunities directly tied to your impact Vision: Be part of building the foundation for humanity's most transformative technology Our Vision We believe data will remain crucial in achieving artificial general intelligence. As AI models become more sophisticated, the need for high-quality, specialized training data will only grow. Join us in developing new products and services that enable the next generation of AI breakthroughs. Labelbox is backed by leading investors including SoftBank, Andreessen Horowitz, B Capital, Gradient Ventures, Databricks Ventures, and Kleiner Perkins. Our customers include Fortune 500 enterprises and leading AI labs. Your Personal Data Privacy : Any personal information you provide Labelbox as a part of your application will be processed in accordance with Labelbox’s Job Applicant Privacy notice. Any emails from Labelbox team members will originate from a @labelbox.com email address. If you encounter anything that raises suspicions during your interactions, we encourage you to exercise caution and suspend or discontinue communications.

Posted 1 day ago

Apply

3.0 years

0 Lacs

Indore, Madhya Pradesh, India

On-site

Linkedin logo

Project Role : Data Engineer Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems. Must have skills : Databricks Unified Data Analytics Platform Good to have skills : Business Agility Minimum 3 Year(s) Of Experience Is Required Educational Qualification : 15 years full time education Summary: As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to migrate and deploy data across various systems. You will collaborate with cross-functional teams to understand data requirements and deliver effective solutions that meet business needs. Additionally, you will monitor and optimize data workflows to enhance performance and reliability, ensuring that data is accessible and actionable for stakeholders. Roles & Responsibilities: - Need Databricks resource with Azure cloud experience - Expected to perform independently and become an SME. - Required active participation/contribution in team discussions. - Contribute in providing solutions to work related problems. - Collaborate with data architects and analysts to design scalable data solutions. - Implement best practices for data governance and security throughout the data lifecycle. Professional & Technical Skills: - Must To Have Skills: Proficiency in Databricks Unified Data Analytics Platform. - Good To Have Skills: Experience with Business Agility. - Strong understanding of data modeling and database design principles. - Experience with data integration tools and ETL processes. - Familiarity with cloud platforms and services related to data storage and processing. Additional Information: - The candidate should have minimum 3 years of experience in Databricks Unified Data Analytics Platform. - This position is based at our Pune office. - A 15 years full time education is required. 15 years full time education

Posted 1 day ago

Apply

5.0 years

0 Lacs

Navi Mumbai, Maharashtra, India

On-site

Linkedin logo

Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : Collibra Data Quality & Observability Good to have skills : Collibra Data Governance Minimum 5 Year(s) Of Experience Is Required Educational Qualification : 15 years full time education Summary: As an Application Developer, you will design, build, and configure applications to meet business process and application requirements. A typical day involves collaborating with various teams to understand their needs, developing solutions that align with business objectives, and ensuring that applications are functioning optimally. You will also engage in problem-solving activities, providing support and enhancements to existing applications while maintaining a focus on quality and efficiency. Key Responsibilities: Configure and implement Collibra Data Quality (CDQ) rules, workflows, dashboards, and data quality scoring metrics. Collaborate with data stewards, data owners, and business analysts to define data quality KPIs and thresholds. Develop data profiling and rule-based monitoring using CDQ's native rule engine or integrations (e.g., with Informatica, Talend, or BigQuery). Build and maintain Data Quality Dashboards and Issue Management workflows within Collibra. Integrate CDQ with Collibra Data Intelligence Cloud for end-to-end governance visibility. Drive root cause analysis and remediation plans for data quality issues. Support metadata and lineage enrichment to improve data traceability. Document standards, rule logic, and DQ policies in the Collibra Catalog. Conduct user training and promote data quality best practices across teams. Required Skills and Experience: 3+ years of experience in data quality, metadata management, or data governance. Hands-on experience with Collibra Data Quality & Observability (CDQ) platform. Knowledge of Collibra Data Intelligence Cloud including Catalog, Glossary, and Workflow Designer. Proficiency in SQL and understanding of data profiling techniques. Experience integrating CDQ with enterprise data sources (Snowflake, BigQuery, Databricks, etc.). Familiarity with data governance frameworks and data quality dimensions (accuracy, completeness, consistency, etc.). Excellent analytical, problem-solving, and communication skills. Additional Information: - The candidate should have minimum 7.5 years of experience in Collibra Data Quality & Observability. - This position is based in Mumbai. - A 15 years full time education is required. 15 years full time education

Posted 1 day ago

Apply

5.0 years

0 Lacs

Ahmedabad, Gujarat, India

On-site

Linkedin logo

Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : Microsoft Azure Analytics Services Good to have skills : NA Minimum 5 Year(s) Of Experience Is Required Educational Qualification : BE Summary: As an Application Lead for Packaged Application Development, you will be responsible for designing, building, and configuring applications using Microsoft Azure Analytics Services. Your typical day will involve leading the effort to deliver high-quality applications, acting as the primary point of contact for the project team, and ensuring timely delivery of project milestones. Roles & Responsibilities: - Lead the effort to design, build, and configure applications using Microsoft Azure Analytics Services. - Act as the primary point of contact for the project team, ensuring timely delivery of project milestones. - Collaborate with cross-functional teams to ensure the successful delivery of high-quality applications. - Provide technical guidance and mentorship to team members, ensuring adherence to best practices and standards. Professional & Technical Skills: - Must To Have Skills: Strong experience with Microsoft Azure Analytics Services. - Good To Have Skills: Experience with other Azure services such as Azure Data Factory, Azure Databricks, and Azure Synapse Analytics. - Experience in designing, building, and configuring applications using Microsoft Azure Analytics Services. - Must have databricks and pyspark Skills. - Strong understanding of data warehousing concepts and best practices. - Experience with ETL processes and tools such as SSIS or Azure Data Factory. - Experience with SQL and NoSQL databases. - Experience with Agile development methodologies. Additional Information: - The candidate should have a minimum of 5 years of experience in Microsoft Azure Analytics Services. - The ideal candidate will possess a strong educational background in computer science or a related field, along with a proven track record of delivering high-quality applications. - This position is based at our Bengaluru office. BE

Posted 1 day ago

Apply

7.0 years

0 Lacs

Ahmedabad, Gujarat, India

Remote

Linkedin logo

Job Title : Technical Project Manager Location : Ahmedabad, Gujarat Job Type : Full Time Department : Project management About Simform Simform is a premier digital engineering company specializing in Cloud, Data, AI/ML, and Experience Engineering to create seamless digital experiences and scalable products. Simform is a strong partner for Microsoft, AWS, Google Cloud, and Databricks. With a presence in 5+ countries, Simform primarily serves North America, the UK, and the Northern European market. Simform takes pride in being one of the most reputed employers in the region, having created a thriving work culture with a high work-life balance that gives a sense of freedom and opportunity to grow. Role Overview We are a digital product engineering company that partners with clients to create innovative, high-performing software solutions. We're looking for an experienced Project Manager who can lead from the front-bridging the gap between business and technology while ensuring smooth project execution.. Key Responsibilities Lead Agile/Scrum Projects - Manage the full project lifecycle, from planning to execution, ensuring timely delivery. Define & Refine Requirements - Gather and transform client needs into clear documentation, user stories, and deliverables. Sprint Planning & Execution - Assign tasks, run daily stand-ups, track progress, and mitigate risks. Client Communication - Be the point of contact for stakeholders, providing clear updates on scope, timelines, and deliverables. Technical Oversight - Leverage your expertise in web, mobile, and cloud technologies to guide development teams and ensure quality. Architectural & Design Guidance - Help define project architecture, offer technical mentorship, and ensure best practices. Quality Assurance - Oversee testing, release planning, and ensure glitch-free, high-performing applications. UX & Design Collaboration - Work closely with designers to maintain visual and UX design excellence, including responsive design principles. Required Skills & Qualifications 7+ years of technical experience, working with multiple technologies and understanding their core concepts. Proven leadership in managing and mentoring teams to successfully deliver projects. Expertise in Agile/SCRUM methodologies for medium-to-large scale applications. Strong interpersonal skills - mentoring, coaching, collaborating, and building high-performing teams. Problem-solving mindset - ability to zoom into details while keeping the big picture in focus. Solid grasp of design patterns and software architecture best practices. Hands-on experience managing projects from concept to production deployment. Why Join Us Young Team, Thriving Culture Flat-hierarchical, friendly, engineering-oriented, and growth-focused culture. Well-balanced learning and growth opportunities Free health insurance. Office facilities with a game zone, in-office kitchen with affordable lunch service, and free snacks. Sponsorship for certifications/events and library service. Flexible work timing, leaves for life events, WFH and hybrid options (ref:hirist.tech)

Posted 1 day ago

Apply

12.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Linkedin logo

Role Overview The Technical Architect - Databricks designs and implements scalable data architectures and solutions. The jobholder has expertise in Databricks Lakehouse, data modeling, and cloud integration, ensuring high performance, security, and reliability. Responsibilities Design and implement Databricks-based data architectures to meet business requirements. Develop and optimize data pipelines using PySpark, Scala, or SQL. Establish the Databricks Lakehouse architecture for batch and streaming data. Collaborate with cross-functional teams to integrate Databricks with cloud platforms (e.g., AWS, Azure, GCP). Ensure data security and compliance with best practices. Monitor and troubleshoot Databricks environments for performance and reliability. Stay updated on Databricks advancements and industry trends. Key Technical Skills & Responsibilities 12+ years of experience in data engineering using Databricks or Apache Spark-based platforms. Proven track record of building and optimizing ETL/ELT pipelines for batch and streaming data ingestion. Hands-on experience with Azure services such as Azure Data Factory, Azure Data Lake Storage, Azure Synapse Analytics, or Azure SQL Data Warehouse. Proficiency in programming languages such as Python, Scala, or SQL for data processing and transformation. Expertise in Spark (PySpark, Spark SQL, or Scala) and Databricks notebooks for large-scale data processing. Familiarity with Delta Lake, Delta Live Tables, and medallion architecture for data lakehouse implementations. Build and query deltalake storage solutions Process streaming data with Azure Databricks structured streaming Design Azure Databricks security and data protection solutions Flatten nested structures and explode arrays with spark Transfer data outside using sparkpools using pyspark connector Optimizing spark jobs Implementing best practices in spark/databricks Experience with orchestration tools like Azure Data Factory or Databricks Jobs for scheduling and automation. Knowledge of Git for source control and CI/CD integration for Databricks workflows, cost optimization, performance tuning. Familiarity with Unity Catalog, RBAC, or enterprise-level Databricks setups. Ability to create reusable components, templates, and documentation to standardize data engineering workflows. Solutioning and presales - Architecting frameworks, defining roadmaps, and engaging with stakeholders. Experience in defining data strategy, evaluating new tools/technologies, and driving adoption across the organization. Must have experience of working with streaming data sources and Kafka (preferred). Eligibility Criteria Bachelor’s degree in computer science, Information Technology, or related field Proven experience as a Databricks Architect or similar role Complete knowledge in Azure Databricks platform architecture Databricks certification (e.g., Certified Data Engineer, Associate Developer) Expertise in Python/Scala/ SQL/R Experience with cloud platforms like AWS, Azure, or GCP Strong understanding of data modeling and cloud integration Experience with cluster sizing and security implementation Excellent problem-solving and communication skills

Posted 1 day ago

Apply

7.0 years

0 Lacs

India

On-site

Linkedin logo

Pay Range: 18 - 30 LPA (INR) Required Skills & Experience 7+ years' experience with database engineering - building out and deploying pipelines, ideally working with financial data. 2+ years’ experience working with Azure applications (Azure Data Factory, Azure Batch, Azure SQL Server, Azure Data Warehouse, Azure Databricks etc.) and building out Azure pipelines. 2+ years; experience working with Azure Databricks. 1+ year of experience working in an Agile environment. Experience creating PowerBI reports. 2+ years' experience with Snowflake. 2+ years' experience with Python, PySpark & SQL. 2+ years' experience with infrastructure administration. Working knowledge of CI/CD. Working knowledge of building data integrity checks as part of delivery of applications. Nice to Have Skills & Experience Retail and/or e-commerce background; experience working for a multi-channel retailer. Job Description A large, North American, retail company is seeking an Azure Data Engineer in Bangalore, India. You will be joining our client’s FP&A Team, focusing on their Finance Data Hub (FDH). The FDH hosts all of our client’s sales data, inventory data, PNL data, financial data, etc. FDH houses Oracle RMS, Oracle EPM, and Oracle EBS and the Data Engineer will integrate data coming from these 3 systems, into the FDH. Owning data pipelines that gather data coming from multiple sources and consolidating that data for different use cases. Lead development and maintenance of data science and analytics processes, procedures, and policies. Lead identification, design, and implementation of integration, modelling, and orchestration of complex data. Act as subject matter expert for data engineering needs. Understanding of modern data platforms including data lakes and data warehouse, with good knowledge of the underlying architecture, preferably in Snowflake.

Posted 1 day ago

Apply

10.0 years

0 Lacs

India

Remote

Linkedin logo

Join phData, a dynamic and innovative leader in the modern data stack. We partner with major cloud data platforms like Snowflake, AWS, Azure, GCP, Fivetran, Pinecone, Glean and dbt to deliver cutting-edge services and solutions. We're committed to helping global enterprises overcome their toughest data challenges. phData is a remote-first global company with employees based in the United States, Latin America and India. We celebrate the culture of each of our team members and foster a community of technological curiosity, ownership and trust. Even though we're growing extremely fast, we maintain a casual, exciting work environment. We hire top performers and allow you the autonomy to deliver results. 5x Snowflake Partner of the Year (2020, 2021, 2022, 2023, 2024) Fivetran, dbt, Atlation, Matillion Partner of the Year #1 Partner in Snowflake Advanced Certifications 600+ Expert Cloud Certifications (Sigma, AWS, Azure, Dataiku, etc) Recognized as an award-winning workplace in US, India and LATAM Required Experience: 10+ years as a hands-on Solutions Architect and/or Data Engineer designing and implementing data solutions Team lead, and/or mentorship of other engineers Ability to develop end-to-end technical solutions into production — and to help ensure performance, security, scalability, and robust data integration. Programming expertise in Java, Python and/or Scala Core cloud data platforms including Snowflake, Spark, AWS, Azure, Databricks and GCP SQL and the ability to write, debug, and optimize SQL queries Client-facing written and verbal communication skills and experience Create and deliver detailed presentations Detailed solution documentation (e.g. including POCS and roadmaps, sequence diagrams, class hierarchies, logical system views, etc.) 4-year Bachelor's degree in Computer Science or a related field Prefer any of the following: Production experience in core data platforms: Snowflake, AWS, Azure, GCP, Hadoop, Databricks Cloud and Distributed Data Storage: S3, ADLS, HDFS, GCS, Kudu, ElasticSearch/Solr, Cassandra or other NoSQL storage systems Data integration technologies: Spark, Kafka, event/streaming, Streamsets, Matillion, Fivetran, NiFi, AWS Data Migration Services, Azure DataFactory, Informatica Intelligent Cloud Services (IICS), Google DataProc or other data integration technologies Multiple data sources (e.g. queues, relational databases, files, search, API) Complete software development lifecycle experience including design, documentation, implementation, testing, and deployment Automated data transformation and data curation: dbt, Spark, Spark streaming, automated pipelines Workflow Management and Orchestration: Airflow, AWS Managed Airflow, Luigi, NiFi Why phData? We Offer: Remote-First Workplace Medical Insurance for Self & Family Medical Insurance for Parents Term Life & Personal Accident Wellness Allowance Broadband Reimbursement Continuous learning and growth opportunities to enhance your skills and expertise Other benefits include paid certifications, professional development allowance, and bonuses for creating for company-approved content phData celebrates diversity and is committed to creating an inclusive environment for all employees. Our approach helps us to build a winning team that represents a variety of backgrounds, perspectives, and abilities. So, regardless of how your diversity expresses itself, you can find a home here at phData. We are proud to be an equal opportunity employer. We prohibit discrimination and harassment of any kind based on race, color, religion, national origin, sex (including pregnancy), sexual orientation, gender identity, gender expression, age, veteran status, genetic information, disability, or other applicable legally protected characteristics. If you would like to request an accommodation due to a disability, please contact us at People Operations.

Posted 1 day ago

Apply

3.0 years

0 Lacs

India

On-site

Linkedin logo

This is a hands-on Data Platform Engineering role with a strong focus on consultative data engagements across business and technical teams. Responsibilities Design and implement resilient data pipelines for batch and real-time processing Work closely with product managers, engineers, analysts, and data scientists to deliver scalable data platforms Provide guidance on architecture, infrastructure, and implementation best practices Collaborate with architects and developers to define data structures, pipelines, and orchestration strategies Ensure data privacy, processing, modeling, analytics, AI integration, and API connectivity Embrace Agile principles for project execution Develop frameworks to solve data challenges at scale Technical Skills Required 3+ years in data engineering with experience in lakehouse implementations (Databricks, Snowflake, or Synapse) Hands-on with Azure data stack (Databricks, Synapse, ADF) and supporting services (Key Vault, Storage, Firewall) Proficient in SQL, Python, and Spark Familiar with tools like JIRA, Git, Jenkins, TFS, Shell, PowerShell, Bitbucket Experience in Agile environments and familiarity with DBT and PowerBI is a plus

Posted 1 day ago

Apply

5.0 - 9.0 years

22 - 37 Lacs

Bengaluru

Hybrid

Naukri logo

Work closely with cross-functional teams of business partners, data scientists, data engineers, solutions, and data architects to quickly deliver scalable Artificial Intelligence (AI) solutions including DL, ML, NLP, optimization etc. Development, deployment, and maintenance of scalable AI solutions including the optimization of data queries, code refactoring, shared library usage and documentation of process and solution artifacts. Evolving, implementing, and advocating a global MLOps model across project teams to ensure standardization and consistency of AI solutions design, R&D though deployment and maintenance. Design and implement automation processes during each MLOps project phase ensuring reusable code library development, testing & adoption. Partner with business stakeholders and AI project teams to clearly understand business requirements; design and develop scalable AI solutions and promotion of their benefits back to business partners with an emphasis on value. Become a Subject Matter Expert (SME) on best usage of the teams global data science platform driving MLOps processes and reducing the cost of scalable AI delivery for business partners. Research and promote latest technologies, design patterns and best practice delivery models that drive optimal business value and ensure continuous improvement of team, processes, and platforms. Employ innovative thinking across everything you do, always seeking the best ways of working for our teams and business partners

Posted 1 day ago

Apply

10.0 years

0 Lacs

Greater Kolkata Area

On-site

Linkedin logo

Responsibilities : About Lexmark: Founded in 1991 and headquartered in Lexington, Kentucky, Lexmark is recognized as a global leader in print hardware, service, software solutions and security by many of the technology industry’s leading market analyst firms. Lexmark creates cloud-enabled imaging and IoT solutions that help customers in more than 170 countries worldwide quickly realize business outcomes. Lexmark’s digital transformation objectives accelerate business transformation, turning information into insights, data into decisions, and analytics into action. Lexmark India, located in Kolkata, is one of the research and development centers of Lexmark International Inc. The India team works on cutting edge technologies & domains like cloud, AI/ML, Data science, IoT, Cyber security on creating innovative solutions for our customers and helping them minimize the cost and IT burden in providing a secure, reliable, and productive print and imaging environment. At our core, we are a technology company – deeply committed to building our own R&D capabilities, leveraging emerging technologies and partnerships to bring together a library of intellectual property that can add value to our customer's business. Caring for our communities and creating growth opportunities by investing in talent are woven into our culture. It’s how we care, grow, and win together. Job Description/Responsibilities: We are looking for a highly skilled and strategic Data Architect with deep expertise in the Azure Data ecosystem . This role requires a strong command over Azure Databricks , Azure Data Lake , Azure Data Factory , data warehouse design , SQL optimization , and AI/ML integration . The Data Architect will design and oversee robust, scalable, and secure data architectures to support advanced analytics and machine learning workloads. Qualification: BE/ME/MCA with 10+ Years in IT Experience. Must Have Skills/Skill Requirement: Define and drive the overall Azure-based data architecture strategy aligned with enterprise goals. Architect and implement scalable data pipelines, data lakes, and data warehouses using Azure Data Lake, ADF, and Azure SQL/Synapse. Provide technical leadership on Azure Databricks (Spark, Delta Lake, Notebooks, MLflow etc.) for large-scale data processing and advanced analytics use cases. Integrate AI/ML models into data pipelines and support end-to-end ML lifecycle (training, deployment, monitoring). Collaborate with cross-functional teams including data scientists, DevOps engineers, and business analysts. Evaluate and recommend tools, platforms, and design patterns for data and ML infrastructure. Mentor data engineers and junior architects on best practices and architectural standards. Strong experience with data modeling, ETL/ELT frameworks, and data warehousing concepts. Proficient in SQL, Python, PySpark. Solid understanding of AI/ML workflows and tools. Exposure on Azure DevOps. Excellent communication and stakeholder management skills. How to Apply ? Are you an innovator? Here is your chance to make your mark with a global technology leader. Apply now! Global Privacy Notice Lexmark is committed to appropriately protecting and managing any personal information you share with us. Click here to view Lexmark's Privacy Notice.

Posted 1 day ago

Apply

5.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Linkedin logo

Sonatype is the software supply chain security company. We provide the world’s best end-to-end software supply chain security solution, combining the only proactive protection against malicious open source, the only enterprise grade SBOM management and the leading open source dependency management platform. This empowers enterprises to create and maintain secure, quality, and innovative software at scale. As founders of Nexus Repository and stewards of Maven Central, the world’s largest repository of Java open-source software, we are software pioneers and our open source expertise is unmatched. We empower innovation with an unparalleled commitment to build faster, safer software and harness AI and data intelligence to mitigate risk, maximize efficiencies, and drive powerful software development. More than 2,000 organizations, including 70% of the Fortune 100 and 15 million software developers, rely on Sonatype to optimize their software supply chains. The Opportunity We’re looking for a Senior Data Engineer to join our growing Data Platform team. You’ll play a key role in designing and scaling the infrastructure and pipelines that power analytics, machine learning, and business intelligence across Sonatype. You’ll work closely with stakeholders across product, engineering, and business teams to ensure data is reliable, accessible, and actionable. This role is ideal for someone who thrives on solving complex data challenges at scale and enjoys building high-quality, maintainable systems. What You’ll Do Design, build, and maintain scalable data pipelines and ETL/ELT processes Architect and optimize data models and storage solutions for analytics and operational use Collaborate with data scientists, analysts, and engineers to deliver trusted, high-quality datasets Own and evolve parts of our data platform (e.g., Airflow, dbt, Spark, Redshift, or Snowflake) Implement observability, alerting, and data quality monitoring for critical pipelines Drive best practices in data engineering, including documentation, testing, and CI/CD Contribute to the design and evolution of our next-generation data lakehouse architecture Minimum Qualifications 5+ years of experience as a Data Engineer or in a similar backend engineering role Strong programming skills in Python, Scala, or Java Hands-on experience with HBase or similar NoSQL columnar stores Hands-on experience with distributed data systems like Spark, Kafka, or Flink Proficient in writing complex SQL and optimizing queries for performance Experience building and maintaining robust ETL/ELT (Data Warehousing) pipelines in production Familiarity with workflow orchestration tools (Airflow, Dagster, or similar) Understanding of data modeling techniques (star schema, dimensional modeling, etc.) Bonus Points Experience working with Databricks, dbt, Terraform, or Kubernetes Familiarity with streaming data pipelines or real-time processing Exposure to data governance frameworks and tools Experience supporting data products or ML pipelines in production Strong understanding of data privacy, security, and compliance best practices Why You’ll Love Working Here Data with purpose: Work on problems that directly impact how the world builds secure software Modern tooling: Leverage the best of open-source and cloud-native technologies Collaborative culture: Join a passionate team that values learning, autonomy, and impact At Sonatype, we value diversity and inclusivity. We offer perks such as parental leave, diversity and inclusion working groups, and flexible working practices to allow our employees to show up as their whole selves. We are an equal-opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status. If you have a disability or special need that requires accommodation, please do not hesitate to let us know.

Posted 1 day ago

Apply

5.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Linkedin logo

Sonatype is the software supply chain security company. We provide the world’s best end-to-end software supply chain security solution, combining the only proactive protection against malicious open source, the only enterprise grade SBOM management and the leading open source dependency management platform. This empowers enterprises to create and maintain secure, quality, and innovative software at scale. As founders of Nexus Repository and stewards of Maven Central, the world’s largest repository of Java open-source software, we are software pioneers and our open source expertise is unmatched. We empower innovation with an unparalleled commitment to build faster, safer software and harness AI and data intelligence to mitigate risk, maximize efficiencies, and drive powerful software development. More than 2,000 organizations, including 70% of the Fortune 100 and 15 million software developers, rely on Sonatype to optimize their software supply chains. The Opportunity We’re looking for a Senior Data Engineer to join our growing Data Platform team. This role is a hybrid of data engineering and business intelligence, ideal for someone who enjoys solving complex data challenges while also building intuitive and actionable reporting solutions. You’ll play a key role in designing and scaling the infrastructure and pipelines that power analytics, dashboards, machine learning, and decision-making across Sonatype. You’ll also be responsible for delivering clear, compelling, and insightful business intelligence through tools like Looker Studio and advanced SQL queries. What You’ll Do Design, build, and maintain scalable data pipelines and ETL/ELT processes. Architect and optimize data models and storage solutions for analytics and operational use. Create and manage business intelligence reports and dashboards using tools like Looker Studio, Power BI, or similar. Collaborate with data scientists, analysts, and stakeholders to ensure datasets are reliable, meaningful, and actionable. Own and evolve parts of our data platform (e.g., Airflow, dbt, Spark, Redshift, or Snowflake). Write complex, high-performance SQL queries to support reporting and analytics needs. Implement observability, alerting, and data quality monitoring for critical pipelines. Drive best practices in data engineering and business intelligence, including documentation, testing, and CI/CD. Contribute to the evolution of our next-generation data lakehouse and BI architecture. What We’re Looking For 5+ years of experience as a Data Engineer or in a hybrid data/reporting role. Strong programming skills in Python, Java, or Scala. Proficiency with data tools such as Databricks, data modeling techniques (e.g., star schema, dimensional modeling), and data warehousing solutions like Snowflake or Redshift. Hands-on experience with modern data platforms and orchestration tools (e.g., Spark, Kafka, Airflow). Proficient in SQL with experience in writing and optimizing complex queries for BI and analytics. Experience with BI tools such as Looker Studio, Power BI, or Tableau. Experience in building and maintaining robust ETL/ELT pipelines in production. Understanding of data quality, observability, and governance best practices. 5+ years of experience as a Data Engineer or in a hybrid data/reporting role. Strong programming skills in Python, Java, or Scala. Hands-on experience with modern data platforms and orchestration tools (e.g., Spark, Kafka, Airflow). Proficient in SQL with experience in writing and optimizing complex queries for BI and analytics. Experience with BI tools such as Looker Studio, Power BI, or Tableau. Familiarity with data modeling techniques (star schema, dimensional modeling, etc.). Experience in building and maintaining robust ETL/ELT pipelines in production. Understanding data quality, observability, and governance best practices. Bonus Points Experience with dbt, Terraform, or Kubernetes. Familiarity with real-time data processing or streaming architectures. Understanding of data privacy, compliance, and security best practices in analytics and reporting. Why You’ll Love Working Here Data with purpose: Work on problems that directly impact how the world builds secure software. Full-spectrum impact: Use both engineering and analytical skills to shape product, strategy, and operations. Modern tooling: Leverage the best of open-source and cloud-native technologies. Collaborative culture: Join a passionate team that values learning, autonomy, and real-world impact. At Sonatype, we value diversity and inclusivity. We offer perks such as parental leave, diversity and inclusion working groups, and flexible working practices to allow our employees to show up as their whole selves. We are an equal-opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status. If you have a disability or special need that requires accommodation, please do not hesitate to let us know.

Posted 1 day ago

Apply

6.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Linkedin logo

Sonatype is the software supply chain security company. We provide the world’s best end-to-end software supply chain security solution, combining the only proactive protection against malicious open source, the only enterprise grade SBOM management and the leading open source dependency management platform. This empowers enterprises to create and maintain secure, quality, and innovative software at scale. As founders of Nexus Repository and stewards of Maven Central, the world’s largest repository of Java open-source software, we are software pioneers and our open source expertise is unmatched. We empower innovation with an unparalleled commitment to build faster, safer software and harness AI and data intelligence to mitigate risk, maximize efficiencies, and drive powerful software development. More than 2,000 organizations, including 70% of the Fortune 100 and 15 million software developers, rely on Sonatype to optimize their software supply chains. About The Role The Engineering Manager – Data role at Sonatype blends hands-on data engineering with leadership and strategic influence. You will lead high-performing data engineering teams to build the infrastructure, pipelines, and systems that fuel analytics, business intelligence, and machine learning across our global products. We’re looking for a leader who brings deep technical experience in modern data platforms, is fluent in programming, and understands the nuances of open-source consumption and software supply chain security. This hybrid role is based out of our Hyderabad office. What You’ll Do Lead, mentor, and grow a team of data engineers responsible for building scalable, secure, and maintainable data solutions. Design and architect data pipelines, Lakehouse systems, and warehouse models using tools such as Databricks, Airflow, Spark, and Snowflake/Redshift. Stay hands-on—write, review, and guide production-level code in Python, Java, or similar languages. Ensure strong foundations in data modeling, governance, observability, and data quality. Collaborate with cross-functional teams including Product, Security, Engineering, and Data Science to translate business needs into data strategies and deliverables. Apply your knowledge of open-source component usage, dependency management, and software composition analysis to ensure our data platforms support secure development practices. Embed application security principles into data platform design, supporting Sonatype’s mission to secure the software supply chain. Foster an engineering culture that prioritizes continuous improvement, technical excellence, and team ownership. Who You Are A technical leader with a strong background in data engineering, platform design, and secure software development. Comfortable operating across domains—data infrastructure, programming, architecture, security, and team leadership. Passionate about delivering high-impact results through technical contributions, mentoring, and strategic thinking. Familiar with modern data engineering practices, open-source ecosystems, and the challenges of managing data securely on a scale. A collaborative communicator who thrives in hybrid and cross-functional team environments. What You Need 6+ years of experience in data engineering, backend systems, or infrastructure development. 2+ year of experience in a technical leadership or engineering management role with hands-on contribution. Expertise in data technologies: Databricks, Spark, Airflow, Snowflake/Redshift, dbt, etc. Strong programming skills in Python, Java, or Scala with experience building robust, production-grade systems. Experience in data modeling (dimensional modeling, star/snowflake schema), data warehousing, and ELT/ETL pipeline development. Understanding software dependency management and open-source consumption patterns. Familiarity with application security principles and a strong interest in secure software supply chains. Experience supporting real-time data systems or streaming architectures. Exposure to machine learning pipelines or data productization. Experience with tools like Terraform, Kubernetes, and CI/CD for data engineering workflows. Knowledge of data governance frameworks and regulatory compliance (GDPR, SOC2, etc.). Why Join Us? Help secure the software supply chain for millions of developers worldwide. Build meaningful software in a collaborative, fast-moving environment with strong technical peers. Stay hands-on while leading—technical leadership is part of the job, not separate from it. Join a global engineering organization with deep local roots and a strong team culture. Competitive salary, great benefits, and opportunities for growth and innovation. At Sonatype, we value diversity and inclusivity. We offer perks such as parental leave, diversity and inclusion working groups, and flexible working practices to allow our employees to show up as their whole selves. We are an equal-opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status. If you have a disability or special need that requires accommodation, please do not hesitate to let us know.

Posted 1 day ago

Apply

4.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Linkedin logo

The Role The Data Engineer is accountable for developing high quality data products to support the Bank’s regulatory requirements and data driven decision making. A Mantas Scenario Developer will serve as an example to other team members, work closely with customers, and remove or escalate roadblocks. By applying their knowledge of data architecture standards, data warehousing, data structures, and business intelligence they will contribute to business outcomes on an agile team. Responsibilities Developing and supporting scalable, extensible, and highly available data solutions Deliver on critical business priorities while ensuring alignment with the wider architectural vision Identify and help address potential risks in the data supply chain Follow and contribute to technical standards Design and develop analytical data models Required Qualifications & Work Experience First Class Degree in Engineering/Technology (4-year graduate course) 3 to 4 years’ experience implementing data-intensive solutions using agile methodologies Experience of relational databases and using SQL for data querying, transformation and manipulation Experience of modelling data for analytical consumers Hands on Mantas (Oracle FCCM) Scenario Development experience throughout the full development life cycle Ability to automate and streamline the build, test and deployment of data pipelines Experience in cloud native technologies and patterns A passion for learning new technologies, and a desire for personal growth, through self-study, formal classes, or on-the-job training Excellent communication and problem-solving skills T echnical Skills (Must Have) ETL: Hands on experience of building data pipelines. Proficiency in at least one of the data integration platforms such as Ab Initio, Apache Spark, Talend and Informatica Mantas: Expert in Oracle Mantas/FCCM, Scenario Manager, Scenario Development, thorough knowledge and hands on experience in Mantas FSDM, DIS, Batch Scenario Manager Big Data: Exposure to ‘big data’ platforms such as Hadoop, Hive or Snowflake for data storage and processing Data Warehousing & Database Management: Understanding of Data Warehousing concepts, Relational (Oracle, MSSQL, MySQL) and NoSQL (MongoDB, DynamoDB) database design Data Modeling & Design: Good exposure to data modeling techniques; design, optimization and maintenance of data models and data structures Languages: Proficient in one or more programming languages commonly used in data engineering such as Python, Java or Scala DevOps: Exposure to concepts and enablers - CI/CD platforms, version control, automated quality control management Technical Skills (Valuable) Ab Initio: Experience developing Co>Op graphs; ability to tune for performance. Demonstrable knowledge across full suite of Ab Initio toolsets e.g., GDE, Express>IT, Data Profiler and Conduct>IT, Control>Center, Continuous>Flows Cloud: Good exposure to public cloud data platforms such as S3, Snowflake, Redshift, Databricks, BigQuery, etc. Demonstrable understanding of underlying architectures and trade-offs Data Quality & Controls: Exposure to data validation, cleansing, enrichment and data controls Containerization: Fair understanding of containerization platforms like Docker, Kubernetes File Formats: Exposure in working on Event/File/Table Formats such as Avro, Parquet, Iceberg, Delta Others: Basics of Job scheduler like Autosys. Basics of Entitlement management Certification on any of the above topics would be an advantage. ------------------------------------------------------ Job Family Group: Technology ------------------------------------------------------ Job Family: Digital Software Engineering ------------------------------------------------------ Time Type: Full time ------------------------------------------------------ Most Relevant Skills Please see the requirements listed above. ------------------------------------------------------ Other Relevant Skills For complementary skills, please see above and/or contact the recruiter. ------------------------------------------------------ Citi is an equal opportunity employer, and qualified candidates will receive consideration without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, disability, status as a protected veteran, or any other characteristic protected by law. If you are a person with a disability and need a reasonable accommodation to use our search tools and/or apply for a career opportunity review Accessibility at Citi. View Citi’s EEO Policy Statement and the Know Your Rights poster.

Posted 1 day ago

Apply

5.0 - 10.0 years

20 - 25 Lacs

Bengaluru

Remote

Naukri logo

Role & responsibilities At least 5 years of experience in data engineering with a strong background on Azure Databricks and Scala/Python. •Experience in handling unstructured data processing and transformation with programming knowledge. •Hands on experience in building data pipelines using Scala/Python •Big data technologies such as Apache Spark, Structured Streaming, SQL, Databricks Delta Lake •Strong analytical and problem solving skills with the ability to troubleshoot spark applications and resolve data pipeline issues. •Familiarity with version control systems like Git, CICD pipelines using Jenkins.

Posted 1 day ago

Apply

3.0 years

0 Lacs

Greater Kolkata Area

On-site

Linkedin logo

Line of Service Advisory Industry/Sector FS X-Sector Specialism Data, Analytics & AI Management Level Associate Job Description & Summary At PwC, our people in data and analytics focus on leveraging data to drive insights and make informed business decisions. They utilise advanced analytics techniques to help clients optimise their operations and achieve their strategic goals. In business intelligence at PwC, you will focus on leveraging data and analytics to provide strategic insights and drive informed decision-making for clients. You will develop and implement innovative solutions to optimise business performance and enhance competitive advantage. *Why PWC At PwC, you will be part of a vibrant community of solvers that leads with trust and creates distinctive outcomes for our clients and communities. This purpose-led and values-driven work, powered by technology in an environment that drives innovation, will enable you to make a tangible impact in the real world. We reward your contributions, support your wellbeing, and offer inclusive benefits, flexibility programmes and mentorship that will help you thrive in work and life. Together, we grow, learn, care, collaborate, and create a future of infinite experiences for each other. Learn more about us . At PwC, we believe in providing equal employment opportunities, without any discrimination on the grounds of gender, ethnic background, age, disability, marital status, sexual orientation, pregnancy, gender identity or expression, religion or other beliefs, perceived differences and status protected by law. We strive to create an environment where each one of our people can bring their true selves and contribute to their personal growth and the firm’s growth. To enable this, we have zero tolerance for any discrimination and harassment based on the above considerations. " Responsibilities: Senior Associate Exp : 3 - 6 Years Location: Kolkata Technical Skills: · Strong expertise in Azure Databricks, Azure Data Factory (ADF), PySpark, SQL Server, and Python. · Solid understanding of Azure Functions and their application in data processing workflows. · Understanding of DevOps practices and CI/CD pipelines for data solutions. · Experience with other ETL tools such as Informatica Intelligent Cloud Services (IICS) is a plus. · Strong problem-solving skills and ability to work independently and collaboratively in a fast-paced environment. · Excellent communication skills to effectively convey technical concepts to non-technical stakeholders. Key Responsibilities: · Develop, maintain, and optimize scalable data pipelines using Azure Databricks, Azure Data Factory (ADF), and PySpark. · Collaborate with data architects and business stakeholders to translate requirements into technical solutions. · Implement and manage data integration processes using SQL Server and Python. · Design and deploy Azure Functions to support data processing workflows. · Monitor and troubleshoot data pipeline performance and reliability issues. · Ensure data quality, security, and compliance with industry standards and best practices. · Document technical specifications and maintain clear and concise project documentation. Mandatory skill sets: Azure Databricks, Azure Data Factory (ADF), and PySpark. Preferred skill sets: Azure Databricks, Azure Data Factory (ADF), and PySpark. Years of experience required: 3-6 Years Education qualification: B.E.(B.Tech)/M.E/M.Tech Education (if blank, degree and/or field of study not specified) Degrees/Field of Study required: Bachelor of Engineering, Master of Engineering Degrees/Field of Study preferred: Certifications (if blank, certifications not specified) Required Skills Microsoft Azure Optional Skills Accepting Feedback, Accepting Feedback, Active Listening, Business Case Development, Business Data Analytics, Business Intelligence and Reporting Tools (BIRT), Business Intelligence Development Studio, Communication, Competitive Advantage, Continuous Process Improvement, Data Analysis and Interpretation, Data Architecture, Database Management System (DBMS), Data Collection, Data Pipeline, Data Quality, Data Science, Data Visualization, Emotional Regulation, Empathy, Inclusion, Industry Trend Analysis, Intellectual Curiosity, Java (Programming Language), Market Development {+ 11 more} Desired Languages (If blank, desired languages not specified) Travel Requirements Not Specified Available for Work Visa Sponsorship? No Government Clearance Required? No Job Posting End Date

Posted 1 day ago

Apply

0 years

0 Lacs

Greater Kolkata Area

On-site

Linkedin logo

Line of Service Advisory Industry/Sector Not Applicable Specialism Microsoft Management Level Senior Associate Job Description & Summary At PwC, our people in software and product innovation focus on developing cutting-edge software solutions and driving product innovation to meet the evolving needs of clients. These individuals combine technical experience with creative thinking to deliver innovative software products and solutions. Those in software engineering at PwC will focus on developing innovative software solutions to drive digital transformation and enhance business performance. In this field, you will use your knowledge to design, code, and test cutting-edge applications that revolutionise industries and deliver exceptional user experiences. Why PWC At PwC, you will be part of a vibrant community of solvers that leads with trust and creates distinctive outcomes for our clients and communities. This purpose-led and values-driven work, powered by technology in an environment that drives innovation, will enable you to make a tangible impact in the real world. We reward your contributions, support your wellbeing, and offer inclusive benefits, flexibility programmes and mentorship that will help you thrive in work and life. Together, we grow, learn, care, collaborate, and create a future of infinite experiences for each other. Learn more about us . At PwC, we believe in providing equal employment opportunities, without any discrimination on the grounds of gender, ethnic background, age, disability, marital status, sexual orientation, pregnancy, gender identity or expression, religion or other beliefs, perceived differences and status protected by law. We strive to create an environment where each one of our people can bring their true selves and contribute to their personal growth and the firm’s growth. To enable this, we have zero tolerance for any discrimination and harassment based on the above considerations. Job Description & Summary: We are seeking a Data Engineer to design, develop, and maintain data ingestion processes to a data platform built using Microsoft Technologies, ensuring data quality and integrity. The role involves collaborating with data architects and business analysts to implement solutions using tools like ADF, Azure Databricks, and requires strong SQL skills. Responsibilities: Key responsibilities include developing, testing, and optimizing ETL workflows and maintaining documentation. ETL development experience in Microsoft data track are required. Work with business team to translate the business requirement to technical requirements. Demonstrated expertise in Agile methodologies, including Scrum, Kanban, or SAFe. Mandatory skill sets: · Strong proficiency in Azure Databricks, including Spark and Delta Lake. · Experience with Azure Data Factory, Azure Data Lake Storage, and Azure SQL Database. · Proficiency in data integration and ETL processes and T-SQL. · Experienced working in Python for data engineering · Experienced working in Postgres Database · Experienced working in graph database · Experienced in architecture design and data modelling Good To Have Skill Sets: · Unity Catalog / Purview · Familiarity with Fabric/Snowflake service offerings · Visualization tool – PowerBI Preferred skill sets: Hands on knowledge of python, Pyspark and strong SQL knowledge. ETL and data warehousing is must. Relevant certifications (Any one) (e.g., Databricks Data Engineer Associate Microsoft Certified: Azure Data Engineer Associate Azure Solution Architect) are mandatory Years of experience required: 5+yrs Education qualification: Bachelor's degree in Computer Science, IT, or a related field. Education (if blank, degree and/or field of study not specified) Degrees/Field of Study required: Bachelor of Engineering Degrees/Field of Study preferred: Certifications (if blank, certifications not specified) Required Skills Data Engineering Optional Skills Acceptance Test Driven Development (ATDD), Acceptance Test Driven Development (ATDD), Accepting Feedback, Active Listening, Analytical Thinking, Android, API Management, Appian (Platform), Application Development, Application Frameworks, Application Lifecycle Management, Application Software, Business Process Improvement, Business Process Management (BPM), Business Requirements Analysis, C#.NET, C++ Programming Language, Client Management, Code Review, Coding Standards, Communication, Computer Engineering, Computer Science, Continuous Integration/Continuous Delivery (CI/CD), Creativity {+ 46 more} Desired Languages (If blank, desired languages not specified) Travel Requirements Available for Work Visa Sponsorship? Government Clearance Required? Job Posting End Date

Posted 1 day ago

Apply

10.0 years

0 Lacs

India

On-site

Linkedin logo

Coursera was launched in 2012 by Andrew Ng and Daphne Koller, with a mission to provide universal access to world-class learning. It is now one of the largest online learning platforms in the world, with 175 million registered learners as of March 31, 2025. Coursera partners with over 350 leading universities and industry leaders to offer a broad catalog of content and credentials, including courses, Specializations, Professional Certificates, and degrees. Coursera’s platform innovations enable instructors to deliver scalable, personalized, and verified learning experiences to their learners. Institutions worldwide rely on Coursera to upskill and reskill their employees, citizens, and students in high-demand fields such as GenAI, data science, technology, and business. Coursera is a Delaware public benefit corporation and a B Corp. Join us in our mission to create a world where anyone, anywhere can transform their life through access to education. We're seeking talented individuals who share our passion and drive to revolutionize the way the world learns. At Coursera, we are committed to building a globally diverse team and are thrilled to extend employment opportunities to individuals in any country where we have a legal entity. We require candidates to possess eligible working rights and have a compatible timezone overlap with their team to facilitate seamless collaboration. Coursera has a commitment to enabling flexibility and workspace choices for employees. Our interviews and onboarding are entirely virtual, providing a smooth and efficient experience for our candidates. As an employee, we enable you to select your main way of working, whether it's from home, one of our offices or hubs, or a co-working space near you. Overview: At Coursera, our Data team is helping to build the future of education through data-driven decision making and data-powered products. We drive marketing, product and business strategy through ensuring high quality data, measurement, experimentation, and causal inference. We are seeking a dynamic, collaborative, and experienced Senior Manager to lead our Data Science team, responsible for KPI reporting and business analytics for business teams and functions. Reporting to the Director of Data Science, this role will play a pivotal role in shaping strategies and initiatives to enable our leaders, stakeholders and partners with reliable data, automation and analytics at Coursera, to drive business growth. Responsibilities: Hire, mentor, and lead a team of data scientists, fostering a culture of continuous learning and integration of analytics into business teams to enhance our online learning platform. Collaborate with stakeholders to understand requirements, develop technical roadmaps, and identify opportunities for automation and efficiency, ensuring the delivery of impactful data products. Develop scalable data solutions by building reliable data layers, improving self-service capabilities, and driving data-driven decision-making across the organization. Conduct deep-dive analyses and hypothesis testing in partnership with teams and stakeholders to uncover actionable insights and improve business outcomes. Build forecasting models to identify patterns, predict future demand, and optimize resource allocation, enabling informed strategic decision-making. Influence strategic direction by partnering with senior leadership, providing data-driven recommendations, and contributing to the company’s long-term vision. Drive innovation and skill development by staying updated on industry trends, emerging technologies, and best practices in business intelligence and analytics. Bridge technical and non-technical teams, ensuring clear communication, aligning priorities, and fostering strong, collaborative relationships to create a positive and productive work environment. Basic Qualifications: Bachelor's degree in Computer Science, Information Systems, Mathematics, Statistics, or related field; Master's degree preferred. 10+ years of experience in data analytics, business intelligence, or related roles, with 4+ years of managerial experience. Proficiency in programming languages such as SQL, Python, or R and experience with relevant data science libraries and frameworks. Hands-on experience with data visualization tools such as Looker, Tableau, or Amplitude. Strong understanding of data warehousing concepts and experience working with cloud-based data platforms (e.g., Databricks, AWS, Google Cloud). Excellent communication skills, with the ability to effectively communicate technical concepts to non-technical stakeholders. Proven leadership abilities, with a track record of successfully managing teams and driving projects to completion. Strong analytical and problem-solving skills, with the ability to translate data into actionable insights and recommendations. Preferred Qualifications: Experience at EdTech or Content Subscription business. Experience working with Salesforce data. Experience with Airflow, Databricks. Experience with PM tools like JIRA, confluence. If this opportunity interests you, you might like these courses on Coursera: Go Beyond the Numbers: Translate Data into Insights Applied AI with DeepLearning Probability & Statistics for Machine Learning & Data Science Coursera is an Equal Employment Opportunity Employer and considers all qualified applicants without regard to race, color, religion, sex, sexual orientation, gender identity, age, marital status, national origin, protected veteran status, disability, or any other legally protected class. If you are an individual with a disability and require a reasonable accommodation to complete any part of the application process, please contact us at accommodations@coursera.org. For California Candidates, please review our CCPA Applicant Notice here. For our Global Candidates, please review our GDPR Recruitment Notice here.

Posted 1 day ago

Apply

10.0 years

15 - 17 Lacs

India

Remote

Linkedin logo

Note: This is a remote role with occasional office visits. Candidates from Mumbai or Pune will be preferred About The Company A fast-growing enterprise technology consultancy operating at the intersection of cloud computing, big-data engineering and advanced analytics . The team builds high-throughput, real-time data platforms that power AI, BI and digital products for Fortune 500 clients across finance, retail and healthcare. By combining Databricks Lakehouse architecture with modern DevOps practices, they unlock insight at petabyte scale while meeting stringent security and performance SLAs. Role & Responsibilities Architect end-to-end data pipelines (ingestion → transformation → consumption) using Databricks, Spark and cloud object storage. Design scalable data warehouses/marts that enable self-service analytics and ML workloads. Translate logical data models into physical schemas; own database design, partitioning and lifecycle management for cost-efficient performance. Implement, automate and monitor ETL/ELT workflows, ensuring reliability, observability and robust error handling. Tune Spark jobs and SQL queries, optimizing cluster configurations and indexing strategies to achieve sub-second response times. Provide production support and continuous improvement for existing data assets, championing best practices and mentoring peers. Skills & Qualifications Must-Have 6–10 years building production-grade data platforms, including 3 years+ hands-on Apache Spark/Databricks experience. Expert proficiency in PySpark, Python and advanced SQL, with a track record of performance-tuning distributed jobs. Demonstrated ability to model data warehouses/marts and orchestrate ETL/ELT pipelines with tools such as Airflow or dbt. Hands-on with at least one major cloud platform (AWS or Azure) and modern lakehouse / data-lake patterns. Strong problem-solving skills, DevOps mindset and commitment to code quality; comfortable mentoring fellow engineers. Preferred Deep familiarity with the AWS analytics stack (Redshift, Glue, S3) or the broader Hadoop ecosystem. Bachelor’s or Master’s degree in Computer Science, Engineering or a related field. Experience building streaming pipelines (Kafka, Kinesis, Delta Live Tables) and real-time analytics solutions. Exposure to ML feature stores, MLOps workflows and data-governance/compliance frameworks. Relevant professional certifications (Databricks, AWS, Azure) or notable open-source contributions. Benefits & Culture Highlights Remote-first & flexible hours with 25+ PTO days and comprehensive health cover. Annual training budget & certification sponsorship (Databricks, AWS, Azure) to fuel continuous learning. Inclusive, impact-focused culture where engineers shape the technical roadmap and mentor a vibrant data community Skills: data modeling,big data technologies,team leadership,agile methodologies,performance tuning,data,aws,airflow

Posted 1 day ago

Apply

10.0 years

15 - 17 Lacs

India

Remote

Linkedin logo

Note: This is a remote role with occasional office visits. Candidates from Mumbai or Pune will be preferred About The Company A fast-growing enterprise technology consultancy operating at the intersection of cloud computing, big-data engineering and advanced analytics . The team builds high-throughput, real-time data platforms that power AI, BI and digital products for Fortune 500 clients across finance, retail and healthcare. By combining Databricks Lakehouse architecture with modern DevOps practices, they unlock insight at petabyte scale while meeting stringent security and performance SLAs. Role & Responsibilities Architect end-to-end data pipelines (ingestion → transformation → consumption) using Databricks, Spark and cloud object storage. Design scalable data warehouses/marts that enable self-service analytics and ML workloads. Translate logical data models into physical schemas; own database design, partitioning and lifecycle management for cost-efficient performance. Implement, automate and monitor ETL/ELT workflows, ensuring reliability, observability and robust error handling. Tune Spark jobs and SQL queries, optimizing cluster configurations and indexing strategies to achieve sub-second response times. Provide production support and continuous improvement for existing data assets, championing best practices and mentoring peers. Skills & Qualifications Must-Have 6–10 years building production-grade data platforms, including 3 years+ hands-on Apache Spark/Databricks experience. Expert proficiency in PySpark, Python and advanced SQL, with a track record of performance-tuning distributed jobs. Demonstrated ability to model data warehouses/marts and orchestrate ETL/ELT pipelines with tools such as Airflow or dbt. Hands-on with at least one major cloud platform (AWS or Azure) and modern lakehouse / data-lake patterns. Strong problem-solving skills, DevOps mindset and commitment to code quality; comfortable mentoring fellow engineers. Preferred Deep familiarity with the AWS analytics stack (Redshift, Glue, S3) or the broader Hadoop ecosystem. Bachelor’s or Master’s degree in Computer Science, Engineering or a related field. Experience building streaming pipelines (Kafka, Kinesis, Delta Live Tables) and real-time analytics solutions. Exposure to ML feature stores, MLOps workflows and data-governance/compliance frameworks. Relevant professional certifications (Databricks, AWS, Azure) or notable open-source contributions. Benefits & Culture Highlights Remote-first & flexible hours with 25+ PTO days and comprehensive health cover. Annual training budget & certification sponsorship (Databricks, AWS, Azure) to fuel continuous learning. Inclusive, impact-focused culture where engineers shape the technical roadmap and mentor a vibrant data community Skills: data modeling,big data technologies,team leadership,agile methodologies,performance tuning,data,aws,airflow

Posted 1 day ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies