Home
Jobs

1243 Inference Jobs - Page 7

Filter Interviews
Min: 0 years
Max: 25 years
Min: ₹0
Max: ₹10000000
Setup a job Alert
Filter
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

3.0 years

0 Lacs

Chennai, Tamil Nadu, India

On-site

Linkedin logo

Job Description Global Data Insight & Analytics organization is looking for a top-notch Software Engineer who has also got Machine Learning knowledge & Experience to add to our team to drive the next generation of AI/ML (Mach1ML) platform. In this role you will work in a small, cross-functional team. The position will collaborate directly and continuously with other engineers, business partners, product managers and designers from distributed locations, and will release early and often. The team you will be working on is focused on building Mach1ML platform – an AI/ML enablement platform to democratize Machine Learning across Ford enterprise (like OpenAI’s GPT, Facebook’s FBLearner, etc.) to deliver next-gen analytics innovation. We strongly believe that data has the power to help create great products and experiences which delight our customers. We believe that actionable and persistent insights, based on high quality data platform, help business and engineering make more impactful decisions. Our ambitions reach well beyond existing solutions, and we are in search of innovative individuals to join this Agile team. This is an exciting, fast-paced role which requires outstanding technical and organization skills combined with critical thinking, problem-solving and agile management tools to support team success. Responsibilities What you'll be able to do: As a Software Engineer, you will work on developing features for Mach1ML platform, support customers in model deployment using Mach1ML platform on GCP and On-prem. You will follow Rally to manage your work. You will incorporate an understanding of product functionality and customer perspective for model deployment. You will work on the cutting-edge technologies such as GCP, Kubernetes, Docker, Seldon, Tekton, Airflow, Rally, etc. Position Responsibilities: Work closely with Tech Anchor, Product Manager and Product Owner to deliver machine learning use cases using Ford Agile Framework. Work with Data Scientists and ML engineers to tackle challenging AI problems. Work specifically on the Deploy team to drive model deployment and AI/ML adoption with other internal and external systems. Help innovate by researching state-of-the-art deployment tools and share knowledge with the team. Lead by example in use of Paired Programming for cross training/upskilling, problem solving, and speed to delivery. Leverage latest GCP, CICD, ML technologies Critical Thinking: Able to influence the strategic direction of the company by finding opportunities in large, rich data sets and crafting and implementing data driven strategies that fuel growth including cost savings, revenue, and profit. Modelling: Assessments, and evaluating impacts of missing/unusable data, design and select features, develop, and implement statistical/predictive models using advanced algorithms on diverse sources of data and testing and validation of models, such as forecasting, natural language processing, pattern recognition, machine vision, supervised and unsupervised classification, decision trees, neural networks, etc. Analytics: Leverage rigorous analytical and statistical techniques to identify trends and relationships between different components of data, draw appropriate conclusions and translate analytical findings and recommendations into business strategies or engineering decisions - with statistical confidence Data Engineering: Experience with crafting ETL processes to source and link data in preparation for Model/Algorithm development. This includes domain expertise of data sets in the environment, third-party data evaluations, data quality Visualization: Build visualizations to connect disparate data, find patterns and tell engaging stories. This includes both scientific visualization as well as geographic using applications such as Seaborn, Qlik Sense/PowerBI/Tableau/Looker Studio, etc. Qualifications Minimum Requirements we seek: Bachelor’s or master’s degree in computer science engineering or related field or a combination of education and equivalent experience. 3+ years of experience in full stack software development 3+ years’ experience in Cloud technologies & services, preferably GCP 3+ years of experience of practicing statistical methods and their accurate application e.g. ANOVA, principal component analysis, correspondence analysis, k-means clustering, factor analysis, multi-variate analysis, Neural Networks, causal inference, Gaussian regression, etc. 3+ years’ experience with Python, SQL, BQ. Experience in SonarQube, CICD, Tekton, terraform, GCS, GCP Looker, Google cloud build, cloud run, Vertex AI, Airflow, TensorFlow, etc., Experience in Train, Build and Deploy ML, DL Models Experience in HuggingFace, Chainlit, Streamlit, React Ability to understand technical, functional, non-functional, security aspects of business requirements and delivering them end-to-end. Ability to adapt quickly with opensource products & tools to integrate with ML Platforms Building and deploying Models (Scikit learn, DataRobots, TensorFlow PyTorch, etc.) Developing and deploying On-Prem & Cloud environments Kubernetes, Tekton, OpenShift, Terraform, Vertex AI Our Preferred Requirements: Master’s degree in computer science engineering, or related field or a combination of education and equivalent experience. Demonstrated successful application of analytical methods and machine learning techniques with measurable impact on product/design/business/strategy. Proficiency in programming languages such as Python with a strong emphasis on machine learning libraries, generative AI frameworks, and monitoring tools. Utilize tools and technologies such as TensorFlow, PyTorch, scikit-learn, and other machine learning libraries to build and deploy machine learning solutions on cloud platforms. Design and implement cloud infrastructure using technologies such as Kubernetes, Terraform, and Tekton to support scalable and reliable deployment of machine learning models, generative AI models, and applications. Integrate machine learning and generative AI models into production systems on cloud platforms such as Google Cloud Platform (GCP) and ensure scalability, performance, and proactive monitoring. Implement monitoring solutions to track the performance, health, and security of systems and applications, utilizing tools such as Prometheus, Grafana, and other relevant monitoring tools. Conduct code reviews and provide constructive feedback to team members on machine learning-related projects. Knowledge and experience in agentic workflow based application development and DevOps Stay up to date with the latest trends and advancements in machine learning and data science.

Posted 1 week ago

Apply

10.0 years

7 - 9 Lacs

Hyderābād

On-site

GlassDoor logo

Summary At Novartis, we are reimagining medicine by harnessing the power of data and AI. As a Senior Architect – AI Products supporting our Commercial function, you will drive the architectural strategy that enables seamless integration of data and AI products across omnichannel engagement, customer analytics, field operations, and real-world insights. You will work across commercial business domains, data platforms, and AI product teams to design scalable, interoperable, and compliant solutions that maximize the impact of data and advanced analytics on how we engage with healthcare professionals and patients. About the Role Position Title: Assoc. Dir. DDIT US&I AI Architect (Commercial) Location – Hyd-India #LI Hybrid About the Role At Novartis, we are reimagining medicine by harnessing the power of data and AI. As a Senior Architect – AI Products supporting our Commercial function, you will drive the architectural strategy that enables seamless integration of data and AI products across omnichannel engagement, customer analytics, field operations, and real-world insights. You will work across commercial business domains, data platforms, and AI product teams to design scalable, interoperable, and compliant solutions that maximize the impact of data and advanced analytics on how we engage with healthcare professionals and patients. Your responsibilities include but are not limited to: Commercial Architecture Strategy: Define and drive the reference architecture for commercial data and AI products, ensuring alignment with enterprise standards and business priorities. Cross-Product Integration: Architect how data products (e.g., HCP 360, engagement data platforms, real-world data assets) connect with AI products (e.g., field force recommendations, predictive models, generative AI copilots) and downstream tools. Modular, Scalable Design: Ensure architecture promotes reuse, scalability, and interoperability across multiple markets, brands, and data domains within the commercial landscape. Stakeholder Alignment: Partner with commercial product managers, data science teams, platform engineering, and global/local stakeholders to guide solution design, delivery, and lifecycle evolution. Data & AI Lifecycle Enablement: Support the full lifecycle of data and AI—from ingestion and transformation to model training, inference, and monitoring—within compliant and secure environments. Governance & Compliance: Ensure architecture aligns with GxP, data privacy, and commercial compliance requirements (e.g., consent management, data traceability). Innovation & Optimization: Recommend architectural improvements, modern technologies, and integration patterns to support personalization, omnichannel engagement, segmentation, targeting, and performance analytics. What you’ll bring to the role: Proven ability to lead cross-functional architecture efforts across business, data, and technology teams. Good understanding of security, compliance, and privacy regulations in a commercial pharma setting. Experience with pharmaceutical commercial ecosystems and data (e.g., IQVIA, Veeva, Symphony). Familiarity with customer data platforms (CDPs), identity resolution, and marketing automation tools. Desirable Requirements: Bachelor's or master’s degree in computer science, Engineering, Data Science, or a related field. 10+ years of experience in enterprise or solution architecture, with significant experience in commercial functions (preferably in pharma or life sciences). Strong background in data platforms, pipelines, and governance (e.g., Snowflake, Databricks, CDP, Salesforce integration). Hands-on experience integrating solutions across Martech, CRM, and omnichannel systems. Strong knowledge of AI/ML architectures, particularly those supporting commercial use cases (recommendation engines, predictive analytics, NLP, LLMs). Exposure to GenAI applications in commercial (e.g., content generation, intelligent assistants). Understanding of global-to-local deployment patterns and data sharing requirements Commitment to Diversity & Inclusion: Novartis embraces diversity, equal opportunity, and inclusion. We are committed to building diverse teams, representative of the patients and communities we serve, and we strive to create an inclusive workplace that cultivates bold innovation through collaboration and empowers our people to unleash their full potential. Why Novartis: Helping people with disease and their families takes more than innovative science. It takes a community of smart, passionate people like you. Collaborating, supporting and inspiring each other. Combining to achieve breakthroughs that change patients’ lives. Ready to create a brighter future together? https://www.novartis.com/about/strategy/people-and-culture Join our Novartis Network: Not the right Novartis role for you? Sign up to our talent community to stay connected and learn about suitable career opportunities as soon as they come up: https://talentnetwork.novartis.com/network Benefits and Rewards: Read our handbook to learn about all the ways we’ll help you thrive personally and professionally: https://www.novartis.com/careers/benefits-rewards Division Operations Business Unit CTS Location India Site Hyderabad (Office) Company / Legal Entity IN10 (FCRS = IN010) Novartis Healthcare Private Limited Functional Area Technology Transformation Job Type Full time Employment Type Regular Shift Work No

Posted 1 week ago

Apply

8.0 years

4 - 8 Lacs

Gurgaon

On-site

GlassDoor logo

JOB DESCRIPTION AI Lead - Innovation & Product Development About Us KPMG is a dynamic and forward-thinking Professional service firm committed to leveraging cutting-edge artificial intelligence to create transformative products and solutions. We are building a team of passionate innovators who thrive on solving complex challenges and pushing the boundaries of what's possible with AI. Job Summary We are seeking an experienced and visionary AI Lead to spearhead our AI innovation and product development. The ideal candidate will be a hands-on leader with a strong background in solution architecture, a proven track record in developing AI-based products, and deep expertise in Generative AI applications, including Agentic AI. This role requires a comprehensive understanding of AI models, frameworks, and Agentic AI, along with exposure to GPU infrastructure, to design, build, and deploy scalable AI solutions. You will drive our AI strategy, lead cross-functional teams, and transform complex ideas into tangible, market-ready products, with a strong understanding of enterprise requirements from a professional services perspective. Key Responsibilities Strategic Leadership & Innovation: o Define and drive the AI innovation roadmap, identifying emerging trends in AI, Generative AI and Agentic AI. o Lead research, evaluation, and adoption of new AI models, algorithms, and frameworks. o Foster a culture of continuous learning, experimentation, and innovation. AI Product Development & Management: o Lead end-to-end development of AI-based products, from ideation to deployment and optimization. o Collaborate with product managers, designers, and stakeholders to translate business requirements into viable AI solutions. o Ensure successful delivery of high-quality, scalable, and performant AI products. o Client Engagement & Solutioning: Work with multiple clients to understand requirements, design tailored AI solutions, develop proofs-of-concept (POCs), and ensure successful implementation in a professional services context. Solution Architecture & Design: o Design robust, scalable, and secure AI solution architectures across multi-cloud platforms and on-premise infrastructure. o Provide technical guidance and architectural oversight for AI initiatives, focusing on optimizing for GPU infrastructure . o Evaluate and recommend AI technologies, tools, and infrastructure, including Large Language Models (LLMs) and Small Language Models (SLMs) on cloud and on-premise. Team Leadership & Mentorship: o Lead, mentor, and grow a team of talented AI engineers, data scientists, and machine learning specialists. o Conduct code reviews and ensure adherence to coding standards and architectural principles. o Promote collaboration and knowledge sharing. Technical Expertise & Implementation: o Hands-on experience in developing and deploying Generative AI applications (e.g., LLMs, RAG, GraphRags , image generation, code generation), including Agentic AI and Model Context Protocol (MCP). o Proficiency with Agentic AI orchestration frameworks such as LangChain, LlamaIndex, and/or similar tools. o Experience with leading LLM providers and models including OpenAI, Llama, Anthropic, and others. o Familiarity with AI-powered tools and platforms such as Microsoft Copilot, GitHub Copilot etc. o Strong understanding of various machine learning models (deep learning, supervised, unsupervised, reinforcement learning). o Experience with large datasets, ensuring data quality, feature engineering, and efficient data processing for AI model training. o Deep understanding of GPU infrastructure, for AI model training or/ and inference. Qualifications Bachelor's or Master's degree in Computer Science, AI, ML, Data Science, or a related quantitative field. 8+ years in AI/ML development, with at least 3 years in a leadership or lead architect role. Mandatory: Proven experience in leading the development and deployment of AI-based products and solutions. Mandatory: Extensive hands-on experience with Generative AI models and frameworks (e.g., TensorFlow, PyTorch, Hugging Face, OpenAI APIs, etc.), including practical application of Agentic AI. Proficiency with Agentic AI orchestration frameworks such as LangChain, LlamaIndex, and/or similar tools. Experience in leveraging and integrating various LLM providers and models, including but not limited to OpenAI, Llama, and Anthropic. Familiarity with AI-powered development tools and platforms such as Microsoft Copilot, GitHub Copilot, and other code generation/assistance tools. Strong understanding of solution architecture principles for large-scale AI systems, including multi-cloud platforms and on-premise deployments. Mandatory: Exposure to and understanding of GPU infrastructure, especially NVIDIA, for AI workloads. Experience with Large Language Models (LLMs) and Small Language Models (SLMs) on both cloud and on-premise environments. Proficiency in programming languages such as Python, with strong software engineering fundamentals. Familiarity with MLOps practices, including model versioning, deployment, monitoring, and retraining. Mandatory: Demonstrated industry exposure to professional services, with a proven track record of working with multiple clients to solution requirements, conduct POCs, and understand enterprise-level needs. Excellent communication, interpersonal, and presentation skills, with the ability to articulate complex technical concepts to diverse audiences. Strong problem-solving abilities and a strategic mindset. What We Offer Opportunity to work on cutting-edge AI technologies and shape the future of our products. A collaborative and innovative work environment. Competitive salary and benefits package. Professional development and growth opportunities. The chance to make a significant impact on our business and our customers. If you are a passionate AI leader with a drive for innovation and a desire to build groundbreaking AI products, we encourage you to apply!

Posted 1 week ago

Apply

3.0 - 5.0 years

6 - 11 Lacs

Thiruvananthapuram

On-site

GlassDoor logo

Experience Required: 3-5 years of hands-on experience in full-stack development, system design, and supporting AI/ML data-driven solutions in a production environment. Key Responsibilities Implementing Technical Designs: Collaborate with architects and senior stakeholders to understand high-level designs and break them down into detailed engineering tasks. Implement system modules and ensure alignment with architectural direction. Cross-Functional Collaboration: Work closely with software developers, data scientists, and UI/UX teams to translate system requirements into working code. Clearly communicate technical concepts and implementation plans to internal teams. Stakeholder Support: Participate in discussions with product and client teams to gather requirements. Provide regular updates on development progress and raise flags early to manage expectations. System Development & Integration: Develop, integrate, and maintain components of AI/ML platforms and data-driven applications. Contribute to scalable, secure, and efficient system components based on guidance from architectural leads. Issue Resolution: Identify and debug system-level issues, including deployment and performance challenges. Proactively collaborate with DevOps and QA to ensure resolution. Quality Assurance & Security Compliance: Ensure that implementations meet coding standards, performance benchmarks, and security requirements. Perform unit and integration testing to uphold quality standards. Agile Execution: Break features into technical tasks, estimate efforts, and deliver components in sprints. Participate in sprint planning, reviews, and retrospectives with a focus on delivering value. Tool & Framework Proficiency: Use modern tools and frameworks in your daily workflow, including AI/ML libraries, backend APIs, front-end frameworks, databases, and cloud services, contributing to robust, maintainable, and scalable systems. Continuous Learning & Contribution: Keep up with evolving tech stacks and suggest optimizations or refactoring opportunities. Bring learnings from the industry into internal knowledge-sharing sessions. Proficiency in using AI-copilots for Coding: Adaptation to emerging tools and knowledge of prompt engineering to effectively use AI for day-to-day coding needs. Technical Skills Hands-on experience with Python-based AI/ML development using libraries such as TensorFlow , PyTorch , scikit-learn , or Keras . Hands-on exposure to self-hosted or managed LLMs , supporting integration and fine-tuning workflows as per system needs while following architectural blueprints. Practical implementation of NLP/CV modules using tools like SpaCy , NLTK , Hugging Face Transformers , and OpenCV , contributing to feature extraction, preprocessing, and inference pipelines. Strong backend experience using Django , Flask , or Node.js , and API development (REST or GraphQL). Front-end development experience with React , Angular , or Vue.js , with a working understanding of responsive design and state management. Development and optimization of data storage solutions , using SQL (PostgreSQL, MySQL) and NoSQL (MongoDB, Cassandra), with hands-on experience configuring indexes, optimizing queries, and using caching tools like Redis and Memcached . Working knowledge of microservices and serverless patterns , participating in building modular services, integrating event-driven systems, and following best practices shared by architectural leads. Application of design patterns (e.g., Factory, Singleton, Observer) during implementation to ensure code reusability, scalability, and alignment with architectural standards. Exposure to big data tools like Apache Spark , and Kafka for processing datasets. Familiarity with ETL workflows and cloud data warehouse , using tools such as Airflow , dbt , BigQuery , or Snowflake . Understanding of CI/CD , containerization (Docker), IaC (Terraform), and cloud platforms (AWS, GCP, or Azure). Implementation of cloud security guidelines , including setting up IAM roles , configuring TLS/SSL , and working within secure VPC setups, with support from cloud architects. Exposure to MLOps practices , model versioning, and deployment pipelines using MLflow , FastAPI , or AWS SageMaker . Configuration and management of cloud services such as AWS EC2 , RDS , S3 , Load Balancers , and WAF , supporting scalable infrastructure deployment and reliability engineering efforts. Personal Attributes Proactive Execution and Communication: Able to take architectural direction and implement it independently with minimal rework with regular communication with stakeholders Collaboration: Comfortable working across disciplines with designers, data engineers, and QA teams. Responsibility: Owns code quality and reliability, especially in production systems. Problem Solver: Demonstrated ability to debug complex systems and contribute to solutioning. Preferred Skills: Key : Python, Django, Django ORM, HTML, CSS, Bootstrap, JavaScript, jQuery, Multi-threading, Multi-processing, Database Design, Database Administration, Cloud Infrastructure, Data Science, self-hosted LLMs Qualifications Bachelor’s or Master’s degree in Computer Science, Information Technology, Data Science, or a related field. Relevant certifications in cloud or machine learning are a plus. Package: 6-11 LPA Job Types: Full-time, Permanent Pay: ₹600,000.00 - ₹1,100,000.00 per year Schedule: Day shift Monday to Friday

Posted 1 week ago

Apply

1.0 - 4.0 years

0 Lacs

Bengaluru, Karnataka, India

On-site

Linkedin logo

Title: Bioinformatician Date: 20 Jun 2025 Job Location: Bangalore Pay Grade Year of Experience: Job Title: Bioinformatician Job Location: Bangalore About Syngene: Syngene ( www.syngeneintl.com ) is an innovation-led contract research, development and manufacturing organization offering integrated scientific services from early discovery to commercial supply. At Syngene, safety is at the heart of everything we do personally and professionally. Syngene has placed safety at par with business performance with shared responsibility and accountability, including following safety guidelines, procedures, and SOPs, in letter and spirit Overall adherence to safe practices and procedures of oneself and the teams aligned Contributing to development of procedures, practices and systems that ensures safe operations and compliance to company’s integrity & quality standards Driving a corporate culture that promotes environment, health, and safety (EHS) mindset and operational discipline at the workplace at all times. Ensuring safety of self, teams, and lab/plant by adhering to safety protocols and following environment, health, and safety (EHS) requirements at all times in the workplace. Ensure all assigned mandatory trainings related to data integrity, health, and safety measures are completed on time by all members of the team including self Compliance to Syngene’ s quality standards at all times Hold self and their teams accountable for the achievement of safety goals Govern and Review safety metrics from time to time We are seeking a highly skilled and experienced computational biologist to join our team. The ideal candidate will have a proven track record in multi-omics data analysis. They will be responsible for integrative analyses and contributing to the development of novel computational approaches to uncover biological insights. Experience: 1-4 years Core Purpose of the Role To support data-driven biological research by performing computational analysis of omics data, and generating translational insights through bioinformatics tools and pipelines. Position Responsibilities Conduct comprehensive analyses of multi-omics datasets, including genomics, transcriptomics, proteomics, metabolomics, and epigenomics. Develop computational workflows to integrate various -omics data to generate inference and hypotheses for testing. Conduct differential expression and functional enrichment analyses. Implement and execute data processing workflows and automate the pipelines with best practices for version control, modularization, and documentation. Apply advanced multivariate data analysis techniques, including regression, clustering, and dimensionality reduction, to uncover patterns and relationships in large datasets. Collaborate with researchers, scientists, and other team members to translate computational findings into actionable biological insights. Educational Qualifications Master’s degree in bioinformatics. Mandatory Technical Skills Programming: Proficiency in Python for data analysis, visualization, and pipeline development. Multi-omics analysis: Proven experience in analyzing and integrating multi-omics datasets. Statistics: Knowledge of probability distributions, correlation analysis, and hypothesis testing. Data visualization: Strong understanding of data visualization techniques and tools (e.g., ggplot2, matplotlib, seaborn). Preferred Machine learning: Familiarity with AI/ML concepts Behavioral Skills Excellent communication skills Objective thinking Problem solving Proactivity Syngene Values All employees will consistently demonstrate alignment with our core values Excellence Integrity Professionalism Equal Opportunity Employer It is the policy of Syngene to provide equal employment opportunity (EEO) to all persons regardless of age, color, national origin, citizenship status, physical or mental disability, race, religion, creed, gender, sex, sexual orientation, gender identity and/or expression, genetic information, marital status, status with regard to public assistance, veteran status, or any other characteristic protected by applicable legislation or local law. In addition, Syngene will provide reasonable accommodations for qualified individuals with disabilities.

Posted 1 week ago

Apply

6.0 years

0 Lacs

Bengaluru, Karnataka, India

On-site

Linkedin logo

About The Role Have you ever wondered why it's taking so long for an earner to be matched to your trip, how the price is determined for your trip, or how an earner is picked from the many around you? If so, the Mobility Marketplace Health Science team is for you! The Marketplace Health Science team at Uber plays a pivotal role in monitoring marketplace performance, detecting issues in real time, and driving solutions through algorithmic and data-driven interventions. Our work is essential to maintaining Uber's market leadership and delivering reliable experiences to riders and earners. We are seeking experienced data scientists who thrive on solving complex problems at scale. The ideal candidate brings a strong foundation in causal inference, experimentation and analytics, along with a deep understanding of marketplace dynamics and metric trade-offs. What The Candidate Will Need / Bonus Points ---- What the Candidate Will Do ---- Refine ambiguous questions and generate new hypotheses about whether marketplace levers such as Rider and Driver Pricing, Matching, Surge etc are functioning appropriately through a deep understanding of the data, our customers, and our business. Define how our teams measure success by developing Key Performance Indicators and other user/business metrics, in close partnership with Product and other subject areas such as engineering, operations, and marketing. Collaborate with applied scientists and engineers to build and improve the availability, integrity, accuracy, and reliability of our models, tables etc. Design and develop algorithms to increase the speed and accuracy with which we react to marketplace changes. Develop data-driven business insights and work with cross-functional partners to find opportunities and recommend prioritization of product, growth, and optimization initiatives. Basic Qualifications Undergraduate and/or graduate degree in Math, Economics, Statistics, Engineering, Computer Science, or other quantitative fields. 6+ years of experience as a Data Scientist, Product Analyst, Senior Data Analyst, or other types of data analysis-focused functions. Deep understanding of core statistical concepts such as hypothesis testing, regression, and causal inference Advanced SQL expertise. Experience with either Python or R for data analysis. Knowledge of experimental design and analysis (A/B, Switchbacks, Synthetic Control, Diff in Diff, etc.). Experience with exploratory data analysis, statistical analysis and testing, and model development. Proven track record to wrangle large datasets, extract insights from data, and summarize learnings/takeaways. Experience with Excel and some dashboarding/data visualization (i.e., Tableau, Mixpanel, Looker, or similar). Preferred Qualifications Proven aptitude toward Data Storytelling and Root Cause Analysis using data. Excellent communication skills across technical, non-technical, and executive audiences. Have a growth mindset; love solving ambiguous, ambitious, and impactful problems. Ability to work in a self-guided manner. Ability to deliver on tight timelines and prioritize multiple tasks while maintaining quality and detail.

Posted 1 week ago

Apply

4.0 years

0 Lacs

Hyderabad, Telangana, India

Remote

Linkedin logo

About Media.net : Media.net is a leading, global ad tech company that focuses on creating the most transparent and efficient path for advertiser budgets to become publisher revenue. Our proprietary contextual technology is at the forefront of enhancing Programmatic buying, the latest industry standard in ad buying for digital platforms. The Media.net platform powers major global publishers and ad tech businesses at scale across ad formats like display, video, mobile, native, as well as search. Media.net’s U.S. HQ is based in New York, and the Global HQ is in Dubai. With office locations and consultant partners across the world. Media.net takes pride in the value-add it offers to its 50+ demand and 21K+ publisher partners, in terms of both products and services. Role : Software Development Engineer 2 Location : Hyderabad ( Remote ) What is the job like? As a Developer, you will contribute to the product engineering efforts of multiple areas of our intent discovery secret sauce and work with world-class R&D teams to develop game-changing search and text inference algorithms that will help millions of internet users discover what they are looking for. Our Search/Ad Platform involves myriad technologies, diverse platforms, complex algorithms and the latest application paradigms such as NoSQL databases, eventual consistency, distributed queues and are deployed across hundreds of servers in a super-scalable fashion where a 10ms delay in response time could mean the difference between success and failure. In this role, you will manage/work with a team of energized developers and will be responsible for the entire lifecycle of one or more areas, including architecture, design, coding, deployment etc. We believe that ‘code speaks louder than words and as such expect everyone at every level in the engineering team to be comfortable with rolling up their sleeves, firing up their favourite IDE and writing clean, testable and well-designed code. Who should apply for this role? 2–4 years of software development experience in Python Strong understanding and hands-on experience with deep learning frameworks such as PyTorch, including implementation of real-world projects Good knowledge of relational and NoSQL databases Ability to write complex and optimized SQL queries Solid programming fundamentals, including OOP, Design Patterns, and Data Structures Excellent analytical, logical, and problem-solving skills Familiarity with cloud platforms is a plus Experience with big data technologies like Spark, and Hive is a plus Ability to understand business requirements, work independently, and take full ownership of tasks Passionate and enthusiastic about building and maintaining large-scale, high-performance systems.

Posted 1 week ago

Apply

0 years

0 Lacs

Pune, Maharashtra, India

On-site

Linkedin logo

Summary Under limited supervision designs, develops and maintains test procedures, tester hardware and software for electronic circuit board production. ESSENTIAL DUTIES AND RESPONSIBILITIES include the following. Other duties may be assigned. Leadership And Management Responsibilities Recruitment and Retention: Recruit and interview Process Technicians. Communicate criteria to recruiters for Process Technician position candidates. Coach technicians in the interviewing/hiring process. Monitor team member turnover; identify key factors that can be improved; make improvements. Employee and Team Development: Identify individual and team strengths and development needs on an ongoing basis. Create and/or validate training curriculum in area of responsibility. Coach and mentor Process Technicians to deliver excellence to every internal and external customer. Performance Management: Establish clear measurable goals and objectives by which to determine individual and team results (i.e. operational metrics, results against project timelines, training documentation, attendance records, knowledge of operational roles and responsibilities, personal development goals). Solicit ongoing feedback from Assistant Test Engineering Manager, Workcell Manager (WCM), Business Unit Manager (BUM), peers and team member on team member’s contribution to the Workcell team. Provide ongoing coaching and counseling to team member based on feedback. Express pride in staff and encourage them to feel good about their accomplishments. Perform team member evaluations professionally and on time. Drive individuals and the team to continuously improve in key operational metrics and the achievement of the organizational goals. Coordinate activities of large teams and keep them focused in times of crises. Ensure recognition and rewards are managed fairly and consistently in area of responsibility. Communication: Provide communication forum for the exchange of ideas and information with the department. Organize verbal and written ideas clearly and use an appropriate business style. Ask questions; encourage input from team members. Assess communication style of individual team members and adapt own communication style accordingly. Technical Management Responsibilities Review circuit board designs for testability requirements. Support manufacturing with failure analysis, tester debugging, reduction of intermittent failures and downtime of test equipment. Prepare recommendations for testing and documentation of procedures to be used from the product design phase through to initial production. Generate reports and analysis of test data, prepares documentation and recommendations. Review test equipment designs, data and RMA issues with customers regularly. Design, and direct engineering and technical personnel in fabrication of testing and test control apparatus and equipment. Direct and coordinate engineering activities concerned with development, procurement, installation, and calibration of instruments, equipment, and control devices required to test, record, and reduce test data. Determine conditions under which tests are to be conducted and sequences and phases of test operations. Direct and exercise control over operational, functional, and performance phases of tests. Perform moderately complex assignments of the engineering test function for standard and/or custom devices. Analyze and interpret test data and prepares technical reports for use by test engineering and management personnel. Develop or use computer software and hardware to conduct tests on machinery and equipment. Perform semi-routine technique development and maintenance, subject to established Jabil standards, including ISO and QS development standards. Provide training in new procedures to production testing staff. Adhere to all safety and health rules and regulations associated with this position and as directed by supervisor. Comply and follow all procedures within the company security policy. Minimum Requirements Bachelors of Science in Electronics or Electrical Engineering from four-year college or university; and three to five years experience Language Skills Ability to read, analyze, and interpret general business periodicals, professional journals, technical procedures, or governmental regulations. Ability to write reports, business correspondence, and procedure manuals. Ability to effectively present information and respond to questions from groups of managers, clients, customers, and the general public. Mathematical Skills Ability to work with mathematical concepts such as probability and statistical inference, and fundamentals of plane and solid geometry and trigonometry. Ability to apply concepts such as fractions, percentages, ratios, and proportions to practical situations. REASONING ABILITY Ability to define problems, collect data, establish facts, and draw valid conclusions. Ability to interpret an extensive variety of technical instructions in mathematical or diagram form and deal with several abstract and concrete variables. PHYSICAL DEMANDS The physical demands described here are representative of those that must be met by an employee to successfully perform the essential functions of this job. The employee is frequently required to walk, and to lift and carry PC’s and test equipment weighing up to 50 lbs. Specific vision abilities required by this job include close vision and use of computer monitor screens a great deal of time. WORK ENVIRONMENT The work environment characteristics described here are representative of those an employee encounters while performing the essential functions of this job. Individual’s primary workstation is located in the office area, with some time spent each day on the manufacturing floor. The noise level in this environment ranges from low to moderate. , BE AWARE OF FRAUD: When applying for a job at Jabil you will be contacted via correspondence through our official job portal with a jabil.com e-mail address; direct phone call from a member of the Jabil team; or direct e-mail with a jabil.com e-mail address. Jabil does not request payments for interviews or at any other point during the hiring process. Jabil will not ask for your personal identifying information such as a social security number, birth certificate, financial institution, driver’s license number or passport information over the phone or via e-mail. If you believe you are a victim of identity theft, contact your local police department. Any scam job listings should be reported to whatever website it was posted in.

Posted 1 week ago

Apply

5.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Linkedin logo

We are seeking a passionate AI/ML Engineer to join our team in building the core AI-driven functionality an intelligent visual data encryption system. The role involves designing, training, and deploying AI models (e.g., CLIP, DCGANs, Decision Trees), integrating them into a secure backend, and operationalizing the solution via AWS cloud services and Python-based APIs. Key Responsibilities: AI/ML Development Design and train deep learning models for image classification and sensitivity tagging using CLIP, DCGANs, and Decision Trees. Build synthetic datasets using DCGANs for balancing. Fine-tune pre-trained models for customized encryption logic. Implement explainable classification logic for model outputs. Validate model performance using custom metrics and datasets. API Development Design and develop Python RESTful APIs using FastAPI or Flask for: Image upload and classification Model inference endpoints Encryption trigger calls Integrate APIs with AWS Lambda and Amazon API Gateway. AWS Integration Deploy and manage AI models on Amazon SageMaker for training and real-time inference. Use AWS Lambda for serverless backend compute. Store encrypted image data on Amazon S3 and metadata on Amazon RDS (PostgreSQL). Use AWS Cognito for secure user authentication and KMS for key management. Monitor job status via CloudWatch and enable secure, scalable API access. Required Skills & Experience: Must-Have 3–5 years of experience in AI/ML (especially vision-based systems). Strong experience with PyTorch or TensorFlow for model development. Proficient in Python with experience building RESTful APIs. Hands-on experience with Amazon SageMaker, Lambda, API Gateway, and S3. Knowledge of OpenSSL/PyCryptodome or basic cryptographic concepts. Understanding of model deployment, serialization, and performance tuning. Nice-to-Have Experience with CLIP model fine-tuning. Familiarity with Docker, GitHub Actions, or CI/CD pipelines. Experience in data classification under compliance regimes (e.g., GDPR, HIPAA). Familiarity with multi-tenant SaaS design patterns. Tools & Technologies: Python, PyTorch, TensorFlow FastAPI, Flask AWS: SageMaker, Lambda, S3, RDS, Cognito, API Gateway, KMS Git, Docker, Postgres, OpenCV, OpenSSL

Posted 1 week ago

Apply

6.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Linkedin logo

Job Title: AI Engineer Location: Gurgaon (On-site) Type: Full-Time Experience: 2–6 Years Role Overview We are seeking a hands-on AI Engineer to architect and deploy production-grade AI systems that power our real-time voice intelligence suite. You will lead AI model development, optimize low-latency inference pipelines, and integrate GenAI, ASR, and RAG systems into scalable platforms. This role combines deep technical expertise with team leadership and a strong product mindset. Key Responsibilities Build and deploy ASR models (e.g., Whisper, Wav2Vec2.0) and diarization systems for multi-lingual, real-time environments. Design and optimize GenAI pipelines using OpenAI, Gemini, LLaMA, and RAG frameworks (LangChain, LlamaIndex). Architect and implement vector database systems (FAISS, Pinecone, Weaviate) for knowledge retrieval and indexing. Fine-tune LLMs using SFT, LoRA, RLHF, and craft effective prompt strategies for summarization and recommendation tasks. Lead AI engineering team members and collaborate cross-functionally to ship robust, high-performance systems at scale. Preferred Qualification 2–6 years of experience in AI/ML, with demonstrated deployment of NLP, GenAI, or STT models in production. Proficiency in Python, PyTorch/TensorFlow, and real-time architectures (WebSockets, Kafka). Strong grasp of transformer models, MLOps, and low-latency pipeline optimization. Bachelor’s/Master’s in CS, AI/ML, or related field from a reputed institution (IITs, BITS, IIITs, or equivalent). What We Offer Compensation: Competitive salary + equity + performance bonuses Ownership: Lead impactful AI modules across voice, NLP, and GenAI Growth: Work with top-tier mentors, advanced compute resources, and real-world scaling challenges Culture: High-trust, high-speed, outcome-driven startup environment

Posted 1 week ago

Apply

4.0 - 6.0 years

0 Lacs

New Delhi, Delhi, India

On-site

Linkedin logo

About Knowdis.ai Knowdis.ai is an AI-first company specializing in e-commerce applications. We harness the power of machine learning and AI to enhance e-commerce operations, optimize customer experiences, and drive growth. If you are passionate about AI-driven product innovation, this is the perfect opportunity to make a meaningful impact. Key Responsibilities: - Infrastructure Management: Build scalable and robust infrastructure for ML models, ensuring seamless production integration. - CI/CD Expertise: Develop and maintain CI/CD pipelines with a focus on ML model deployment. - Model Deployment and Monitoring: Deploy ML models using TensorFlow Serving, Pytorch Serving, Triton Inference Server, or TensorRT and monitor their performance in production. - Collaboration: Work closely with data scientists and software engineers to transition ML models from research to production. - Security and Compliance: Uphold security protocols and ensure regulatory compliance in ML systems. Skills and Experience Required: - Proficiency in Docker and Kubernetes for containerization and orchestration. - Experience with CI/CD pipeline development and maintenance. - Experience in deploying ML models using TensorFlow Serving, Pytorch Serving, Triton Inference Server, and TensorRT. - Experience with cloud platforms like AWS, Azure, and GCP. - Strong problem-solving, communication, and teamwork skills. Qualifications: - Bachelor’s/Master’s degree in Computer Science, Engineering, or a related field. - 4-6 years of experience in ML project management, with a recent focus on MLOps. Additional Competencies: - AI Technologies Deployment, Data Engineering, IT Performance, Scalability Testing, and Security Practices. SELECTION PROCESS: Interested Candidates are mandatorily required to apply through this listing on Jigya. Only applications received through Jigya will be evaluated further. Shortlisted candidates may be required to appear in an Online Assessment administered by Jigya on behalf of the Client. Candidates selected after the screening test will be interviewed by the Client

Posted 1 week ago

Apply

0 years

0 Lacs

Chennai, Tamil Nadu, India

On-site

Linkedin logo

Roles & Responsibilities Design, implement, and train deep learning models for: Text-to-Speech (e.g., SpeechT5, StyleTTS2, YourTTS, XTTS-v2 or similar models) Voice Cloning with speaker embeddings (x-vectors, d-vectors), few-shot adaptation, prosody and emotion transfer Engineer multilingual audio-text preprocessing pipelines: Text normalization, grapheme-to-phoneme (G2P) conversion, Unicode normalization (NFC/NFD) Silence trimming, VAD-based audio segmentation, audio enhancement for noisy corpora, speech prosody modification and waveform manipulation Build scalable data loaders using PyTorch for: Large-scale, multi-speaker datasets with variable-length sequences and chunked streaming Extract and process acoustic features: Log-mel spectrograms, pitch contours, MFCCs, energy, speaker embeddings Optimize training using: Mixed precision (FP16/BFloat16), gradient checkpointing, label smoothing, quantization-aware training Build serving infrastructure for inference using: TorchServe, ONNX Runtime, Triton Inference Server, FastAPI (for REST endpoints), including batch and real-time modes Optimize models for production: Quantization, model pruning, ONNX conversion, parallel decoding, GPU/CPU memory profiling Create automated and human evaluation logics: MOS, PESQ, STOI, BLEU, WER/CER, multi-speaker test sets, multilingual subjective listening tests Implement ethical deployment safeguards: Digital watermarking, impersonation detection, and voice verification for cloned speech Conduct literature reviews and reproduce state-of-the-art papers; adapt and improve on open benchmarks Mentor junior contributors, review code, and maintain shared research and model repositories Collaborate across teams (MLOps, backend, product, linguists) to translate research into deployable, user-facing solutions Required Skills Advanced proficiency in Python and PyTorch (TensorFlow a plus) Strong grasp of deep learning concepts: Sequence-to-sequence models, Transformers, autoregressive and non-autoregressive decoders, attention mechanisms, VAEs, GANs Experience with modern speech processing toolkits: ESPnet, NVIDIA NeMo, Coqui TTS, OpenSeq2Seq, or equivalent Design custom loss function for custom models based on: Mel loss, GAN loss, KL divergence, attention losses, etc.,, learning rate schedules, training stability Hands-on experience with multilingual and low-resource language modeling Understanding of transformer architecture, LLMs and working with existing AI models, tools and APIs Model serving & API integration: TorchServe, FastAPI, Docker, ONNX Runtime Preferred (Bonus) Skills CUDA kernel optimization, custom GPU operations, memory footprint profiling Experience deploying on AWS/GCP with GPU acceleration Experience developing RESTful APIs for real-time TTS/voice cloning endpoints Publications or open-source contributions in TTS, ASR, or speech processing Working knowledge of multilingual translation pipelines Knowledge of speaker diarization, voice anonymization, and speech synthesis for agglutinative/morphologically rich languages Milestones & Expectations (First 3–6 Months) Deliver at least one production-ready TTS or Voice Cloning model integrated with India Speaks’ Dubbing Studio or SaaS APIs Create a fully reproducible experiment pipeline for multilingual speech modeling, complete with model cards and performance benchmarks Contribute to custom evaluation tools for measuring quality across Indian languages Deploy optimized models to live staging environments using Triton, TorchServe, or ONNX Demonstrate impact through real-world integration in education, media, or defence deployments

Posted 1 week ago

Apply

50.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Linkedin logo

Data Axle Inc. has been an industry leader in data, marketing solutions, sales, and research for over 50 years in the USA. Data Axle now has an established strategic global centre of excellence in Pune. This centre delivers mission critical data services to its global customers powered by its proprietary cloud-based technology platform and by leveraging proprietary business and consumer databases. Data Axle India is recognized as a Great Place to Work! This prestigious designation is a testament to our collective efforts in fostering an exceptional workplace culture and creating an environment where every team member can thrive. Roles & Responsibilities We are looking for Associate Data Scientist to join the Data Science Client Services team to continue our success of identifying high quality target audiences that generate profitable marketing return for our clients. We are looking for experienced data science, machine learning and MLOps practitioners to design, build and deploy impactful predictive marketing solutions that serve a wide range of verticals and clients. The right candidate will enjoy contributing to and learning from a highly talented team and working on a variety of projects. Ownership of design, implementation, and deployment of machine learning algorithms in a modern Python-based cloud architecture. Design or enhance ML workflows for data ingestion, model design, model inference and scoring. Oversight on team project execution and delivery. Establish peer review guidelines for high quality coding to help develop junior team members' skill set growth, cross-training, and team efficiencies. Visualize and publish model performance results and insights to internal and external audiences.

Posted 1 week ago

Apply

3.0 years

0 Lacs

Mumbai, Maharashtra, India

Remote

Linkedin logo

At AryaXAI , we’re building the future of explainable, scalable, and aligned AI —designed specifically for high-stakes environments where trust, transparency, and performance are non-negotiable. From financial services to energy and other regulated industries, our platform powers intelligent decision-making through safe and robust AI systems. We’re looking for a Data Scientist with a deep understanding of both classical and deep learning techniques, experience building enterprise-scale ML pipelines, and the ambition to tackle real-world, high-impact problems. You will work at the intersection of modeling, infrastructure, and regulatory alignment—fine-tuning models that must be auditable, performant, and production-ready. Responsibilities: Modeling & AI Development Design, build, and fine-tune machine learning models (both classical and deep learning) for complex mission-critical use cases in domains like banking, finance, energy, etc. Work on supervised, unsupervised, and semi-supervised learning problems using structured, unstructured, and time-series data. Fine-tune foundation models for specialized use cases requiring high interpretability and performance. Platform Integration Develop and deploy models on AryaXAI’s platform to serve real-time or batch inference needs. Leverage explainability tools (e.g., DLBacktrace, SHAP, LIME, or AryaXAI’s native xai_evals stack) to ensure transparency and regulatory compliance. Design pipelines for data ingestion, transformation, model training, evaluation, and deployment using MLOps best practices. Enterprise AI Architecture Collaborate with product and engineering teams to implement scalable and compliant ML pipelines across cloud and hybrid environments. Contribute to designing secure, modular AI workflows that meet enterprise needs—latency, throughput, auditability, and policy constraints. Ensure models meet strict regulatory and ethical requirements (e.g., bias mitigation, traceability, explainability). Requirements : 3+ years of experience building ML systems in production, ideally in regulated or enterprise environments. Strong proficiency in Python , with experience in libraries like scikit-learn, XGBoost, PyTorch, TensorFlow , or similar. Experience with end-to-end model lifecycle : from data preprocessing and feature engineering to deployment and monitoring. Deep understanding of enterprise ML architecture —model versioning, reproducibility, CI/CD for ML, and governance. Experience working with regulatory, audit, or safety constraints in data science or ML systems. Familiarity with ML Ops tools (MLflow, SageMaker, Vertex AI, etc.) and cloud platforms (AWS, Azure, GCP). Strong communication skills and an ability to translate technical outcomes into business impact. Bonus Points For Prior experience in regulated industries : banking, insurance, energy, or critical infrastructure. Experience with time-series modeling , anomaly detection, underwriting, fraud detection or risk scoring systems. Knowledge of RAG architectures , generative AI , or foundation model fine-tuning . Exposure to privacy-preserving ML , model monitoring , and bias mitigation frameworks. What You’ll Get Competitive compensation with performance-based upside Comprehensive health coverage for you and your family Opportunity to work on mission-critical AI systems where your models drive real-world decisions Ownership of core components in a platform used by top-tier enterprises Career growth in a fast-paced, high-impact startup environment Remote-first, collaborative, and high-performance team culture If you’re excited to build data science solutions that truly matter , especially in the most demanding industries, we want to hear from you.

Posted 1 week ago

Apply

5.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Linkedin logo

At Roche you can show up as yourself, embraced for the unique qualities you bring. Our culture encourages personal expression, open dialogue, and genuine connections, where you are valued, accepted and respected for who you are, allowing you to thrive both personally and professionally. This is how we aim to prevent, stop and cure diseases and ensure everyone has access to healthcare today and for generations to come. Join Roche, where every voice matters. The Position A healthier future. That’s what drives us. We are looking for a highly skilled Artificial Intelligence (AI) / Machine Learning (ML) Engineer with expertise in building AI-powered applications. We will be building AI & GenAI solutions end-to-end: from concept, through prototyping, production, to operations. The Opportunity: Generative AI Application Development: Collaborate with developers and stakeholders in Agile teams to integrate LLMs and classical AI techniques into end-user applications, focusing on user experience, and real-time performance Algorithm Development: Design, develop, customize, optimize, and fine-tune LLM-based and other AI-infused algorithms tailored to specific use cases such as text generation, summarization, information extraction, chatbots, AI agents, code generation, document analysis, sentiment analysis, data analysis, etc LLM Fine-Tuning and Customization: Fine-tune pre-trained LLMs to specific business needs, leveraging prompt engineering, transfer learning, and few-shot techniques to enhance model performance in real-world scenarios End-to-End Pipeline Development: Build and maintain production-ready end-to-end ML pipelines, including data ingestion, preprocessing, training, evaluation, deployment, and monitoring; automate workflows using MLOps best practices to ensure scalability and efficiency Performance Optimization: Optimize model inference speed, reduce latency, and manage resource usage across cloud services and GPU/TPU architectures Scalable Model Deployment: Collaborate with other developers to deploy models at scale, using cloud-based infrastructure (AWS, Azure) and ensuring high availability and fault tolerance Monitoring and Maintenance: Implement continuous monitoring and refining strategies for deployed models, using feedback loops and e.g. incremental fine-tuning to ensure ongoing accuracy and reliability; address drifts and biases as they arise Software Development: Apply software development best practices, including writing unit tests, configuring CI/CD pipelines, containerizing applications, prompt engineering and setting up APIs; ensure robust logging, experiment tracking, and model monitoring Who are: Minimum overall 5-7 years of experience and hold B.Sc., B.Eng., M.Sc., M.Eng., Ph.D. or D.Eng. in Computer Science or equivalent degree Experience: 3+ years of experience in AI/ML engineering, with exposure to both classical machine learning methods and language model-based applications Technical Skills: Advanced proficiency in Python and experience with deep learning frameworks such as PyTorch or TensorFlow; expertise with Transformer architectures; hands-on experience with LangChain or similar LLM frameworks; experience with designing end-to-end RAG systems using state of the art orchestration frameworks (hands on experience with fine-tuning LLMs for specific tasks and use cases considered as an additional advantage) MLOps Knowledge: Strong understanding of MLOps tools and practices, including version control, CI/CD pipelines, containerization, orchestration, Infrastructure as Code, automated deployment Deployment: Experience in deploying LLM and other AI models with cloud platforms (AWS, Azure) and machine learning workbenches for robust and scalable productizations Practical overview and experience with AWS services to design cloud solutions, familiarity with Azure is a plus; experience with working with GenAI specific services like Azure OpenAI, Amazon Bedrock, Amazon SageMaker JumpStart, etc. Data Engineering: Expertise in working with structured and unstructured data, including data cleaning, feature engineering with data stores like vector, relational, NoSQL databases and data lakes through APIs Model Evaluation and Metrics: Proficiency in evaluating both classical ML models and LLMs using relevant metrics Relocation benefits are not available for this posting. Who we are A healthier future drives us to innovate. Together, more than 100’000 employees across the globe are dedicated to advance science, ensuring everyone has access to healthcare today and for generations to come. Our efforts result in more than 26 million people treated with our medicines and over 30 billion tests conducted using our Diagnostics products. We empower each other to explore new possibilities, foster creativity, and keep our ambitions high, so we can deliver life-changing healthcare solutions that make a global impact. Let’s build a healthier future, together. Roche is an Equal Opportunity Employer.

Posted 1 week ago

Apply

0 years

0 Lacs

Noida, Uttar Pradesh, India

On-site

Linkedin logo

Senior Gen AI Engineer Job Description Brightly Software is seeking an experienced candidate to join our Product team in the role of Gen AI engineer to drive best in class client-facing AI features by creating and delivering insights that advise client decisions tomorrow. Role As a Gen AI Engineer , you will play a critical role in building AI offerings for Brightly. Y ou will partner with our various software Product teams to drive client facing insights to inform smarter decisions faster . This will include the following: Lead the evaluation and selection of foundation models and vector databases based on performance and business needs Design and implement applications powered by generative AI (e.g., LLMs, diffusion models), delivering contextual and actionable insights for clients. Establish best practices and documentation for prompt engineering, model fine-tuning, and evaluation to support cross-domain generative AI use cases. Build, test, and deploy generative AI applications using standard tools and frameworks for model inference, embeddings, vector stores, and orchestration pipelines. Key Responsibilities Guide the design of multi-step RAG, agentic, or tool-augmented workflows Implement governance, safety layers, and responsible AI practices (e.g., guardrails, moderation, auditability) Mentor junior engineers and review GenAI design and implementation plans Drive experimentation, benchmarking, and continuous improvement of GenAI capabilities Collaborate with leadership to align GenAI initiatives with product and business strategy Build and optimize Retrieval-Augmented Generation (RAG) pipelines using vector stores like Pinecone, FAISS, or AWS Opensearch Perform exploratory data analysis (EDA), data cleaning, and feature engineering to prepare data for model building. Design, develop, train, and evaluate machine learning models (e.g., classification, regression, clustering, natural language processing) with strong exerience in predictive and stastical modelling. Implement and deploy machine learning models into production using AWS services, with a strong focus on Amazon SageMaker (e.g., SageMaker Studio, training jobs, inference endpoints, SageMaker Pipelines). Understanding and development of state management workflows using Langraph . Develop GenAI applications using Hugging Face Transformers, LangChain , and Llama related frameworks Engineer and evaluate prompts, including prompt chaining and output quality assessment Apply NLP and transformer model expertise to solve language tasks Deploy GenAI models to cloud platforms (preferably AWS) using Docker and Kubernetes Monitor and optimize model and pipeline performance for scalability and efficiency Communicate techn

Posted 1 week ago

Apply

2.0 years

0 Lacs

Noida, Uttar Pradesh, India

On-site

Linkedin logo

Gen AI Engineer Job Description Brightly Software is seeking a high performer to join our Product team in the role of Gen AI engineer to drive best in class client - facing AI features by creating and delivering insights that advise client decisions tomorrow. Role As a Gen AI Engineer , you will play a critical role in building AI offering s for Brightly. Y ou will partner with our various software Product teams to drive client facing insights to inform smarter decisions faster . This will include the following: Design and implement applications powered by generative AI (e.g., LLMs, diffusion models), delivering contextual and actionable insights for clients. Establish best practices and documentation for prompt engineering, model fine-tuning, and evaluation to support cross-domain generative AI use cases. Build, test, and deploy generative AI applications using standard tools and frameworks for model inference, embeddings, vector stores, and orchestration pipelines. Key Responsibilities Build and optimize Retrieval-Augmented Generation (RAG) pipelines using vector stores like Pinecone, FAISS, or AWS OpenSearch D evelop GenAI applications using Hugging Face Transformers, LangChain , and Llama related frameworks Perform exploratory data analysis (EDA), data cleaning, and feature engineering to prepare data for model building. Design, develop, train, and evaluate machine learning models (e.g., classification, regression, clustering, natural language processing) with strong ex erience in predictive and stastical modelling . Implement and deploy machine learning models into production using AWS services, with a strong focus on Amazon SageMaker (e.g., SageMaker Studio, training jobs, inference endpoints, SageMaker Pipelines). Understanding and development of state management workflows using Langraph . Engineer and evaluate prompts, including prompt chaining and output quality assessment Apply NLP and transformer model expertise to solve language tasks Deploy GenAI models to cloud platforms (preferably AWS) using Docker and Kubernetes Monitor and optimize model and pipeline performance for scalability and efficiency Communicate technical concepts clearly to cross-functional and non-technical stakeholders Thrive in a fast-paced, lean environment and contribute to scalable GenAI system design Qualifications Bachelor’s degree is required 2-4 years of experience of total experience with a strong focus on AI and ML and 1+ years in core GenAI Engineer ing Demonstrated expertise in working with large language models (LLMs) and generative AI systems, including both text-based and multimodal models. S trong programming skills in Python, including proficiency with data science libraries such as NumPy, Pandas, Scikit-learn, TensorFlow, and/or PyTorch . Familiarity with MLOps principles and tools for automating and streamlining the ML lifecycle. Experience working with agentic AI . Capable of building Retrieval-Augmented Generation (RAG) pipelines leveraging vector stores like Pinecone, Chroma, or FAISS. St rong programming skills in Python, with experience using leading AI/ML libraries such as Hugging Face Transformers and LangChain . Practical experience in working with vector databases and embedding methodologies for efficient information retrieval. P ossess experience in developing and exposing API endpoints for accessing AI model capabilities using frameworks like FastAPI . Knowledgeable in prompt engineering techniques, including prompt chaining and performance evaluation strategies . Solid grasp of natural language processing (NLP) fundamentals and transformer-based model architectures. Experience in deploying machine learning models to cloud platforms (preferably AWS) and containerized environments using Docker or Kubernetes. Skilled in fine-tuning and assessing open-source models using methods such as LoRA , PEFT, and supervised training. Strong communication skills with the ability to convey complex technical concepts to non-technical stakeholders. Able to operate successfully in a lean, fast-paced organization, and to create a vision and organization that can scale quickly Senior Gen AI Engineer

Posted 1 week ago

Apply

4.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Linkedin logo

Job Title: Data Scientist – AIML, GenAI & Agentic AI Location: Pune/ Bangalore/ Indore/ Kolkata Job Type: Full-time Experience Level: 4+ Years NP : Immediate Joiner OR15 Days Max Job Description We are seeking a highly skilled and innovative Data Scientist / AI Engineer with deep expertise in AI/ML, Generative AI, and Agentic AI frameworks to join our advanced analytics and AI team. The ideal candidate will possess a robust background in data science and machine learning, along with hands-on experience in building and deploying end-to-end intelligent systems using modern AI technologies including RAG (Retrieval-Augmented Generation), LLMs, and agent orchestration tools. Key Responsibilities Design, build, and deploy machine learning models and Generative AI solutions for a wide range of use cases (text, vision, and tabular data). Develop and maintain AI/ML pipelines for large-scale training and inference in production environments. Leverage frameworks such as LangChain, LangGraph, CrewAI for building Agentic AI workflows. Fine-tune and prompt-engineer LLMs (e.g., GPT, BERT) for enterprise-grade RAG and NLP solutions. Collaborate with business and engineering teams to translate business problems into AI/ML models that deliver measurable value. Apply advanced analytics techniques such as regression, classification, clustering, sequence modeling, association rules, computer vision, and NLP. Architect and implement scalable AI solutions using Python , PyTorch , TensorFlow , and cloud-native technologies. Ensure integration of AI solutions within existing enterprise architecture using containerized services and orchestration (e.g., Docker, Kubernetes). Maintain documentation and present insights and technical findings to stakeholders. Required Skills and Qualifications Bachelor’s/Master’s/PhD in Computer Science, Data Science, Statistics, or related field. Strong proficiency in Python and libraries such as Pandas, NumPy, Scikit-learn, etc. Extensive experience with deep learning frameworks : PyTorch and TensorFlow. Proven experience with Generative AI , LLMs , RAG , BERT , and related architectures. Familiarity with LangChain , LangGraph , and CrewAI and strong knowledge of agent orchestration and autonomous workflows. Experience with large-scale ML pipelines , MLOps practices, and cloud platforms (AWS, GCP, or Azure). Deep understanding of software engineering principles , design patterns, and enterprise architecture. Strong problem-solving, analytical thinking, and debugging skills. Excellent communication, presentation, and cross-functional collaboration abilities. Preferred Qualifications Experience in fine-tuning LLMs and optimizing prompt engineering techniques. Publications, open-source contributions, or patents in AI/ML/NLP/GenAI. Experience with vector databases and tools such as Pinecone, FAISS, Weaviate, or Milvus. Why Join Us? Work on cutting-edge AI/ML and GenAI innovations. Collaborate with top-tier scientists, engineers, and product teams. Opportunity to shape the next generation of intelligent agents and enterprise AI solutions. Flexible work arrangements and continuous learning culture. To Apply: Please submit your resume and portfolio of relevant AI/ML work (e.g., GitHub, papers, demos) to Shanti.upase@calsoftinc.com

Posted 1 week ago

Apply

5.0 years

0 Lacs

India

On-site

Linkedin logo

Role Summary We’re hiring a Founding Full-Stack AI/ML Engineer to help build and scale the backbone of our AI system. You’ll lead development across agent orchestration, tool execution, Model Context Protocol (MCP), API integration, and browser-based research workflows. You’ll work closely with the founder on hands-on roadmap development, rapid prototyping, and fast iteration cycles to evolve the product quickly based on real user needs. Responsibilities Build multi-agent systems capable of reasoning, tool use, and autonomous action Implement Model Context Protocol (MCP) strategies to manage complex, multi-source context Integrate third-party APIs (e.g., Crunchbase, PitchBook, CB Insights), scraping APIs, and data aggregators Develop browser-based agents enhanced with computer vision for dynamic research, scraping, and web interaction Optimize inference pipelines, task planning, and system performance Collaborate on architecture, prototyping, and iterative development Experiment with prompt chaining, tool calling, embeddings, and vector search Requirements 5+ years of experience in software engineering or AI/ML development Strong Python skills and experience with LangChain, LlamaIndex, or agentic frameworks Proven experience with multi-agent systems, tool calling, or task planning agents Familiarity with Model Context Protocol (MCP), Retrieval-Augmented Generation (RAG), and multi-modal context handling Experience with browser automation frameworks (e.g., Playwright, Puppeteer, Selenium) Cloud deployment and systems engineering experience (GCP, AWS, etc.) Self-starter attitude with strong product sense and iteration speed Bonus Points Experience with AutoGen, CrewAI, OpenAgents, or ReAct-style frameworks Background in building AI systems that blend structured and unstructured data Experience working in a fast-paced startup environment Previous startup or technical founding team experience This is a unique opportunity to work directly with an industry leader in AI to build a cutting-edge, next-generation AI system from the ground up.

Posted 1 week ago

Apply

8.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Linkedin logo

Java Full Stack Developer_Full-Time_Hyderabad, 5 Days WFO Job Title: Java Full Stack Developer Job Type: Full-Time Experience: 8 to 10 Years Location: Hyderabad, 5 Days WFO Must Have: Java Full Stack Developer with React & React Native Job Description: Must-Have Skills: • professionals with 6 to 10 years of industry experience, preferably with a background in product startups. • 5+ years of hands-on experience in Java Spring Boot and Microservices architecture • Strong proficiency in React.js and React Native (web & mobile development) • Al/ML knowledge — using pre-trained models for inference • Solid experience with MySQL and PostgreSQL — data modelling and query optimization • Expertise in MongoDB and handling document-based data • Familiar with Kafka (producer & consumer) and event-driven systems, WebRTC, WebSocket protocols. • Experience deploying on AWS Cloud, EC2, S3, RDS, EKS/Kubernetes • Cl/CD implementation experience • Must have proven experience in building scalable products and infrastructure on Video driven platforms Good to Have: • API Gateway experience (Kong Konnect or similar) • Exposure to Video Analytics or Computer Vision • Experience in building mobile apps from scratch • Familiarity with Low-code and Agentic workflow platforms • Previous startup experience is a big plus!

Posted 1 week ago

Apply

0.0 - 12.0 years

0 Lacs

Bengaluru, Karnataka

On-site

Indeed logo

Company Overview: Schneider Electric is a global leader in energy management and automation, committed to providing innovative solutions that ensure Life Is On everywhere, for everyone, and at every moment. We are part of small and medium buildings league, We are expanding our team in Gurugram and looking for a Principal Architect to enhance our Edge applications and drive the roadmap for next generations IOT gateways. Job Description: We are looking for an experienced Principal Architect - IoT Edge to lead the design and deployment of edge middleware for smart building environments. This role focuses on architecting robust, secure, and scalable IoT gateway-based systems that enable real-time monitoring, control, and optimization of micro-BMS, including HVAC, lighting, and energy. As a Principal Architect at Schneider Electric, you will play a crucial role in developing and implementing IoT solutions across our global infrastructure, with a primary focus on Edge software. This position requires a blend of strategic architectural design and practical hands-on ability to implement and manage, and optimize edge middleware solutions, ensuring efficient data processing for a large-scale edge gateways and devices (100s of thousands) deployed in the field. Some of the core services supported by IoT gateways aims at providing various services and features such as Secure firmware update Log management. Product configuration (identity, network connectivity, date/time…) Service/message bus (for intra and inter service communication) Controls logic to control and schedule downstream devices Device management, application management and connectivity to the “Cloud system” Edge intelligence – e.g. data buffering, computing metrics on edge Dockerised services Local Web Interface Connectivity protocols (MQTT, Zigbee, Wi-Fi, LoRaWAN, Modbus, BacNet …) Key Responsibilities: Provide architecture guidelines, identify technical solutions, and write technical requirements, answering to the functional requirement of the SMB BMS solution. Architect and develop scalable, high-performance Edge computing solutions for IoT applications. Work closely with POs and solution Architects of SMB- Building Activate platform to ensure proper landing of the middleware features and services Develop and optimize IoT data pipelines, integrating sensors, edge devices, and cloud-based platforms. Collaborate with cross-functional teams to define edge computing strategies, system architectures, and best practices. Work on device-to-cloud communication using MQTT(s), HTTP(s), WebSockets, or other messaging protocols. Ensure software is secure, reliable, and optimized for resource-constrained edge environments. Design and optimize Linux-based networking for edge devices, including network configuration, VPNs, firewalls, and traffic shaping. Implement and manage Linux process management, including systemd services, resource allocation, and performance tuning for IoT applications. Conduct code reviews, mentor junior developers, and provide technical leadership in edge software development. Implement edge analytics and AI/ML inference for predictive maintenance, energy optimization, and occupant comfort. Lead PoCs and pilot deployments in commercial, industrial, or mixed-use buildings. Requirements: Technical 10 – 12 years of overall experience in software engineering with a strong focus on IoT based firmware development Understanding of BACnet/Modbus protocols. Familiarity with cloud IoT platforms (AWS IoT, Azure IoT, Google Cloud IoT) and their integration with edge devices Strong knowledge of Linux networking, including TCP/IP, DNS, firewalls (iptables/nftables), VPNs, and network security. Experience in Linux process management, including systemd, resource limits (cgroups), and performance tuning. Good Understanding of IoT architectures, protocols (MQTT, HTTP/REST), and edge computing frameworks. Hands-on experience with Docker. Proficiency and Experience with Git or any other VCS. Excellent problem-solving skills and the ability to lead complex technical projects. Proficiency in edge programming (Python, GoLang, Rust, Java or C++) Knowledge of cybersecurity best practices for IOT environments. Good to have: Experience with digital twins, building energy modelling, or occupancy analytics. Expertise in Python, with experience in asynchronous programming, task processing frameworks, and Web frameworks Soft Skills: Excellent problem-solving abilities and strong communication skills. Advanced verbal and written communication skills including the ability to explain and present technical concepts to a diverse set of audiences. Comfortable working directly with both technical and non-technical audiences Good judgment, time management, and decision-making skills Strong teamwork and interpersonal skills; ability to communicate and thrive in a cross-functional environment Willingness to work outside documented job description. Has a “whatever is needed” attitude. Qualifications Preferred Qualifications: Bachelor's or Master's degree in computer science, Information Technology, or related field. Working experience on designing robust, scalable & maintainable IOT gateway applications Prior experience in building cloud connected Edge IoT solutions. Prior experience in the energy sector or industrial automation is advantageous. Primary Location : IN-Karnataka-Bangalore Schedule : Full-time Unposting Date : Ongoing

Posted 1 week ago

Apply

5.0 years

0 Lacs

Bengaluru, Karnataka

On-site

Indeed logo

- 8+ years of specific technology domain areas (e.g. software development, cloud computing, systems engineering, infrastructure, security, networking, data & analytics) experience - 5+ years of design, implementation, or consulting in applications and infrastructures experience - 10+ years of IT development or implementation/consulting in the software or Internet industries experience Sales, Marketing and Global Services (SMGS) AWS Sales, Marketing, and Global Services (SMGS) is responsible for driving revenue, adoption, and growth from the largest and fastest growing small- and mid-market accounts to enterprise-level customers including public sector. Do you like startups? Are you interested in Cloud Computing & Generative AI? Yes? We have a role you might find interesting. Startups are the large enterprises of the future. These young companies are founded by ambitious people who have a desire to build something meaningful and to challenge the status quo. To address underserved customers, or to challenge incumbents. They usually operate in an environment of scarcity: whether that’s capital, engineering resource, or experience. This is where you come in. The Startup Solutions Architecture team is dedicated to working with these early stage startup companies as they build their businesses. We’re here to make sure that they can deploy the best, most scalable, and most secure architectures possible – and that they spend as little time and money as possible doing so. We are looking for technical builders who love the idea of working with early stage startups to help them as they grow. In this role, you’ll work directly with a variety of interesting customers and help them make the best (and sometimes the most pragmatic) technical decisions along the way. You’ll have a chance to build enduring relationships with these companies and establish yourself as a trusted advisor. As well as spending time working directly with customers, you’ll also get plenty of time to “sharpen the saw” and keep your skills fresh. We have more than 175 services across a range of different categories and it’s important that we can help startups take advantages of the right ones. You’ll also play an important role as an advocate with our product teams to make sure we are building the right products for the startups you work with. And for the customers you don’t get to work with on a 1:1 basis you’ll get the chance to share your knowledge more broadly by working on technical content and presenting at events. A day in the life You’re surrounded by innovation. You’re empowered with a lot of ownership. Your growth is accelerated. The work is challenging. You have a voice here and are encouraged to use it. Your experience and career development is in your hands. We live our leadership principles every day. At Amazon, it's always "Day 1". Diverse Experiences Amazon values diverse experiences. Even if you do not meet all of the preferred qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Why AWS Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why we strive for flexibility as part of our working culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (gender diversity) conferences, inspire us to never stop embracing our uniqueness. Mentorship and Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional. 5+ years of infrastructure architecture, database architecture and networking experience. Knowledge of AWS services, market segments, customer base and industry verticals Experience working with end user or developer communities. Experience in developing and deploying large scale machine learning, Agentic AI systems and/or systems into production. Experience scaling model training and inference using technologies like Slurm, ParallelCluster, Amazon SageMaker Hands-on experience benchmarking and optimizing performance of models on accelerated computing (GPU, TPU, AI ASICs) clusters with high-speed networking. Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process, including support for the interview or onboarding process, please visit https://amazon.jobs/content/en/how-we-hire/accommodations for more information. If the country/region you’re applying in isn’t listed, please contact your Recruiting Partner.

Posted 1 week ago

Apply

0 years

7 - 11 Lacs

Prayagraj, Uttar Pradesh, India

On-site

Linkedin logo

Institute of Information Science Postdoctoral Researcher 2 Person The Computer Systems Laboratory - Machine Learning Systems Team Focuses On Research Areas Including Parallel And Distributed Computing, Compilers, And Computer Architecture. We Aim To Leverage Computer System Technologies To Accelerate The Inference And Training Of Deep Learning Models And Develop Optimizations For Next-generation AI Models. Our Research Emphasizes The Following Job Description Unit Institute of Information Science JobTitle Postdoctoral Researcher 2 Person Work Content Research on Optimization of Deep Learning Model Inference and Training AI Model Compression and Optimization Model Compression Techniques (e.g., Pruning And Quantization) Reduce The Size And Computational Demands Of AI Models, Which Are Crucial For Resource-constrained Platforms Such As Embedded Systems And Memory-limited AI Accelerators. We Aim To Explore AI compiler: deployment methods for compressed models across servers, edge devices, and heterogeneous systems. High performance computing: efficient execution of compressed models on processors with advanced AI extensions, e.g., Intel AVX512, ARM SVE, RISC-V RVV, and tensor-level accelerations on GPUs and NPUs. AI Accelerator Design We aim to design AI accelerators for accelerating AI model inference, focusing on software and hardware co-design and co-optimization. Optimization of AI Model Inference in Heterogeneous Environments Computer Architectures Are Evolving Toward Heterogeneous Multi-processor Designs (e.g., CPUs + GPUs + AI Accelerators). Integrating Heterogeneous Processors To Execute Complex Models (e.g., Hybrid Models, Multi-models, And Multi-task Models) With High Computational Efficiency Poses a Critical Challenge. We Aim To Explore Efficient scheduling algorithms. Parallel algorithms for the three dimensions: data parallelism, model parallelism, and tensor parallelism. Qualifications Ph.D. degree in Computer Science, Computer Engineering, or Electrical Engineering Experience in parallel computing and parallel programming (CUDA or OpenCL, C/C++ programming) or hardware design (Verilog or HLS) Proficient in system and software development Candidates With The Following Experience Will Be Given Priority Experience in deep learning platforms, including PyTorch, TensorFlow, TVM, etc. Experience in high-performance computing or embedded systems. Experience in algorithm designs. Knowledge of compilers or computer architecture Working Environment Operating Hours 8:30AM-5:30PM Work Place Institute of Information Science, Academia Sinica Treatment According to Academia Sinica standards: Postdoctoral Researchers: NT$64,711-99,317/month. Benefits include: labor and healthcare insurance, and year-end bonuses. Reference Site 洪鼎詠網頁: http://www.iis.sinica.edu.tw/pages/dyhong/index_zh.html, 吳真貞網頁: http://www.iis.sinica.edu.tw/pages/wuj/index_zh.html Please Email Your CV (including Publications, Projects, And Work Experience), Transcripts (undergraduate And Above), And Any Other Materials That May Assist In The Review Process To The Following PIs Acceptance Method Contacts Dr. Ding-Yong Hong Contact Address Room 818, New IIS Building, Academia Sinica Contact Telephone 02-27883799 ext. 1818 Email dyhong@iis.sinica.edu.tw Required Documents Dr. Ding-Yong Hong: dyhong@iis.sinica.edu.tw Dr. Jan-Jan Wu: wuj@iis.sinica.edu.tw Precautions for application Date Publication Date 2025-01-20 Expiration Date 2025-12-31

Posted 1 week ago

Apply

0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Linkedin logo

This job is with Swiss Re, an inclusive employer and a member of myGwork – the largest global platform for the LGBTQ+ business community. Please do not contact the recruiter directly. About The Team And Our Scope We are a forward-thinking tech organization within Swiss Re, delivering transformative AI/ML solutions that redefine how businesses operate. Our mission is to build intelligent, secure, and scalable systems that deliver real-time insights, automation, and high-impact user experiences to clients globally. You'll join a high-velocity AI/ML team working closely with product managers, architects, and engineers to create next-gen enterprise-grade solutions. Our team is built on a startup mindset — bias to action, fast iterations, and ruthless focus on value delivery. We’re not only shaping the future of AI in business — we’re shaping the future of talent. This role is ideal for someone passionate about advanced AI engineering today and curious about evolving into a product leadership role tomorrow. You'll get exposure to customer discovery, roadmap planning, and strategic decision-making alongside your technical contributions. Role Overview As an AI/ML Engineer, you will play a pivotal role in the research, development, and deployment of next-generation GenAI and machine learning solutions . Your scope will go beyond retrieval-augmented generation (RAG) to include areas such as prompt engineering, long-context LLM orchestration, multi-modal model integration (voice, text, image, PDF), and agent-based workflows. You will help assess trade-offs between RAG and context-native strategies, explore hybrid techniques, and build intelligent pipelines that blend structured and unstructured data. You’ll work with technologies such as LLMs, vector databases, orchestration frameworks, prompt chaining libraries, and embedding models, embedding intelligence into complex, business-critical systems. This role sits at the intersection of rapid GenAI prototyping and rigorous enterprise deployment, giving you hands-on influence over both the technical stack and the emerging product direction. Key Responsibilities Build Next-Gen GenAI Pipelines: Design, implement, and optimize pipelines across RAG, prompt engineering, long-context input handling, and multi-modal processing. Prototype, Validate, Deploy: Rapidly test ideas through PoCs, validate performance against real-world business use cases, and industrialize successful patterns. Ingest, Enrich, Embed: Construct ingestion workflows including OCR, chunking, embeddings, and indexing into vector databases to unlock unstructured data. Integrate Seamlessly: Embed GenAI services into critical business workflows, balancing scalability, compliance, latency, and observability. Explore Hybrid Strategies: Combine RAG with context-native models, retrieval mechanisms, and agentic reasoning to build robust hybrid architectures. Drive Impact with Product Thinking: Collaborate with product managers and UX designers to shape user-centric solutions and understand business context. Ensure Enterprise-Grade Quality: Deliver solutions that are secure, compliant (e.g., GDPR), explainable, and resilient — especially in regulated environments. What Makes You a Fit Must-Have Technical Expertise Proven experience with GenAI techniques and LLMs, including RAG, long-context inference, prompt tuning, and multi-modal integration. Strong hands-on skills with Python, embedding models, and orchestration libraries (e.g., LangChain, Semantic Kernel, or equivalents). Comfort with MLOps practices, including version control, CI/CD pipelines, model monitoring, and reproducibility. Ability to operate independently, deliver iteratively, and challenge assumptions with data-driven insight. Understanding of vector search optimization and retrieval tuning. Exposure to multi-modal models Nice-To-Have Qualifications Experience building and operating AI systems in regulated industries (e.g., insurance, finance, healthcare). Familiarity with Azure AI ecosystem (e.g., Azure OpenAI, Azure AI Document Intelligence, Azure Cognitive Search) and deployment practices in cloud-native environments. Experience with agentic AI architectures, tools like AutoGen, or prompt chaining frameworks. Familiarity with data privacy and auditability principles in enterprise AI. Bonus: You Think Like a Product Manager While this role is technical at its core, we highly value candidates who are curious about how AI features become products . If you’re excited by the idea of influencing roadmaps, shaping requirements, or owning end-to-end value delivery — we’ll give you space to grow into it. This is a role where engineering and product are not silos . If you’re keen to move in that direction, we’ll mentor and support your evolution. Why Join Us? You’ll be part of a team that’s pushing AI/ML into uncharted, high-value territory. We operate with urgency, autonomy, and deep collaboration. You’ll prototype fast, deliver often, and see your work shape real-world outcomes — whether in underwriting, claims, or data orchestration. And if you're looking to transition from deep tech to product leadership , this role is a launchpad. Swiss Re is an equal opportunity employer . We celebrate diversity and are committed to creating an inclusive environment for all employees. About Swiss Re Swiss Re is one of the world’s leading providers of reinsurance, insurance and other forms of insurance-based risk transfer, working to make the world more resilient. We anticipate and manage a wide variety of risks, from natural catastrophes and climate change to cybercrime. Combining experience with creative thinking and cutting-edge expertise, we create new opportunities and solutions for our clients. This is possible thanks to the collaboration of more than 14,000 employees across the world. Our success depends on our ability to build an inclusive culture encouraging fresh perspectives and innovative thinking. We embrace a workplace where everyone has equal opportunities to thrive and develop professionally regardless of their age, gender, race, ethnicity, gender identity and/or expression, sexual orientation, physical or mental ability, skillset, thought or other characteristics. In our inclusive and flexible environment everyone can bring their authentic selves to work and their passion for sustainability. If you are an experienced professional returning to the workforce after a career break, we encourage you to apply for open positions that match your skills and experience. swissre_footer { position: relative; margin-top: -50px; height: 30px; clear: both; margin-bottom: 20px; background: #EEE none repeat scroll 0% 0%; line-height: 30px; padding: 0px 10px; color: #AAA; font-family: "Arial,Helvetica,sans-serif"; } .swissre_jobtemplate { width: 970px; max-width: 100%; height: auto; } .jobDisplay .job { font-family: "Arial" !important; font-size: 12px !important; } .joqReqDescription { max-width: 100%; height: auto; align: center; } .joqReqDescription ul { width: 787px; max-width: 100%; } .joqReqDescription p { width: 827px; max-width: 100%; } Keywords Reference Code: 134317

Posted 1 week ago

Apply

0 years

25 - 30 Lacs

Mangaluru, Karnataka, India

On-site

Linkedin logo

About The Opportunity We’re a deep-tech innovator at the intersection of Artificial Intelligence, machine-learning infrastructure, and edge-to-cloud platforms . Our award-winning solutions let Fortune-500 enterprises build, train, and deploy large-scale AI models—seamlessly, securely, and at lightning speed. As global demand for generative AI, RAG pipelines, and autonomous agents accelerates, we’re scaling our MLOps team to keep our customers two steps ahead of the curve. Role & Responsibilities (max 6) Own the full MLOps stack—design, build, and harden GPU-accelerated Kubernetes clusters across on-prem DCs and AWS/GCP/Azure for model training, fine-tuning, and low-latency inference. Automate everything: craft IaC modules (Terraform/Pulumi) and CI/CD pipelines that deliver zero-downtime releases and reproducible experiment tracking. Ship production-grade LLM workloads—optimize RAG/agent pipelines, manage model registries, and implement self-healing workflow orchestration with Kubeflow/Flyte/Prefect. Eliminate bottlenecks: profile CUDA, resolve driver mismatches, and tune distributed frameworks (Ray, DeepSpeed) for multi-node scale-out. Champion reliability: architect HA data lakes, databases, ingress/egress, DNS, and end-to-end observability (Prometheus/Grafana) targeting 99.99 % uptime. Mentor & influence: instill platform-first mind-set, codify best practices, and report progress/road-blocks directly to senior leadership. Skills & Qualifications (max 6) Must-Have 5 + yrs DevOps/Platform experience with Docker & Kubernetes; expert bash/Python/Go scripting. Hands-on building ML infrastructure for distributed GPU training and scalable model serving. Deep fluency in cloud services (EKS/GKE/AKS), networking, load-balancing, RBAC, and Git-based CI/CD. Proven mastery of IaC & config-management (Terraform, Pulumi, Ansible). Preferred Production experience with LLM fine-tuning, RAG architectures, or agentic workflows at scale. Exposure to Kubeflow, Flyte, Prefect, or Ray; track record of setting up observability and data-lake pipelines (Delta Lake, Iceberg). Skills: cloud services,containerization,automation tools,version control,devops

Posted 1 week ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies