Home
Jobs

1317 Inference Jobs - Page 15

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Description When you attract people who have the DNA of pioneers and the DNA of explorers, you build a company of like-minded people who want to invent. And that’s what they think about when they get up in the morning: how are we going to work backwards from customers and build a great service or a great product” – Jeff Bezos Amazon.com’s success is built on a foundation of customer obsession. Have you ever thought about what it takes to successfully deliver millions of packages to Amazon customers seamlessly every day like a clock work? In order to make that happen, behind those millions of packages, billions of decision gets made by machines and humans. What is the accuracy of customer provided address? Do we know exact location of the address on Map? Is there a safe place? Can we make unattended delivery? Would signature be required? If the address is commercial property? Do we know open business hours of the address? What if customer is not home? Is there an alternate delivery address? Does customer have any special preference? What are other addresses that also have packages to be delivered on the same day? Are we optimizing delivery associate’s route? Does delivery associate know locality well enough? Is there an access code to get inside building? And the list simply goes on. At the core of all of it lies quality of underlying data that can help make those decisions in time. The person in this role will be a strong influencer who will ensure goal alignment with Technology, Operations, and Finance teams. This role will serve as the face of the organization to global stakeholders. This position requires a results-oriented, high-energy, dynamic individual with both stamina and mental quickness to be able to work and thrive in a fast-paced, high-growth global organization. Excellent communication skills and executive presence to get in front of VPs and SVPs across Amazon will be imperative. Key Strategic Objectives: Amazon is seeking an experienced leader to own the vision for quality improvement through global address management programs. As a Business Intelligence Engineer of Amazon last mile quality team, you will be responsible for shaping the strategy and direction of customer-facing products that are core to the customer experience. As a key member of the last mile leadership team, you will continually raise the bar on both quality and performance. You will bring innovation, a strategic perspective, a passionate voice, and an ability to prioritize and execute on a fast-moving set of priorities, competitive pressures, and operational initiatives. You will partner closely with product and technology teams to define and build innovative and delightful experiences for customers. You must be highly analytical, able to work extremely effectively in a matrix organization, and have the ability to break complex problems down into steps that drive product development at Amazon speed. You will set the tempo for defect reduction through continuous improvement and drive accountability across multiple business units in order to deliver large scale high visibility/ high impact projects. You will lead by example to be just as passionate about operational performance and predictability as you will be about all other aspects of customer experience. The Successful Candidate Will Be Able To Effectively manage customer expectations and resolve conflicts that balance client and company needs. Develop process to effectively maintain and disseminate project information to stakeholders. Be successful in a delivery focused environment and determining the right processes to make the team successful. This opportunity requires excellent technical, problem solving, and communication skills. The candidate is not just a policy maker/spokesperson but drives to get things done. Possess superior analytical abilities and judgment. Use quantitative and qualitative data to prioritize and influence, show creativity, experimentation and innovation, and drive projects with urgency in this fast-paced environment. Partner with key stakeholders to develop the vision and strategy for customer experience on our platforms. Influence product roadmaps based on this strategy along with your teams. Support the scalable growth of the company by developing and enabling the success of the Operations leadership team. Serve as a role model for Amazon Leadership Principles inside and outside the organization Actively seek to implement and distribute best practices across the operation Devise and implement efficient and secure procedures for data management and analysis with attention to all technical aspects Create and enforce policies for effective data management Formulate management techniques for quality data collection to ensure adequacy, accuracy and legitimacy of data Establish rules and procedures for data sharing with upper management, external stakeholders etc. Basic Qualifications Knowledge of SQL and Excel Experience hiring and leading a high-performance team Knowledge of data engineering pipelines, cloud solutions, ETL management, databases, visualizations and analytical platforms Knowledge of methods for statistical inference (e.g. regression, experimental design, significance testing) Preferred Qualifications Knowledge of product experimentation (A/B testing) Knowledge of a scripting language (Python, R, etc.) Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process, including support for the interview or onboarding process, please visit https://amazon.jobs/content/en/how-we-hire/accommodations for more information. If the country/region you’re applying in isn’t listed, please contact your Recruiting Partner. Company - ADCI HYD 13 SEZ Job ID: A2974488 Show more Show less

Posted 2 weeks ago

Apply

0 years

0 Lacs

Hyderabad, Telangana, India

Remote

We are united in our mission to make a positive impact on healthcare. Join Us! South Florida Business Journal, Best Places to Work 2024 Inc. 5000 Fastest-Growing Private Companies in America 2024 2024 Black Book Awards, ranked #1 EHR in 11 Specialties 2024 Spring Digital Health Awards, “Web-based Digital Health” category for EMA Health Records (Gold) 2024 Stevie American Business Award (Silver), New Product and Service: Health Technology Solution (Klara) Who We Are We Are Modernizing Medicine (WAMM)! We’re a team of bright, passionate, and positive problem-solvers on a mission to place doctors and patients at the center of care through an intelligent, specialty-specific cloud platform. Our vision is a world where the software we build increases medical practice success and improves patient outcomes. Founded in 2010 by Daniel Cane and Dr. Michael Sherling, we have grown to over 3400 combined direct and contingent team members serving eleven specialties, and we are just getting started! ModMed is based in Boca Raton, FL, with office locations in Santiago, Chile, Berlin, Germany, Hyderabad, India, and a robust remote workforce with team members across the US. ModMed is hiring a driven ML Ops Engineer 2 to join our positive, passionate, and high-performing team focused on scalable ML Systems. This is an exciting opportunity to You as you will collaborate with data scientists, engineers, and other cross-functional teams to ensure seamless model deployment, monitoring, and automation. If you're passionate about cloud infrastructure, automation, and optimizing ML pipelines, this is the role for you within a fast-paced Healthcare IT company that is truly Modernizing Medicine! Key Responsibilities Model Deployment & Automation: Develop, deploy, and manage ML models on Databricks using MLflow for tracking experiments, managing models, and registering them in a centralized repository. Infrastructure & Environment Management: Set up scalable and fault-tolerant infrastructure to support model training and inference in cloud environments such as AWS, GCP, or Azure. Monitoring & Performance Optimization: Implement monitoring systems to track model performance, accuracy, and drift over time. Create automated systems for re-training and continuous learning to maintain optimal performance. Data Pipeline Integration: Collaborate with the data engineering team to integrate model pipelines with real-time and batch data processing frameworks, ensuring seamless data flow for training and inference. Skillset & Qualification Model Deployment: Experience with deploying models in production using cloud platforms like AWS Sagemaker, GCP AI Platform, or Azure ML Studio. Version Control & Automation: Experience with MLOps tools such as MLflow, Kubeflow, or Airflow to automate and monitor the lifecycle of machine learning models. Cloud Expertise: Experience with cloud-based machine learning services on AWS, Google Cloud, or Azure, ensuring that models are scalable and efficient. Engineers must be skilled in measuring and optimizing model performance through metrics like AUC, precision, recall, and F1-score, ensuring that models are robust and reliable in production settings. Education: Bachelor’s or Master’s degree in Data Science, Statistics, Mathematics, or a related technical field. ModMed In India Benefit Highlights High growth, collaborative, transparent, fun, and award-winning culture Comprehensive benefits package including medical for you, your family, and your dependent parents The company supported community engagement opportunities along with a paid Voluntary Time Off day to use for volunteering in your community of interest Global presence, and in-person collaboration opportunities; dog-friendly HQ (US), Hybrid office-based roles and remote availability Company-sponsored Employee Resource Groups that provide engaged and supportive communities within ModMed ModMed Benefits Highlight: At ModMed, we believe it’s important to offer a competitive benefits package designed to meet the diverse needs of our growing workforce. Eligible Modernizers can enroll in a wide range of benefits: India Meals & Snacks: Enjoy complimentary office lunches & dinners on select days and healthy snacks delivered to your desk, Insurance Coverage: Comprehensive health, accidental, and life insurance plans, including coverage for family members, all at no cost to employees, Allowances: Annual wellness allowance to support your well-being and productivity, Earned, casual, and sick leaves to maintain a healthy work-life balance, Bereavement leave for difficult times and extended medical leave options, Paid parental leaves, including maternity, paternity, adoption, surrogacy, and abortion leave, Celebration leave to make your special day even more memorable, and company-paid holidays to recharge and unwind. United States Comprehensive medical, dental, and vision benefits, including a company Health Savings Account contribution, 401(k): ModMed provides a matching contribution each payday of 50% of your contribution deferred on up to 6% of your compensation. After one year of employment with ModMed, 100% of any matching contribution you receive is yours to keep. Generous Paid Time Off and Paid Parental Leave programs, Company paid Life and Disability benefits, Flexible Spending Account, and Employee Assistance Programs, Company-sponsored Business Resource & Special Interest Groups that provide engaged and supportive communities within ModMed, Professional development opportunities, including tuition reimbursement programs and unlimited access to LinkedIn Learning, Global presence and in-person collaboration opportunities; dog-friendly HQ (US), Hybrid office-based roles and remote availability for some roles, Weekly catered breakfast and lunch, treadmill workstations, Zen, and wellness rooms within our BRIC headquarters. PHISHING SCAM WARNING: ModMed is among several companies recently made aware of a phishing scam involving imposters posing as hiring managers recruiting via email, text and social media. The imposters are creating misleading email accounts, conducting remote "interviews," and making fake job offers in order to collect personal and financial information from unsuspecting individuals. Please be aware that no job offers will be made from ModMed without a formal interview process, and valid communications from our hiring team will come from our employees with a ModMed email address (first.lastname@modmed.com). Please check senders’ email addresses carefully. Additionally, ModMed will not ask you to purchase equipment or supplies as part of your onboarding process. If you are receiving communications as described above, please report them to the FTC website. Show more Show less

Posted 2 weeks ago

Apply

8.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Job Title: Senior Python Developer – Backend Engineering Company: Darwix AI Location: Gurgaon (On-site) Type: Full-Time Experience Required: 4–8 Years About Darwix AI Darwix AI is building India’s most advanced GenAI-powered platform for enterprise sales teams. We combine speech recognition, LLMs, vector databases, real-time analytics, and multilingual intelligence to power customer conversations across India, the Middle East, and Southeast Asia. We’re solving complex backend problems across speech-to-text pipelines , agent assist systems , AI-based real-time decisioning , and scalable SaaS delivery . Our engineering team sits at the core of our product and works closely with AI research, product, and client delivery to build the future of revenue enablement. Backed by top-tier VCs, AI advisors, and enterprise clients, this is a chance to build something foundational. Role Overview We are hiring a Senior Python Developer to architect, implement, and optimize high-performance backend systems that power our AI platform. You will take ownership of key backend services—from core REST APIs and data pipelines to complex integrations with AI/ML modules. This role is for builders. You’ll work closely with product, AI, and infra teams, write production-grade Python code, lead critical decisions on architecture, and help shape engineering best practices. Key Responsibilities 1. Backend API Development Design and implement scalable, secure RESTful APIs using FastAPI , Flask , or Django REST Framework Architect modular services and microservices to support AI, transcription, real-time analytics, and reporting Optimize API performance with proper indexing, pagination, caching, and load management strategies Integrate with frontend systems, mobile clients, and third-party systems through clean, well-documented endpoints 2. AI Integrations & Inference Orchestration Work closely with AI engineers to integrate GenAI/LLM APIs (OpenAI, Llama, Gemini), transcription models (Whisper, Deepgram), and retrieval-augmented generation (RAG) workflows Build services to manage prompt templates, chaining logic, and LangChain flows Deploy and manage vector database integrations (e.g., FAISS , Pinecone , Weaviate ) for real-time search and recommendation pipelines 3. Database Design & Optimization Model and maintain relational databases using MySQL or PostgreSQL ; experience with MongoDB is a plus Optimize SQL queries, schema design, and indexes to support low-latency data access Set up background jobs for session archiving, transcript cleanup, and audio-data binding 4. System Architecture & Deployment Own backend deployments using GitHub Actions , Docker , and AWS EC2 Ensure high availability of services through containerization, horizontal scaling, and health monitoring Manage staging and production environments, including DB backups, server health checks, and rollback systems 5. Security, Auth & Access Control Implement robust authentication (JWT, OAuth), rate limiting , and input validation Build role-based access controls (RBAC) and audit logging into backend workflows Maintain compliance-ready architecture for enterprise clients (data encryption, PII masking) 6. Code Quality, Documentation & Collaboration Write clean, modular, extensible Python code with meaningful comments and documentation Build test coverage (unit, integration) using PyTest , unittest , or Postman/Newman Participate in pull requests, code reviews, sprint planning, and retrospectives with the engineering team Required Skills & QualificationsTechnical Expertise 3–8 years of experience in backend development with Python, PHP. Strong experience with FastAPI , Flask , or Django (at least one in production-scale systems) Deep understanding of RESTful APIs , microservice architecture, and asynchronous Python patterns Strong hands-on with MySQL (joins, views, stored procedures); bonus if familiar with MongoDB , Redis , or Elasticsearch Experience with containerized deployment using Docker and cloud platforms like AWS or GCP Familiarity with Git , GitHub , CI/CD pipelines , and Linux-based server environments Show more Show less

Posted 2 weeks ago

Apply

0 years

0 Lacs

Noida, Uttar Pradesh, India

On-site

Years of exp : 10 - 15 yrs Location : Noida Join us as Cloud Engineer at Dailoqa , where you will be responsible for operationalizing cutting-edge machine learning and generative AI solutions, ensuring scalable, secure, and efficient deployment across infrastructure. You will work closely with data scientists, ML engineers, and business stakeholders to build and maintain robust MLOps pipelines, enabling rapid experimentation and reliable production implementation of AI models, including LLMs and real-time analytics systems. To be successful as Cloud Engineer you should have experience with: Cloud sourcing, networks, VMs, performance, scaling, availability, storage, security, access management Deep expertise in one or more cloud platforms: AWS, Azure, GCP Strong experience in containerization and orchestration (Docker, Kubernetes, Helm) Familiarity with CI/CD tools: GitHub Actions, Jenkins, Azure DevOps, ArgoCD, etc. Proficiency in scripting languages (Python, Bash, PowerShell) Knowledge of MLOps tools such as MLflow, Kubeflow, SageMaker, Vertex AI, or Azure ML Strong understanding of DevOps principles applied to ML workflows. Key Responsibilities may include: Design and implement scalable, cost-optimized, and secure infrastructure for AI-driven platforms. Implement infrastructure as code using tools like Terraform, ARM, or Cloud Formation. Automate infrastructure provisioning, CI/CD pipelines, and model deployment workflows. Ensure version control, repeatability, and compliance across all infrastructure components. Set up monitoring, logging, and alerting frameworks using tools like Prometheus, Grafana, ELK, or Azure Monitor. Optimize performance and resource utilization of AI workloads including GPU-based training/inference Experience with Snowflake, Databricks for collaborative ML development and scalable data processing. Understanding model interpretability, responsible AI, and governance. Contributions to open-source MLOps tools or communities. Strong leadership, communication, and cross-functional collaboration skills. Knowledge of data privacy, model governance, and regulatory compliance in AI systems. Exposure to LangChain, Vector DBs (e. g. , FAISS, Pinecone), and retrieval-augmented generation (RAG) pipelines. Show more Show less

Posted 2 weeks ago

Apply

12.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Job Title: VP-Digital Expert Support Lead Experience : 12 + Years Location : Pune Position Overview The Digital Expert Support Lead is a senior-level leadership role responsible for ensuring the resilience, scalability, and enterprise-grade supportability of AI-powered expert systems deployed across key domains like Wholesale Banking, Customer Onboarding, Payments, and Cash Management . This role requires technical depth, process rigor, stakeholder fluency , and the ability to lead cross-functional squads that ensure seamless operational performance of GenAI and digital expert agents in production environments. The candidate will work closely with Engineering, Product, AI/ML, SRE, DevOps, and Compliance teams to drive operational excellence and shape the next generation of support standards for AI-driven enterprise systems. Role-Level Expectations Functionally accountable for all post-deployment support and performance assurance of digital expert systems. Operates at L3+ support level , enabling L1/L2 teams through proactive observability, automation, and runbook design. Leads stability engineering squads , AI support specialists, and DevOps collaborators across multiple business units. Acts as the bridge between operations and engineering , ensuring technical fixes feed into product backlog effectively. Supports continuous improvement through incident intelligence, root cause reporting, and architecture hardening . Sets the support governance framework (SLAs/OLAs, monitoring KPIs, downtime classification, recovery playbooks). Position Responsibilities Operational Leadership & Stability Engineering Own the production health and lifecycle support of all digital expert systems across onboarding, payments, and cash management. Build and govern the AI Support Control Center to track usage patterns, failure alerts, and escalation workflows. Define and enforce SLAs/OLAs for LLMs, GenAI endpoints, NLP components, and associated microservices. Establish and maintain observability stacks (Grafana, ELK, Prometheus, Datadog) integrated with model behavior. Lead major incident response and drive cross-functional war rooms for critical recovery. Ensure AI pipeline resilience through fallback logic, circuit breakers, and context caching. Review and fine-tune inference flows, timeout parameters, latency thresholds, and token usage limits. Engineering Collaboration & Enhancements Drive code-level hotfixes or patches in coordination with Dev, QA, and Cloud Ops. Implement automation scripts for diagnosis, log capture, reprocessing, and health validation. Maintain well-structured GitOps pipelines for support-related patches, rollback plans, and enhancement sprints. Coordinate enhancement requests based on operational analytics and feedback loops. Champion enterprise integration and alignment with Core Banking, ERP, H2H, and transaction processing systems. Governance, Planning & People Leadership Build and mentor a high-caliber AI Support Squad – support engineers, SREs, and automation leads. Define and publish support KPIs , operational dashboards, and quarterly stability scorecards. Present production health reports to business, engineering, and executive leadership. Define runbooks, response playbooks, knowledge base entries, and onboarding plans for newer AI support use cases. Manage relationships with AI platform vendors, cloud ops partners, and application owners. Must-Have Skills & Experience 12+ years of software engineering, platform reliability, or AI systems management experience. Proven track record of leading support and platform operations for AI/ML/GenAI-powered systems . Strong experience with cloud-native platforms (Azure/AWS), Kubernetes , and containerized observability . Deep expertise in Python and/or Java for production debugging and script/tooling development. Proficient in monitoring, logging, tracing, and alerts using enterprise tools (Grafana, ELK, Datadog). Familiarity with token economics , prompt tuning, inference throttling, and GenAI usage policies. Experience working with distributed systems, banking APIs, and integration with Core/ERP systems . Strong understanding of incident management frameworks (ITIL) and ability to drive postmortem discipline . Excellent stakeholder management, cross-functional coordination, and communication skills. Demonstrated ability to mentor senior ICs and influence product and platform priorities. Nice-to-Haves Exposure to enterprise AI platforms like OpenAI, Azure OpenAI, Anthropic, or Cohere. Experience supporting multi-tenant AI applications with business-driven SLAs. Hands-on experience integrating with compliance and risk monitoring platforms. Familiarity with automated root cause inference or anomaly detection tooling. Past participation in enterprise architecture councils or platform reliability forums Show more Show less

Posted 2 weeks ago

Apply

6.0 years

0 Lacs

Gurugram, Haryana, India

On-site

Job Title: AI Engineer Location: Gurgaon (On-site) Type: Full-Time Experience: 2–6 Years Role Overview We are seeking a hands-on AI Engineer to architect and deploy production-grade AI systems that power our real-time voice intelligence suite. You will lead AI model development, optimize low-latency inference pipelines, and integrate GenAI, ASR, and RAG systems into scalable platforms. This role combines deep technical expertise with team leadership and a strong product mindset. Key Responsibilities Build and deploy ASR models (e.g., Whisper, Wav2Vec2.0) and diarization systems for multi-lingual, real-time environments. Design and optimize GenAI pipelines using OpenAI, Gemini, LLaMA, and RAG frameworks (LangChain, LlamaIndex). Architect and implement vector database systems (FAISS, Pinecone, Weaviate) for knowledge retrieval and indexing. Fine-tune LLMs using SFT, LoRA, RLHF, and craft effective prompt strategies for summarization and recommendation tasks. Lead AI engineering team members and collaborate cross-functionally to ship robust, high-performance systems at scale. Preferred Qualification 2–6 years of experience in AI/ML, with demonstrated deployment of NLP, GenAI, or STT models in production. Proficiency in Python, PyTorch/TensorFlow, and real-time architectures (WebSockets, Kafka). Strong grasp of transformer models, MLOps, and low-latency pipeline optimization. Bachelor’s/Master’s in CS, AI/ML, or related field from a reputed institution (IITs, BITS, IIITs, or equivalent). What We Offer Compensation: Competitive salary + equity + performance bonuses Ownership: Lead impactful AI modules across voice, NLP, and GenAI Growth: Work with top-tier mentors, advanced compute resources, and real-world scaling challenges Culture: High-trust, high-speed, outcome-driven startup environment Show more Show less

Posted 2 weeks ago

Apply

2.0 years

0 Lacs

Dwarka, Delhi, Delhi

Remote

Job Description: Embedded AI Engineer Location: Dwarka Sector 12 Delhi Job Type: Full-Time / Contract / Freelance Job Description: We are seeking a skilled and motivated Embedded AI Engineer to develop and optimize a voice assistant system on resource-constrained devices like the ESP32-S3. You will be responsible for implementing wake word detection, voice activity detection (VAD), and basic speech command recognition using Espressif's ESP-SR framework, I2S microphones, and embedded ML models. And also, they have experience in computer vision-based project through that we can monitor particular object in Realtime using thermal and normal camera. Responsibilities: - Design and implement embedded voice assistant pipelines using ESP-SR, ESP-IDF, or PlatformIO. - Integrate I2S digital microphones (e.g., INMP441, DFPlayer) with ESP32 for real-time audio capture. - Develop wake word detection, VAD, and command recognition using models like WakeNet, MultiNet, or TinyML-based solutions. - - Optimize AI models and inference for ultra-low-power operation. - Manage real-time tasks using FreeRTOS on ESP32 platforms. - Interface with peripherals like SD cards, LEDs, relays, and Wi-Fi/BLE modules. - Debug, profile, and optimize memory and performance on constrained hardware. Required Skills: - Strong proficiency in C/C++ and embedded development for ESP32. - Experience with ESP-IDF, PlatformIO, or Arduino ESP32 core. - Practical knowledge of voice processing algorithms: VAD, wake word, STT. - Experience using or modifying ESP-SR, ESP-Skainet, or custom keyword spotting models. - Familiarity with I2S, DMA, and audio pre-processing (gain control, filtering). - Understanding of FreeRTOS, low-power modes, and real-time audio handling. Preferred/Bonus Skills: - Experience with TinyML, TensorFlow Lite for Microcontrollers, or Edge Impulse. - Knowledge of Python for data preprocessing and model training. - They have knowledge and experience in computer vision (OpenCV, image processing). - They have also knowledge and experience in Deep learning/AI for vision (like CNN, YOLO or Faster-R CNN, pytorch, TensorFlow, keras) - Having Experience on Nvidia Jetson Nano/Orin based device. - Experience with Bluetooth (BLE) or Wi-Fi communication for IoT applications. - Experience in noise reduction (e.g., NSNet), echo cancellation, or ESP-DSP. Qualifications: - Bachelor's or Master's degree in Electronics, Embedded Systems, Computer Engineering, or related field. - 2+ years of experience in embedded firmware or AI on edge devices. Why Join Us? - Work on cutting-edge embedded AI products for consumer and industrial voice control. - Opportunity to shape next-gen low-power voice assistant hardware. - Flexible remote work options and tech ownership. How to Apply: Send your resume, GitHub/portfolio, and any project demos to: Email: hr@gfofireequipments.com Job Type: Full-time Pay: From ₹100,000.00 per month Schedule: Day shift Work Location: In person

Posted 2 weeks ago

Apply

10.0 years

1 - 1 Lacs

Hyderābād

On-site

JOB DESCRIPTION Elevate your career as the Director of Machine Learning Engineering, where your technical expertise and visionary leadership will shape the future of AI and ML solutions. As a Director of Machine Learning Engineering at JPMorgan Chase within the Corporate Sector – Artificial Intelligence and Machine Learning (AIML) Data Platforms, you will lead a specialized technical area, driving impact across teams, technologies, and projects. In this role, you will leverage your deep knowledge of machine learning, software engineering, and product management to spearhead multiple complex ML projects and initiatives, serving as the primary decision-maker and a catalyst for innovation and solution delivery. You will be responsible for hiring, leading, and mentoring a team of Machine Learning and Software Engineers, focusing on best practices in ML engineering, with the goal of elevating team performance to produce high-quality, scalable ML solutions with operational excellence. You will engage deeply in technical aspects, reviewing code, mentoring engineers, troubleshooting production ML applications, and enabling new ideas through rapid prototyping. Your passion for parallel distributed computing, big data, cloud engineering, micro-services, automation, and operational excellence will be key. Job Responsibilities Lead and manage a team of machine learning engineers, ensuring the implementation, delivery, and support of high-quality ML solutions. Collaborate with product teams to deliver tailored, AI/ML-driven technology solutions. Architect and implement distributed AI/ML infrastructure, including inference, training, scheduling, orchestration, and storage. Develop advanced monitoring and management tools for high reliability and scalability in AI/ML systems. Optimize AI/ML system performance by identifying and resolving inefficiencies and bottlenecks. Drive the adoption and execution of AI/ML Platform tools across various teams. Integrate Generative AI and Classical AI within the ML Platform using state-of-the-art techniques. Lead the entire AI/ML product life cycle through planning, execution, and future development by continuously adapting, developing new AI/ML products and methodologies, managing risks, and achieving business targets like cost, features, reusability, and reliability to support growth. Manage, mentor, and develop a team of AI/ML professionals in a way that promotes a culture of excellence, continuous learning, and supports their professional goals. Required Qualifications, Capabilities, and Skills Formal training or certification in software engineering concepts and 10+ years applied experience. In addition, 5+ years of experience leading technologists to manage, anticipate and solve complex technical items within your domain of expertise 12+ years of experience in engineering management with a strong technical background in machine learning. Extensive hands-on experience with AI/ML frameworks (TensorFlow, PyTorch, JAX, scikit-learn). Deep expertise in Cloud Engineering (AWS, Azure, GCP) and Distributed Micro-service architecture. Experienced with Kubernetes ecosystem, including EKS, Helm, and custom operators. Background in High Performance Computing, ML Hardware Acceleration (e.g., GPU, TPU, RDMA), or ML for Systems. Strategic thinker with the ability to craft and drive a technical vision for maximum business impact. Demonstrated leadership in working effectively with engineers, data scientists, and ML practitioners. Preferred Qualifications, Capabilities, and Skills Strong coding skills and experience in developing large-scale AI/ML systems. Proven track record in contributing to and optimizing open-source ML frameworks. Recognized thought leader within the field of machine learning. Understanding & experience of AI/ML Platforms, LLMs, GenAI, and AI Agents. ABOUT US JPMorganChase, one of the oldest financial institutions, offers innovative financial solutions to millions of consumers, small businesses and many of the world’s most prominent corporate, institutional and government clients under the J.P. Morgan and Chase brands. Our history spans over 200 years and today we are a leader in investment banking, consumer and small business banking, commercial banking, financial transaction processing and asset management. We recognize that our people are our strength and the diverse talents they bring to our global workforce are directly linked to our success. We are an equal opportunity employer and place a high value on diversity and inclusion at our company. We do not discriminate on the basis of any protected attribute, including race, religion, color, national origin, gender, sexual orientation, gender identity, gender expression, age, marital or veteran status, pregnancy or disability, or any other basis protected under applicable law. We also make reasonable accommodations for applicants’ and employees’ religious practices and beliefs, as well as mental health or physical disability needs. Visit our FAQs for more information about requesting an accommodation. ABOUT THE TEAM Our professionals in our Corporate Functions cover a diverse range of areas from finance and risk to human resources and marketing. Our corporate teams are an essential part of our company, ensuring that we’re setting our businesses, clients, customers and employees up for success.

Posted 2 weeks ago

Apply

3.0 - 5.0 years

0 Lacs

Delhi, India

On-site

Company Description Monk Outsourcing is a digital marketing service provider company based in Delhi, India. We offer staffing solutions for US-based companies, software development, web development, content management solutions, and creative design services. Our team of experts works with modern technologies and tools to deliver web-based projects from concept to implementation. We are looking for a talented AI/ML Engineer to join our dynamic team and contribute to our exciting projects involving large language models (LLMs). *Job Overview:* As an AI/ML Engineer specializing in generative AI applications, you will be responsible for developing and optimizing the entire machine learning pipeline. This includes data preprocessing, model training, fine-tuning, and deployment. You will work closely with data scientists, software engineers, and product managers to create efficient and scalable LLM models that meet our enterprise clients' needs. *Key Responsibilities:* • Design, implement, and maintain end-to-end machine learning pipelines for generative AI applications. • Develop and fine-tune large language models (LLMs) to meet specific project requirements. • Implement efficient data preprocessing and augmentation techniques to enhance model performance. • Collaborate with cross-functional teams to define project requirements and deliver AI solutions that align with business objectives. • Conduct experiments to evaluate model performance, using metrics and validation techniques to ensure high-quality results. • Optimize model inference and deployment for scalability and efficiency in production environments. • Stay updated with the latest advancements in AI/ML research and incorporate relevant innovations into our projects. • Provide technical guidance and mentorship to junior team members. *Required Skills and Qualifications:* • Bachelor's or Master's degree in Computer Science, Data Science, Machine Learning, or a related field. • 3-5 years of experience in machine learning, with a focus on generative AI and LLMs. • Proficiency in programming languages such as Python, and experience with ML frameworks like TensorFlow, PyTorch, or similar. • Strong understanding of NLP concepts, including text generation, prompting, and transformer-based architectures. • Experience in building and deploying machine learning models in production environments. • Knowledge of data preprocessing techniques, including text cleaning, tokenization, and augmentation. • Familiarity with cloud platforms (AWS, GCP, Azure) and containerization technologies (Docker, Kubernetes) for scalable model deployment. • Excellent problem-solving skills and the ability to work independently and collaboratively in a fast-paced environment. • Strong communication skills, with the ability to explain complex technical concepts to non-technical stakeholders. *Preferred Qualifications:* • Experience with fine-tuning pre-trained LLMs such as GPT, BERT, or similar. • Familiarity with MLOps practices and tools for continuous integration and deployment (CI/CD) of ML models. • Understanding of ethical considerations and bias mitigation in AI models. • Contributions to open-source projects or publications in AI/ML conferences/journals. Show more Show less

Posted 2 weeks ago

Apply

2.0 years

0 Lacs

Hyderābād

On-site

Company: Qualcomm India Private Limited Job Area: Engineering Group, Engineering Group > Software Engineering General Summary: Job Description Join the exciting Generative AI team at Qualcomm focused on integrating cutting edge GenAI models on Qualcomm chipsets. The team uses Qualcomm chips’ extensive heterogeneous computing capabilities to allow inference of GenAI models on-device without a need for connection to the cloud. Our inference engine is designed to help developers run neural network models trained in a variety of frameworks on Snapdragon platforms at blazing speeds while still sipping the smallest amount of power. Utilize this power efficient hardware and Software stack to run Large Language Models (LLMs) and Large Vision Models (LVM) at near GPU speeds! Responsibilities: In this role, you will spearhead the development and commercialization of the Qualcomm AI Runtime (QAIRT) SDK on Qualcomm SoCs. As an AI inferencing expert, you'll push the limits of performance from large models. Your mastery in deploying large C/C++ software stacks using best practices will be essential. You'll stay on the cutting edge of GenAI advancements, understanding LLMs/Transformers and the nuances of edge-based GenAI deployment. Most importantly, your passion for the role of edge in AI's evolution will be your driving force. Requirements: Master’s/Bachelor’s degree in computer science or equivalent. 2-4 years of relevant work experience in software development. Strong understanding of Generative AI models – LLM, LVM, LMMs and building blocks (self-attention, cross attention, kv caching etc.) Floating-point, Fixed-point representations and Quantization concepts. Experience with optimizing algorithms for AI hardware accelerators (like CPU/GPU/NPU). Strong in C/C++ programming, Design Patterns and OS concepts. Good scripting skills in Python. Excellent analytical and debugging skills. Good communication skills (verbal, presentation, written). Ability to collaborate across a globally diverse team and multiple interests. Preferred Qualifications Strong understanding of SIMD processor architecture and system design. Proficiency in object-oriented software development and familiarity Familiarity with Linux and Windows environment Strong background in kernel development for SIMD architectures. Familiarity with frameworks like llama.cpp, MLX, and MLC is a plus. Good knowledge of PyTorch, TFLite, and ONNX Runtime is preferred. Experience with parallel computing systems and languages like OpenCL and CUDA is a plus. Minimum Qualifications: Bachelor's degree in Engineering, Information Systems, Computer Science, or related field and 2+ years of Software Engineering or related work experience. OR Master's degree in Engineering, Information Systems, Computer Science, or related field and 1+ year of Software Engineering or related work experience. OR PhD in Engineering, Information Systems, Computer Science, or related field. 2+ years of academic or work experience with Programming Language such as C, C++, Java, Python, etc. Applicants : Qualcomm is an equal opportunity employer. If you are an individual with a disability and need an accommodation during the application/hiring process, rest assured that Qualcomm is committed to providing an accessible process. You may e-mail disability-accomodations@qualcomm.com or call Qualcomm's toll-free number found here. Upon request, Qualcomm will provide reasonable accommodations to support individuals with disabilities to be able participate in the hiring process. Qualcomm is also committed to making our workplace accessible for individuals with disabilities. (Keep in mind that this email address is used to provide reasonable accommodations for individuals with disabilities. We will not respond here to requests for updates on applications or resume inquiries). Qualcomm expects its employees to abide by all applicable policies and procedures, including but not limited to security and other requirements regarding protection of Company confidential information and other confidential and/or proprietary information, to the extent those requirements are permissible under applicable law. To all Staffing and Recruiting Agencies : Our Careers Site is only for individuals seeking a job at Qualcomm. Staffing and recruiting agencies and individuals being represented by an agency are not authorized to use this site or to submit profiles, applications or resumes, and any such submissions will be considered unsolicited. Qualcomm does not accept unsolicited resumes or applications from agencies. Please do not forward resumes to our jobs alias, Qualcomm employees or any other company location. Qualcomm is not responsible for any fees related to unsolicited resumes/applications. If you would like more information about this role, please contact Qualcomm Careers.

Posted 2 weeks ago

Apply

5.0 years

10 - 27 Lacs

India

On-site

About MostEdge At MostEdge , our purpose is clear: Accelerate commerce and build sustainable, trusted experiences. With every byte of data, we strive to Protect Every Penny. Power Every Possibility. We empower retailers to make real-time, profitable decisions using cutting-edge AI , smart infrastructure, and operational excellence. Our platforms handle: hundreds of thousands of sales transactions/hour hundreds of vendor purchase invoices/hour few hundred product updates/day With systems built for 99.99999% uptime We are building an AI-native commerce engine , and language models are at the heart of this transformation. Role Overview We are looking for an AI/ML Expert with deep experience in training and deploying Large Language Models (LLMs) to power MostEdge's next-generation operations, cost intelligence, and customer analytics platform . You will be responsible for fine-tuning domain-specific models using internal structured and unstructured data (product catalogs, invoices, chats, documents), embedding real-time knowledge through RAG pipelines, and enabling AI-powered interfaces that drive search, reporting, insight generation, and operational recommendations. Scope & Accountability What You Will Own Fine-tune and deploy LLMs for product, vendor, and shopper-facing use cases. Design hybrid retrieval-augmented generation (RAG) pipelines with LangChain, FastAPI, and vector DBs (e.g., FAISS, Weaviate, Qdrant). Train models on internal datasets (sales, cost, product specs, invoices, support logs) using supervised fine-tuning and LoRA/QLoRA techniques. Orchestrate embedding pipelines, prompt tuning, and model evaluation across customer and field operations use cases. Deploy LLMs efficiently on RunPod, AWS, or GCP , optimizing for multi-GPU, low-latency inference . Collaborate with engineering and product teams to embed model outputs in dashboards, chat UIs, and retail systems. What Success Looks Like 90%+ accuracy on retrieval and reasoning tasks for product/vendor cost and invoice queries. <3s inference time across operational prompts, running on GPU-optimized containers. Full integration of LLMs with backend APIs, sales dashboards, and product portals. 75% reduction in manual effort across selected operational workflows. Skills & Experience Must-Have 5+ years in AI/ML , with 2+ years working on LLMs or transformer architectures . Proven experience training or fine-tuning Mistral, LLaMA, Falcon, or similar open-source LLMs . Strong command over LoRA, QLoRA, PEFT, RAG, embeddings, and quantized inference . Familiarity with LangChain, HuggingFace Transformers, FAISS/Qdrant , and FastAPI for LLM orchestration. Experience deploying models on RunPod, AWS, or GCP using Docker + Kubernetes. Proficient in Python , PyTorch , and data preprocessing (structured and unstructured). Experience with ETL pipelines , multi-modal data, and real-time data integration. Nice-to-Have Experience with retail, inventory, or customer analytics systems . Knowledge of semantic search, OCR post-processing, or auto-tagging pipelines . Exposure to multi-tenant environments and secure model isolation for enterprise use. How You Reflect Our Values Lead with Purpose : You empower smarter decisions with AI-first operations. Build Trust : You make model behavior explainable, dependable, and fair. Own the Outcome : You train and optimize end-to-end pipelines from data to insights. Win Together : You partner across engineering, ops, and customer success teams. Keep It Simple : You design intuitive models, prompts, and outputs that drive action—not confusion. Why Join MostEdge? Shape how AI transforms commerce and operations at scale . Be part of a mission-critical, high-velocity, AI-first company . Build LLMs with purpose—connecting frontline data to real-time results. Job Types: Full-time, Permanent Pay: ₹1,068,726.69 - ₹2,729,919.70 per year Benefits: Health insurance Life insurance Paid sick time Paid time off Provident Fund Schedule: Evening shift Morning shift US shift Supplemental Pay: Performance bonus Yearly bonus Work Location: In person Expected Start Date: 15/07/2025

Posted 2 weeks ago

Apply

0 years

0 Lacs

Delhi

On-site

SUMMARY We are seeking a Machine Learning Analyst with a strong foundation in Engineering or a related Quantitative Sciences discipline. While prior experience in Machine Learning is not mandatory, candidates with exposure to Machine Learning and Deep Learning (if any) are expected to demonstrate a rigorous understanding of the concepts they are familiar with. The ideal candidate must be a quick learner and demonstrate strong analytical skills, clear thinking and structured problem-solving, strong quantitative aptitude, a willingness to learn, high self-motivation, and a diligent work ethic. ABOUT US Wadhwani AI is a nonprofit institute building and deploying applied AI solutions to solve critical issues in public health, agriculture, education, and urban development in underserved communities in the global south. We collaborate with governments, social sector organizations, academic and research institutions, and domain experts to identify real-world problems, and develop practical AI solutions to tackle these issues with the aim of making a substantial positive impact. We have over 30+ AI projects supported by leading philanthropies such as Bill & Melinda Gates Foundation and Google.org. With a team of over 200 professionals, our expertise encompasses AI/ML research and innovation, software engineering, domain knowledge, design and user research. In the Press: Our Founder Donors are among the Top 100 AI Influencers G20 India’s Presidency: AI Healthcare, Agriculture, & Education Solutions Showcased Globally. Unlocking the potentials of AI in Public Health Wadhwani AI Takes an Impact-First Approach to Applying Artificial Intelligence - data.org Winner of the H&M Foundation Global Change Award 2022 Indian Winners of the 2019 Google AI Impact Challenge, and the first in the Asia Pacific to host Google Fellow PRE-REQUISITES ML Analyst position is open to all with prior training in Engineering or any related Quantitative Sciences discipline No prior experience in Machine Learning or Deep Learning is required Candidates with exposure to ML/DL (if any) are expected to have a clear and rigorous understanding of the concepts they are familiar with Strong skills in data handling, and logical problem-solving Demonstrates a quick learning ability, and a strong work ethic Willingness to take on any task, learn new tools, and adapt to evolving project needs ROLES & RESPONSIBILITIES Work closely with data to support the development of ML and DL solutions Conduct experiments under guidance and report results reliably Learn to derive insights from experimental outcomes and determine appropriate next steps Prepare, curate, and analyse datasets for training and evaluation Monitor incoming data streams and perform regular quality checks Assist in training and inference of ML models, including deep learning architectures Contribute to well-documented and maintainable codebases Document work clearly and consistently with high standards Communicate and present experimental findings and results clearly within the team Learn and apply best practices across ML development, coding, documentation, and experimentation Collaborate effectively with project teams to meet milestones and deliverables Proactively seek help and feedback when needed Work efficiently with tools like Unix, VS Code, GitHub, and Docker Develop proficiency with common ML tools and libraries such as Pandas, Scikit-learn, PyTorch, Excel (pivot tables), Matplotlib, Weights & Biases DESIRED COMPETENSIES Demonstrates curiosity, humility, and a strong motivation to learn and grow Takes full ownership of tasks; highly diligent, detail-oriented, and accountable Willing to engage in all types of work from data cleaning and exploration to debugging and tooling Comfortable sitting with raw data to explore, understand and derive insights, and not just focused on modelling Proactively seeks guidance and independently builds knowledge when needed Approaches every task with a quality-first mindset; no task is considered beneath them Identifies recurring patterns and abstracts them into reusable, generalisable workflows Contributes across the entire ML lifecycle including data preparation, experimentation, and analysis Selects and applies appropriate tools; builds efficient, reliable, and repeatable processes Maintains a high standard of error-free work; reviews and validates work thoroughly Collaborates effectively with cross-functional teams Communicates clearly and constructively, with an emphasis on precision and clarity We are committed to promoting diversity and the principle of equal employment opportunity for all our employees and encourage qualified candidates to apply irrespective of religion or belief, ethnic or social background, gender, gender identity, and disability. If you have any questions, please email us at careers@wadhwaniai.org.

Posted 2 weeks ago

Apply

3.0 years

1 - 6 Lacs

Noida

On-site

Level AI was founded in 2019 and is a Series C startup headquartered in Mountain View, California. Level AI revolutionizes customer engagement by transforming contact centers into strategic assets. Our AI-native platform leverages advanced technologies such as Large Language Models to extract deep insights from customer interactions. By providing actionable intelligence, Level AI empowers organizations to enhance customer experience and drive growth. Consistently updated with the latest AI innovations, Level AI stands as the most adaptive and forward-thinking solution in the industry. Empowering contact center stakeholders with real-time insights, our tech facilitates data-driven decision-making for contact centers, enhancing service levels and agent performance. As a vital team member, your work will be cutting-edge technologies and will play a high-impact role in shaping the future of AI-driven enterprise applications. You will directly work with people who've worked at Amazon, Facebook, Google, and other technology companies in the world. With Level AI, you will get to have fun, learn new things, and grow along with us. Ready to redefine possibilities? Join us! We'll love to explore more about you if you have Qualification: B.E/B.Tech/M.E/M.Tech/PhD from tier 1 engineering institutes with relevant work experience with a top technology company in computer science or mathematics-related fields with 3-5 years of experience in machine learning and NLP. Knowledge and practical experience in solving NLP problems in areas such as text classification, entity tagging, information retrieval, question-answering, natural language generation, clustering, etc. 3+ years of experience working with LLMs in large-scale environments. Expert knowledge of machine learning concepts and methods, especially those related to NLP, Generative AI, and working with LLMs. Knowledge and hands-on experience with Transformer-based Language Models like BERT, DeBERTa, Flan-T5, Mistral, Llama, etc. Deep familiarity with internals of at least a few Machine Learning algorithms and concepts. Experience with Deep Learning frameworks like Pytorch and common machine learning libraries like scikit-learn, numpy, pandas, NLTK, etc. Experience with ML model deployments using REST API, Docker, Kubernetes, etc. Knowledge of cloud platforms (AWS/Azure/GCP) and their machine learning services is desirable. Knowledge of basic data structures and algorithms. Knowledge of real-time streaming tools/architectures like Kafka, Pub/Sub is a plus. Your role at Level AI includes but is not limited to Big picture: Understand customers’ needs, innovate and use cutting edge Deep Learning techniques to build data-driven solutions. Work on NLP problems across areas such as text classification, entity extraction, summarization, generative AI, and others. Collaborate with cross-functional teams to integrate/upgrade AI solutions into the company’s products and services. Optimize existing deep learning models for performance, scalability, and efficiency. Build, deploy, and own scalable production NLP pipelines. Build post-deployment monitoring and continual learning capabilities. Propose suitable evaluation metrics and establish benchmarks. Keep abreast with SOTA techniques in your area and exchange knowledge with colleagues. Desire to learn, implement and work with latest emerging model architectures, training and inference techniques, data curation pipelines, etc. To learn more visit : https://thelevel.ai/ Funding : https://www.crunchbase.com/organization/level-ai LinkedIn : https://www.linkedin.com/company/level-ai/

Posted 2 weeks ago

Apply

8.0 - 15.0 years

0 Lacs

Noida, Uttar Pradesh, India

On-site

We, at Cogito, are looking for someone with a zeal and flair to curate knowledge pieces and writings on how Artificial Intelligence, Machine Learning & NLP have taken automation & efficiencies in various industries by storm and how these ever shape-shifting technologies are surprising humans themselves everyday. Cogito Technologies, having a strong presence in USA and its delivery centre in Delhi NCR, India, specializes in Human Empowered Automation. Our mission is to help our customers innovate and scale by solving their day-to-day data needs. Using our skilled on-demand workforce, we partner with Machine Learning, Artificial Intelligence, Technology and eCommerce clients to develop high-quality data sets used to build and enhance various cutting-edge business applications. Cogito is currently working on Artificial Intelligence applications in Annotation work and is looking to engage with "Content Manager" to participate in ongoing projects. The Role: Being Content manager you will be working closely with a team to read, write, summarize knowledge, and interpret meaning. Think of it like being a language arts teacher or a personal tutor for some of the world's most influential technology. Assist with development of Large Language Models Assist with development of Generative AI applications Evaluate and Grade responses of LLM models Content Summarization Model evaluation Model alignment Efficient inference Classification and categorization Sentiment analysis Multimodal model question-answering systems Skills Set: Fluent to native English writing and reading skills required University to Masters degree 8-15 years of extensive experience working as Content Writer Ability to answer free-form prompts (questions) across a general scope of topics Ability to reference online resources and rephrase and write cohesive, accurate, responsive, and sometimes empathetic answers to the prompts in the perspective of an AI model Proven track record working on other projects requiring free form sentence creation Ability to answer specialized domain free-form prompts at university-level difficulty Ability to provide cohesive explanations to their prompt responses (E.g., explaining the process of solving a given math question, listing out the detailed process for a software question, etc.) Extra Added advantage if you have: Up-to-date knowledge on US culture and news. Knowledge of various NLP tasks and algorithms, such as sentiment analysis, named entity recognition, text classification, machine translation, and language generation. The Person This is a technology job that accommodates humanities people and relies more fully on your competencies than your formal experience making it a rare opportunity that's perfect for new grads, career transitioners, and those seeking an exciting encore career in AI. You're the sort of person who is exceptional at generating copy intuitively. You've also got the metacognitive awareness to "show your work." In addition, you're the sort of person with a deep interest in linguistics. You instinctively document learning and continuously refine methods. You are a(n): Dedicated Grammarian Natural Writer Diligent Reader Self-Aware Interpreter Synthetic Sense-Maker Perpetual Learner Instinctive Educator Proven experience required Core Content Writing Background with own website/blogs/written piece of articles etc. Work in a cordial and cooperative manner with Marketing, SEO & Operations teams Most importantly, technical bent of mind with keen interest in new technologies and their applications The ability to think and act like a stakeholder is a plus Non-technical Strong sense of responsibility in quality and idea generation Can work well with others and open for giving/taking suggestions Strong written and verbal communication skills Strong motivation to learn and go in-depth into field of study Job requirements Fluent in English to be able to follow instructions and guidelines Detail-oriented with the ability to understand and follow instructions Ability to meet deadlines Responsible, reliable, and communicative Designation: Manager-Content Gender: Male/Female Work Location: Noida DOJ: ASAP Job Type: Full Time CTC: Best in the industry (will be discussed during call)Please apply here or email your updated resume: hr@anolytics.in --- Regards, Team HR Show more Show less

Posted 2 weeks ago

Apply

1.0 years

0 Lacs

Bengaluru, Karnataka, India

On-site

Company: Qualcomm India Private Limited Job Area: Engineering Group, Engineering Group > Systems Engineering General Summary: Summary - We are seeking experts with a robust background in the field of deep learning (DL) to design state-of-the-art low-level perception (LLP) as well as end-to-end AD models, with a focus on achieving accuracy-latency Pareto optimality. This role involves comprehending state-of-the-art research in this field and deploying networks on the Qualcomm Ride platform for L2/L3 Advanced Driver Assistance Systems (ADAS) and autonomous driving. The ideal candidate must be well-versed in recent advancements in Vision Transformers (Cross-attention, Self-attention), lifting 2D features to Bird's Eye View (BEV) space, and their applications to multi-modal fusion. This position offers extensive opportunities to collaborate with advanced R&D teams of leading automotive Original Equipment Manufacturers (OEMs) as well as Qualcomm's internal stack teams. The team is responsible for enhancing the speed, accuracy, power consumption, and latency of deep networks running on Snapdragon Ride AI accelerators. A thorough understanding of machine learning algorithms, particularly those related to automotive use cases (autonomous driving, vision, and LiDAR processing ML algorithms), is essential. Research experience in the development of efficient networks, various Neural Architecture Search (NAS) techniques, network quantization, and pruning is highly desirable. Strong communication and interpersonal skills are required, and the candidate must be able to work effectively with various horizontal AI teams. Minimum Qualifications: Bachelor's degree in Computer Science, Engineering, Information Systems, or related field and 1+ years of Hardware Engineering, Software Engineering, Systems Engineering, or related work experience. OR Master's degree in Computer Science, Engineering, Information Systems, or related field and 1+ year of Hardware Engineering, Software Engineering, Systems Engineering, or related work experience. OR PhD in Computer Science, Engineering, Information Systems, or related field. Preferred Qualifications: Good at software development with excellent analytical, development, and problem-solving skills. Strong understanding of Machine Learning fundamentals Hands-on experience with deep learning network design and implementation. Ability to define network from scratch in PyTorch, ability to add new loss function, modify network with torch.fx. Adept at version control system like GIT. Experience in neural network quantization, compression, pruning algorithms. Experience in deep learning kernel/compiler optimization Strong communication skills Principal Duties And Responsibilities: Applies Machine Learning knowledge to extend training or runtime frameworks or model efficiency software tools with new features and optimizations. Models, architects, and develops machine learning hardware (co-designed with machine learning software) for inference or training solutions. Develops optimized software to enable AI models deployed on hardware (e.g., machine learning kernels, compiler tools, or model efficiency tools, etc.) to allow specific hardware features; collaborates with team members for joint design and development. Assists with the development and application of machine learning techniques into products and/or AI solutions to enable customers to do the same. Develops, adapts, or prototypes complex machine learning algorithms, models, or frameworks aligned with and motivated by product proposals or roadmaps with minimal guidance from more experienced engineers. Conducts complex experiments to train and evaluate machine learning models and/or software independently. Applicants : Qualcomm is an equal opportunity employer. If you are an individual with a disability and need an accommodation during the application/hiring process, rest assured that Qualcomm is committed to providing an accessible process. You may e-mail myhr.support@qualcomm.com or call Qualcomm's toll-free number found here . Upon request, Qualcomm will provide reasonable accommodations to support individuals with disabilities to be able participate in the hiring process. Qualcomm is also committed to making our workplace accessible for individuals with disabilities. Qualcomm expects its employees to abide by all applicable policies and procedures, including but not limited to security and other requirements regarding protection of Company confidential information and other confidential and/or proprietary information, to the extent those requirements are permissible under applicable law. Minimum Qualifications: Bachelor's degree in Engineering, Information Systems, Computer Science, or related field. Applicants : Qualcomm is an equal opportunity employer. If you are an individual with a disability and need an accommodation during the application/hiring process, rest assured that Qualcomm is committed to providing an accessible process. You may e-mail disability-accomodations@qualcomm.com or call Qualcomm's toll-free number found here. Upon request, Qualcomm will provide reasonable accommodations to support individuals with disabilities to be able participate in the hiring process. Qualcomm is also committed to making our workplace accessible for individuals with disabilities. (Keep in mind that this email address is used to provide reasonable accommodations for individuals with disabilities. We will not respond here to requests for updates on applications or resume inquiries). Qualcomm expects its employees to abide by all applicable policies and procedures, including but not limited to security and other requirements regarding protection of Company confidential information and other confidential and/or proprietary information, to the extent those requirements are permissible under applicable law. To all Staffing and Recruiting Agencies : Our Careers Site is only for individuals seeking a job at Qualcomm. Staffing and recruiting agencies and individuals being represented by an agency are not authorized to use this site or to submit profiles, applications or resumes, and any such submissions will be considered unsolicited. Qualcomm does not accept unsolicited resumes or applications from agencies. Please do not forward resumes to our jobs alias, Qualcomm employees or any other company location. Qualcomm is not responsible for any fees related to unsolicited resumes/applications. If you would like more information about this role, please contact Qualcomm Careers. 3074561 Show more Show less

Posted 2 weeks ago

Apply

0 years

0 Lacs

Gurugram, Haryana, India

On-site

Hi All, We are hiring for a Generative AI Architect for one of our esteemed clients based in Gurgaon . Role Overview: We’re looking for an experienced Generative AI Architect to lead the design and implementation of scalable GenAI solutions for real-world applications. The ideal candidate will have a proven track record of deploying GenAI systems in production environments, handling performance optimization, data governance, model evaluation, and vector search enhancements. Key Responsibilities: Architect GenAI systems integrating LLMs, APIs, vector databases, and application layers Design scalable RAG pipelines and knowledge graph integrations Evaluate build vs. buy decisions; choose between open-source and commercial models Define reference architectures and reusable components for GenAI platforms Integrate MLOps/LLMOps for monitoring, versioning, and feedback mechanisms Technical Skills: Strong grasp of Transformer architectures , NLP pipelines, and fine-tuning techniques Hands-on with LangChain, LLamaIndex, Haystack Orchestration tools: Kubeflow, Airflow, Ray Experience with Vector DBs : Pinecone, FAISS, Weaviate, Milvus Cloud expertise in AWS, Azure, or GCP with hands-on GenAI tools Proficient in developing APIs/microservices for AI deployment Strong understanding of embeddings, tokenization, and inference tuning 📩 Interested candidates can send their profiles to: ruchi@intileo.com Please include the following details in your application: Current CTC Expected CTC Notice Period Current Location Brief note on your suitability for the role Looking forward to connecting with the right talent! #Hiring #GenerativeAI #AIArchitect #LLM #LangChain #VectorDB #AIJobs #Gurgaon #Intileo Show more Show less

Posted 2 weeks ago

Apply

7.0 years

0 Lacs

Noida, Uttar Pradesh, India

On-site

As an AI Architect at Innovaccer, you will be at the forefront of designing, training, and fine-tuning foundational models with healthcare-specific data. You will be critical in integrating these technologies into our healthcare products, ensuring they operate effectively in a production environment. Your expertise will help us to revolutionize healthcare delivery, improve patient outcomes, and maintain compliance with industry regulations. A Day in Life Fine-tuning state-of-the-art small language models (SLMs) like Llama-3.1-8B or Phi-3.5-3B on diverse healthcare datasets. Evaluating and optimizing various inference and serving frameworks (e.g., Triton, TensorRT, VLLM, TGI). Read about the most recent AI model architectures. Collaborating with product teams to review and refine architecture designs incorporating AI. Troubleshooting AI deployment issues in customer environments, ensuring smooth operation and optimal performance. Authoring and presenting technical documentation and white papers and showcasing breakthroughs in internal and external forums. Responsibilities Design, train, and fine-tune advanced foundational models (text, audio, vision) using healthcare-specific datasets, focusing on accuracy and clinical relevance. Collaborate with cross-functional teams (product, engineering, clinical) to seamlessly integrate AI/ML technologies into our product offerings. Deploy, monitor, and manage AI models in a production environment, ensuring high availability, scalability, and performance. Continuously research and evaluate the latest advancements in AI/ML and healthcare industry trends to drive innovation within Innovaccer. Create high-quality intellectual property (white papers, case studies, conference presentations) showcasing Innovaccer's technical leadership. Ensure all AI solutions adhere to healthcare standards and regulatory requirements (HIPAA, GDPR, etc.). Develop and maintain comprehensive documentation for AI models, including development, training, fine-tuning, and deployment procedures. Provide technical guidance and mentorship to junior AI engineers and team members. Collaborate with stakeholders to understand clinical needs and translate them into technical requirements for model fine-tuning and development. Select and curate appropriate datasets for fine-tuning foundational models to address specific healthcare use cases. Qualifications Master’s or Ph.D. in Computer Science, Artificial Intelligence, Machine Learning, or a related field. Minimum of 7 years of hands-on experience in AI/ML, with a demonstrable track record of training and deploying LLMs and other machine learning models. Strong proficiency in Python and familiarity with popular AI/ML frameworks (TensorFlow, PyTorch, Hugging Face Transformers, etc.). Practical experience deploying and managing AI models in production environments, including expertise in serving and inference frameworks (Triton, TensorRT, VLLM, TGI, etc.). Proven experience in healthcare AI applications, including a solid understanding of healthcare data standards (FHIR, HL7) and regulatory compliance (HIPAA, SOC2, NIST). Excellent problem-solving and analytical abilities, capable of tackling complex challenges and evaluating multiple factors. Exceptional communication and collaboration skills, enabling effective teamwork in a dynamic environment. A strong publication record in top-tier AI/ML conferences (NeurIPS, ICML, KDD, CVPR, AAMAS, etc.) is highly desirable. Experience with cloud computing platforms (AWS, GCP, Azure) and containerization technologies (Docker, Kubernetes) is a plus. Show more Show less

Posted 2 weeks ago

Apply

8.0 years

0 Lacs

Mumbai Metropolitan Region

On-site

Job Description Are You Ready to Make It Happen at Mondelēz International? Join our Mission to Lead the Future of Snacking. Make It With Pride. You will be crucial in supporting our business by creating valuable, actionable insights about the data, and communicating your findings to the business. You will work with various stakeholders to determine how to use business data for business solutions/insights. How You Will Contribute You will: Analyze and derive value from data through the application methods such as mathematics, statistics, computer science, machine learning and data visualization. In this role you will also formulate hypotheses and test them using math, statistics, visualization and predictive modeling Understand business challenges, create valuable actionable insights about the data, and communicate your findings to the business. After that you will work with stakeholders to determine how to use business data for business solutions/insights Enable data-driven decision making by creating custom models or prototypes from trends or patterns discerned and by underscoring implications. Coordinate with other technical/functional teams to implement models and monitor results Apply mathematical, statistical, predictive modelling or machine-learning techniques and with sensitivity to the limitations of the techniques. Select, acquire and integrate data for analysis. Develop data hypotheses and methods, train and evaluate analytics models, share insights and findings and continues to iterate with additional data Develop processes, techniques, and tools to analyze and monitor model performance while ensuring data accuracy Evaluate the need for analytics, assess the problems to be solved and what internal or external data sources to use or acquire. Specify and apply appropriate mathematical, statistical, predictive modelling or machine-learning techniques to analyze data, generate insights, create value and support decision making Contribute to exploration and experimentation in data visualization and you will manage reviews of the benefits and value of analytics techniques and tools and recommend improvements What You Will Bring A desire to drive your future and accelerate your career and the following experience and knowledge: Strong quantitative skillset with experience in statistics and linear algebra. A natural inclination toward solving complex problems Knowledge/experience with statistical programming languages including R, Python, SQL, etc., to process data and gain insights from it Knowledge of machine learning techniques including decision-tree learning, clustering, artificial neural networks, etc., and their pros and cons Knowledge and experience in advanced statistical techniques and concepts including, regression, distribution properties, statistical testing, etc. Good communication skills to promote cross-team collaboration Multilingual coding knowledge/experience: Java, JavaScript, C, C++, etc. Experience/knowledge in statistics and data mining techniques including random forest, GLM/regression, social network analysis, text mining, etc. Ability to use data visualization tools to showcase data for stakeholders About The Role We are seeking a highly motivated and experienced Senior Data Scientist to join our team and spearhead the development of Generative AI (GenAI) capabilities within Mondelēz International from an enterprise application perspective. In this pivotal role, you will champion the full lifecycle of GenAI application projects, taking ownership from initial concept and design to successful deployment, ongoing optimization, and continuous improvement. Beyond project execution, you will serve as a strategic advisor, guiding the evolution of our GenAI capabilities and ensuring alignment with overarching business objectives. This includes defining and enforcing internal standards for GenAI application development, guaranteeing compliance with ethical and security guidelines. You will also support the design of enterprise-level logical architectures for typical GenAI applications, and evaluate and recommend the most appropriate tools and technologies to empower our GenAI initiatives. Ultimately, you will play a critical role in shaping the future of GenAI within Mondelēz, driving innovation, maximizing value realization, and fostering responsible AI adoption across the enterprise. Key Responsibilities: GenAI Application development: Serve as the tech lead from data science point of view in the complete lifecycle of GenAI application projects, from initial ideation and design to successful deployment, optimization, and continuous improvement. Include both in house development or partner collaboration. Strategic Advisory: Provide strategic advice on the evolution of our GenAI capabilities to match company goals. Keeping up with the latest GenAI trend and map that to Mondelez application context to enable us do things better and smarter. Standards and Governance: Help establish and enforce programmatic approaches, governance frameworks and best practices for effective GenAI application building. Responsible AI for GenAI applications, data protection in GenAI context, complying to regulatory requirements and cost-effective GenAI application deployment. Enterprise Architecture Support: Support the design and development of enterprise-level logical architectures for typical GenAI applications, ensuring scalability, maintainability, and integration with existing infrastructure. Technology Evaluation and Recommendation: Evaluate and recommend the most appropriate GenAI tools, technologies, and platforms to empower our GenAI initiatives, staying abreast of the latest advancements in the field with a focus on cloud-based tools/technologies. Knowledge Sharing and Mentoring: Share knowledge and expertise with other team members, mentor junior data scientists, and spearhead the development of a strong GenAI community within Mondelēz. Skills and Experiences: Deep understanding of data science methodologies and implications: proficiency in machine learning, deep learning, statistical modelling, optimization, causal inference etc. Experience in mapping business problems to such methodologies with clear understanding/consideration in Ethical implications, risk mitigation, integration requirements, resource demand/optimization etc. Hands-on experience in cloud environment (8 years): Cloud platform, cloud-based data storage, processing, AI/ML model building, model life cycle, process orchestration, cost optimization etc. Working experience with MLOps (5 years): Understand MLOps processes and practice, familiar with MLOps Tools and technologies. Contributed significantly to shaping MLOps practice in an enterprise setup. LLM Application architecture & Integration (3 years): Hands-on experience building RAG applications, clear understanding of the underline technologies. Extensive experience integrating LLM into workflow through different ways (API etc). Experience in build Agent to leverage LLM and other tools for complex tasks. Familiar with Agent frameworks and orchestration tools. Core GenAI expertise (3 years): Prompt Engineering, Agentic Framework, Model fine-tuning. Cloud Based Deployment & Scaling (LLM-Specific): Practical experience deploying and scaling LLM-based applications in cloud environment. Familiar with scaling strategies for LLMs, and integration with other application. Collaboration with cross-functional teams (5 years): Proven ability to collaborate effectively with cross-functional teams. Excellent communication skills, both written and verbal to articulate technical concepts. Qualifications: Master’s degree in a Quantitative Discipline, PhD preferred. Minimum 12+ years of experience in data science/AI. Minimum 3 years of GenAI focused experience. Within Country Relocation support available and for candidates voluntarily moving internationally some minimal support is offered through our Volunteer International Transfer Policy Business Unit Summary At Mondelēz International, our purpose is to empower people to snack right by offering the right snack, for the right moment, made the right way. That means delivering a broad range of delicious, high-quality snacks that nourish life's moments, made with sustainable ingredients and packaging that consumers can feel good about. We have a rich portfolio of strong brands globally and locally including many household names such as Oreo , belVita and LU biscuits; Cadbury Dairy Milk , Milka and Toblerone chocolate; Sour Patch Kids candy and Trident gum. We are proud to hold the top position globally in biscuits, chocolate and candy and the second top position in gum. Our 80,000 makers and bakers are located in more than 80 countries and we sell our products in over 150 countries around the world. Our people are energized for growth and critical to us living our purpose and values. We are a diverse community that can make things happen—and happen fast. Mondelēz International is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, gender, sexual orientation or preference, gender identity, national origin, disability status, protected veteran status, or any other characteristic protected by law. Job Type Regular Data Science Analytics & Data Science Show more Show less

Posted 2 weeks ago

Apply

2.0 - 4.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Position Overview Job Title: AI Engineer Location: Pune, India Role Description Indra is the central program driving the introduction and safe scaling of AI at DB. Focus is identify AI potential across various banking operations, driving funded use cases into production to create value and confidence and scale across the bank, creating selected shared services with embedded safety to enable low cost scale, developing an AI Workbench for developers for safe AI development at pace, and introducing AI controls whilst aiming to maintain time to market. What We’ll Offer You As part of our flexible scheme, here are just some of the benefits that you’ll enjoy Best in class leave policy Gender neutral parental leaves 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Employee Assistance Program for you and your family members Comprehensive Hospitalization Insurance for you and your dependents Accident and Term life Insurance Complementary Health screening for 35 yrs. and above Your Key Responsibilities Model Deployment: Collaborate with data scientists to deploy machine learning models into production environments. Implement deployment strategies such as A/B testing or canary releases to ensure safe and controlled rollouts. Infrastructure Management: Design and manage the infrastructure required for hosting ML models, including cloud resources and on-premises servers. Utilize containerization technologies like Docker to package models and dependencies. Continuous Integration/Continuous Deployment (CI/CD): Develop and maintain CI/CD pipelines for automating the testing, integration, and deployment of ML models. Implement version control to track changes in both code and model artifacts. Monitoring and Logging: Establish monitoring solutions to track the performance and health of deployed models. Set up logging mechanisms to capture relevant information for debugging and auditing purposes. Scalability and Resource Optimization: Optimize ML infrastructure for scalability and cost-effectiveness. Implement auto-scaling mechanisms to handle varying workloads efficiently. Security and Compliance: Enforce security best practices to safeguard both the models and the data they process. Ensure compliance with industry regulations and data protection standards. Data Management: Oversee the management of data pipelines and data storage systems required for model training and inference. Implement data versioning and lineage tracking to maintain data integrity. Collaboration with Cross-Functional Teams: Work closely with data scientists, software engineers, and other stakeholders to understand model requirements and system constraints. Collaborate with DevOps teams to align MLOps practices with broader organizational goals. Performance Optimization: Continuously optimize and fine-tune ML models for better performance. Identify and address bottlenecks in the system to enhance overall efficiency. Documentation: Maintain clear and comprehensive documentation of MLOps processes, infrastructure, and model deployment procedures. Document best practices and troubleshooting guides for the team. Your Skills And Experience University degree in a technical or quantitative field (e.g., computer science, mathematics, physics, economics, etc.), preferably a Master’s or Doctoral degree 2-4 years of experience in applying AI, machine learning and/or data science in business and/or academia. Strong knowledge of at least one programming language (e.g., Python, JavaScript) and relevant data science or engineering framework (e.g., scikit-learn, TensorFlow, Spark, etc.). Ideally, practical experience in finance and banking Comfortable working with and managing uncertainty and ambiguity Excellent oral and written communication skills in English How We’ll Support You Training and development to help you excel in your career Coaching and support from experts in your team A culture of continuous learning to aid progression A range of flexible benefits that you can tailor to suit your needs About Us And Our Teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment. Show more Show less

Posted 2 weeks ago

Apply

15.0 years

0 Lacs

Pune, Maharashtra, India

On-site

Job Title: VP-Digital Expert Support Lead Experience : 15 + Years Location : Pune Position Overview The Digital Expert Support Lead is a senior-level leadership role responsible for ensuring the resilience, scalability, and enterprise-grade supportability of AI-powered expert systems deployed across key domains like Wholesale Banking, Customer Onboarding, Payments, and Cash Management . This role requires technical depth, process rigor, stakeholder fluency , and the ability to lead cross-functional squads that ensure seamless operational performance of GenAI and digital expert agents in production environments. The candidate will work closely with Engineering, Product, AI/ML, SRE, DevOps, and Compliance teams to drive operational excellence and shape the next generation of support standards for AI-driven enterprise systems. Role-Level Expectations Functionally accountable for all post-deployment support and performance assurance of digital expert systems. Operates at L3+ support level , enabling L1/L2 teams through proactive observability, automation, and runbook design. Leads stability engineering squads , AI support specialists, and DevOps collaborators across multiple business units. Acts as the bridge between operations and engineering , ensuring technical fixes feed into product backlog effectively. Supports continuous improvement through incident intelligence, root cause reporting, and architecture hardening . Sets the support governance framework (SLAs/OLAs, monitoring KPIs, downtime classification, recovery playbooks). Position Responsibilities Operational Leadership & Stability Engineering Own the production health and lifecycle support of all digital expert systems across onboarding, payments, and cash management. Build and govern the AI Support Control Center to track usage patterns, failure alerts, and escalation workflows. Define and enforce SLAs/OLAs for LLMs, GenAI endpoints, NLP components, and associated microservices. Establish and maintain observability stacks (Grafana, ELK, Prometheus, Datadog) integrated with model behavior. Lead major incident response and drive cross-functional war rooms for critical recovery. Ensure AI pipeline resilience through fallback logic, circuit breakers, and context caching. Review and fine-tune inference flows, timeout parameters, latency thresholds, and token usage limits. Engineering Collaboration & Enhancements Drive code-level hotfixes or patches in coordination with Dev, QA, and Cloud Ops. Implement automation scripts for diagnosis, log capture, reprocessing, and health validation. Maintain well-structured GitOps pipelines for support-related patches, rollback plans, and enhancement sprints. Coordinate enhancement requests based on operational analytics and feedback loops. Champion enterprise integration and alignment with Core Banking, ERP, H2H, and transaction processing systems. Governance, Planning & People Leadership Build and mentor a high-caliber AI Support Squad – support engineers, SREs, and automation leads. Define and publish support KPIs , operational dashboards, and quarterly stability scorecards. Present production health reports to business, engineering, and executive leadership. Define runbooks, response playbooks, knowledge base entries, and onboarding plans for newer AI support use cases. Manage relationships with AI platform vendors, cloud ops partners, and application owners. Must-Have Skills & Experience 15+ years of software engineering, platform reliability, or AI systems management experience. Proven track record of leading support and platform operations for AI/ML/GenAI-powered systems . Strong experience with cloud-native platforms (Azure/AWS), Kubernetes , and containerized observability . Deep expertise in Python and/or Java for production debugging and script/tooling development. Proficient in monitoring, logging, tracing, and alerts using enterprise tools (Grafana, ELK, Datadog). Familiarity with token economics , prompt tuning, inference throttling, and GenAI usage policies. Experience working with distributed systems, banking APIs, and integration with Core/ERP systems . Strong understanding of incident management frameworks (ITIL) and ability to drive postmortem discipline . Excellent stakeholder management, cross-functional coordination, and communication skills. Demonstrated ability to mentor senior ICs and influence product and platform priorities. Nice-to-Haves Exposure to enterprise AI platforms like OpenAI, Azure OpenAI, Anthropic, or Cohere. Experience supporting multi-tenant AI applications with business-driven SLAs. Hands-on experience integrating with compliance and risk monitoring platforms. Familiarity with automated root cause inference or anomaly detection tooling. Past participation in enterprise architecture councils or platform reliability forums Show more Show less

Posted 2 weeks ago

Apply

8.0 years

0 Lacs

Noida, Uttar Pradesh, India

On-site

Role name: Automation Test Lead Years of exp : 5 - 8 yrs About Dailoqa Dailoqa’s mission is to bridge human expertise and artificial intelligence to solve the challenges facing financial services. Our founding team of 20+ international leaders, including former CIOs and senior industry experts, combines extensive technical expertise with decades of real-world experience to create tailored solutions that harness the power of combined intelligence. With a focus on Financial Services clients, we have deep expertise across Risk & Regulations, Retail & Institutional Banking, Capital Markets, and Wealth & Asset Management. Dailoqa has global reach in UK, Europe, Africa, India, ASEAN, and Australia. We integrate AI into business strategies to deliver tangible outcomes and set new standards for the financial services industry. Working at Dailoqa will be hard work, our environment is fluid and fast-moving and you'll be part of a community that values innovation, collaboration, and relentless curiosity. We’re looking at people who: Are proactive, curious adaptable, and patient Shape the company's vision and will have a direct impact on its success. Have the opportunity for fast career growth. Have the opportunity to participate in the upside of an ultra-growth venture. Have fun 🙂 Don’t apply if: You want to work on a single layer of the application. You prefer to work on well-defined problems. You need clear, pre-defined processes. You prefer a relaxed and slow paced environment. Role Overview As an Automation Test Lead at Dailoqa, you’ll architect and implement robust testing frameworks for both software and AI/ML systems. You’ll bridge the gap between traditional QA and AI-specific validation, ensuring seamless integration of automated testing into CI/CD pipelines while addressing unique challenges like model accuracy, GenAI output validation, and ethical AI compliance. Key Responsibilities Test Automation Strategy & Framework Design Design and implement scalable test automation frameworks for frontend (UI/UX), backend APIs, and AI/ML model-serving endpoints using tools like Selenium, Playwright, Postman, or custom Python/Java solutions. Build GenAI-specific test suites for validating prompt outputs, LLM-based chat interfaces, RAG systems, and vector search accuracy. Develop performance testing strategies for AI pipelines (e.g., model inference latency, resource utilization). Continuous Testing & CI/CD Integration Establish and maintain continuous testing pipelines integrated with GitHub Actions, Jenkins, or GitLab CI/CD. Implement shift-left testing by embedding automated checks into development workflows (e.g., unit tests, contract testing). AI/ML Model Validation Collaborate with data scientists to test AI/ML models for accuracy, fairness, stability, and bias mitigation using tools like TensorFlow Model Analysis or MLflow. Validate model drift and retraining pipelines to ensure consistent performance in production. Quality Metrics & Reporting Define and track KPIs: Test coverage (code, data, scenarios) Defect leakage rate Automation ROI (time saved vs. maintenance effort) Model accuracy thresholds Report risks and quality trends to stakeholders in sprint reviews. Drive adoption of AI-specific testing tools (e.g., LangChain for LLM testing, Great Expectations for data validation). Technical Requirements Must-Have 5–8 years in test automation, with 2+ years validating AI/ML systems. Expertise in: Automation tools: Selenium, Playwright, Cypress, REST Assured, Locust/JMeter CI/CD: Jenkins, GitHub Actions, GitLab AI/ML testing: Model validation, drift detection, GenAI output evaluation Languages: Python, Java, or JavaScript Certifications: ISTQB Advanced, CAST, or equivalent. Experience with MLOps tools: MLflow, Kubeflow, TFX Familiarity with vector databases (Pinecone, Milvus) and RAG workflows. Strong programming/scripting experience in JavaScript, Python, Java, or similar Experience with API testing, UI testing, and automated pipelines Understanding of AI/ML model testing, output evaluation, and non-deterministic behavior validation Experience with testing AI chatbots, LLM responses, prompt engineering outcomes, or AI fairness/bias Familiarity with MLOps pipelines and automated validation of model performance in production Exposure to Agile/Scrum methodology and tools like Azure Boards Soft Skills Strong problem-solving skills for balancing speed and quality in fast-paced AI development. Ability to communicate technical risks to non-technical stakeholders. Collaborative mindset to work with cross-functional teams (data scientists, ML engineers, DevOps). Show more Show less

Posted 2 weeks ago

Apply

4.0 years

0 Lacs

Bangalore Urban, Karnataka, India

On-site

About Rapido We are India’s largest bike-taxi platform, steadily venturing into Auto, Delivery, Rental, and more. Currently, present in ~100 cities, we are growing close to ~50% year-on-year with steady funding. We have changed the concept of intra-city travel and made last-mile connectivity affordable to all. Along with being the #1 choice of 40 million people, we have also built a solid captain base of over 5 million registered captains, who have bettered their lives with Rapido. As an employer, we provide a lot of ownership to our team members providing multiple avenues for them to grow within the company. You will only grow with us with the right balance of ambition, fun, and transparent work culture! Job Summary: As a Senior Data Scientist your will be responsible for understanding product/business priorities, conducting opportunity analysis, defining product level metrics, setting metric targets, building data-driven hypothesis, and running high-quality experiments to hit those targets.A Senior Data Scientist should have an experienced of solving complex data analysis and machine learning problems in a real-world product setting. The role requires someone who is very comfortable working across teams, good at communicating and able to drive analysis and manage multiple streams. Your role would also require you to mentor junior Data Scientists in their learning and development. You will be part of a dynamic team and regularly work with colleagues from other verticals to co-ordinate and collaborate within an increasingly interconnected Data Science team here @ Rapido. Job Responsibilities: ● Translating business requirements into analytical solutions ● Do hands-on data analysis, using statistical techniques ● Pre-processing of structured and unstructured data ● Should be able to steer direction related to improving data collection for analytical systems ● Should be proficient in model development, validation and then roll-out ● Understand business challenges and goals to formulate the approach for data analysis and model creation ● Tune the model(s) to improve results provided over time ● Create an analytics roadmap for the project ● Collaborate with a technology/data engineering team to transfer the business understanding, get the model productionized and validate the output ● Work in highly collaborative teams that strive to build quality systems and provide business value ● Articulate insights and story telling for the senior management ● Mentor Junior Data team members in their learning and development Job Requirements: ● Work experience - 4 years as a Data Scientist or equivalent position and overall 6 years ● Masters or Ph.D. in Statistics, Math, CS, Econ, Physics, Engineering or related scientific disciplines; ● Passion for understanding business problems and trying to address them by leveraging data - characterized by high-volume, high dimensionality from multiple sources ● Experience with building predictive statistical, behavioral or other models via supervised and unsupervised machine learning, statistical analysis, and other predictive modeling techniques. ● Worked in a big data environment before alongside a data engineering team (and data visualization team, data and business analysts) ● Statistical knowledge - Proven experience in statistical methods like and not limited to Markov Models, Stochastic models, Bayesian Models, Classification Models, Cluster Analysis, Multivariate Stats, Regression Models, Ensemble Techniques ● Machine Learning - Prior work experience in one or more of these knowledge areas (domain agnostic): Price Modeling, Demand Forecasting, Recommender Systems, User Profiling, Fraud Detectors ● Experience in SQL is must ● Technologies - Proficiency in Python (must have) and any other prog. language; Specific libraries may include - Scikit-learn, pandas, TensorFlow, Keras, Torch, Caffe, Theano, etc; ● Experience of Pyspark is a plus ● Experience with A/B Testing framework ● Experience in managing a small team of data scientists is a plus What’s there for you? ● If you are excited to solve Complex Business problems at scale where you will cater to 10 million rides in a day across 100 cities, no better place to be. ● You can go swimming in our 100s of TBs of data on self-managed databases ● Opportunity to contribute in solving challenging problems like supply-demand matching, demand forecasting, ROI optimization, Segmentation, Causal Inference and many more. ● We have a very flat hierarchy and collaboration is our success mantra ● Chance to work with a lean team of highly talented and motivated people across all levels ● You will be working with cross functional teams and have a great learning curve and business exposure. Show more Show less

Posted 2 weeks ago

Apply

3.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Company Qualcomm India Private Limited Job Area Engineering Group, Engineering Group > Software Engineering General Summary As a leading technology innovator, Qualcomm pushes the boundaries of what's possible to enable next-generation experiences and drives digital transformation to help create a smarter, connected future for all. As a Qualcomm Software Engineer, you will design, develop, create, modify, and validate embedded and cloud edge software, applications, and/or specialized utility programs that launch cutting-edge, world class products that meet and exceed customer needs. Qualcomm Software Engineers collaborate with systems, hardware, architecture, test engineers, and other teams to design system-level software solutions and obtain information on performance requirements and interfaces. Minimum Qualifications Bachelor's degree in Engineering, Information Systems, Computer Science, or related field and 3+ years of Software Engineering or related work experience. OR Master's degree in Engineering, Information Systems, Computer Science, or related field and 2+ years of Software Engineering or related work experience. OR PhD in Engineering, Information Systems, Computer Science, or related field and 1+ year of Software Engineering or related work experience. 2+ years of academic or work experience with Programming Language such as C, C++, Java, Python, etc. Job Location: Hyderabad More Details Below About the team: Join the growing team at Qualcomm focused on advancing state-of-the-art in Machine Learning. The team uses Qualcomm chips’ extensive heterogeneous computing capabilities to allow inference of trained neural networks on-device without a need for connection to the cloud. Our inference engine is designed to help developers run neural network models trained in a variety of frameworks on Snapdragon platforms at blazing speeds while still sipping the smallest amount of power. See your work directly impact billions of devices around the world. Responsibilities In this position, you will be responsible for the development and commercialization of ML solutions like Snapdragon Neural Processing Engine (SNPE) SDK on Qualcomm SoCs. You will be developing various SW features in our ML stack. You would be porting AI/ML solutions to various platforms and optimize the performance on multiple hardware accelerators (like CPU/GPU/NPU). You will have expert knowledge in deployment aspects of large software C/C++ dependency stacks using best practices. You will also have to keep up with the fast-paced development happening in the industry and academia to continuously enhance our solution from software engineering as well as machine learning standpoint. Work Experience 7-9 years of relevant work experience in software development. Live and breathe quality software development with excellent analytical and debugging skills. Strong understanding about Processor architecture, system design fundamentals. Experience with embedded systems development or equivalent. Strong development skills in C and C++. Excellent communication skills (verbal, presentation, written). Ability to collaborate across a globally diverse team and multiple interests. Preferred Qualifications Experience in embedded system development. Experience in C, C++, OOPS and Design patterns. Experience in Linux kernel or driver development is a plus. Strong OS concepts. Applicants : Qualcomm is an equal opportunity employer. If you are an individual with a disability and need an accommodation during the application/hiring process, rest assured that Qualcomm is committed to providing an accessible process. You may e-mail disability-accomodations@qualcomm.com or call Qualcomm's toll-free number found here. Upon request, Qualcomm will provide reasonable accommodations to support individuals with disabilities to be able participate in the hiring process. Qualcomm is also committed to making our workplace accessible for individuals with disabilities. (Keep in mind that this email address is used to provide reasonable accommodations for individuals with disabilities. We will not respond here to requests for updates on applications or resume inquiries). Qualcomm expects its employees to abide by all applicable policies and procedures, including but not limited to security and other requirements regarding protection of Company confidential information and other confidential and/or proprietary information, to the extent those requirements are permissible under applicable law. To all Staffing and Recruiting Agencies : Our Careers Site is only for individuals seeking a job at Qualcomm. Staffing and recruiting agencies and individuals being represented by an agency are not authorized to use this site or to submit profiles, applications or resumes, and any such submissions will be considered unsolicited. Qualcomm does not accept unsolicited resumes or applications from agencies. Please do not forward resumes to our jobs alias, Qualcomm employees or any other company location. Qualcomm is not responsible for any fees related to unsolicited resumes/applications. If you would like more information about this role, please contact Qualcomm Careers. 3072303 Show more Show less

Posted 2 weeks ago

Apply

2.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Company Qualcomm India Private Limited Job Area Engineering Group, Engineering Group > Software Engineering General Summary Job Description Join the exciting Generative AI team at Qualcomm focused on integrating cutting edge GenAI models on Qualcomm chipsets. The team uses Qualcomm chips’ extensive heterogeneous computing capabilities to allow inference of GenAI models on-device without a need for connection to the cloud. Our inference engine is designed to help developers run neural network models trained in a variety of frameworks on Snapdragon platforms at blazing speeds while still sipping the smallest amount of power. Utilize this power efficient hardware and Software stack to run Large Language Models (LLMs) and Large Vision Models (LVM) at near GPU speeds! Responsibilities In this role, you will spearhead the development and commercialization of the Qualcomm AI Runtime (QAIRT) SDK on Qualcomm SoCs. As an AI inferencing expert, you'll push the limits of performance from large models. Your mastery in deploying large C/C++ software stacks using best practices will be essential. You'll stay on the cutting edge of GenAI advancements, understanding LLMs/Transformers and the nuances of edge-based GenAI deployment. Most importantly, your passion for the role of edge in AI's evolution will be your driving force. Requirements Master’s/Bachelor’s degree in computer science or equivalent. 2-4 years of relevant work experience in software development. Strong understanding of Generative AI models – LLM, LVM, LMMs and building blocks (self-attention, cross attention, kv caching etc.) Floating-point, Fixed-point representations and Quantization concepts. Experience with optimizing algorithms for AI hardware accelerators (like CPU/GPU/NPU). Strong in C/C++ programming, Design Patterns and OS concepts. Good scripting skills in Python. Excellent analytical and debugging skills. Good communication skills (verbal, presentation, written). Ability to collaborate across a globally diverse team and multiple interests. Preferred Qualifications Strong understanding of SIMD processor architecture and system design. Proficiency in object-oriented software development and familiarity Familiarity with Linux and Windows environment Strong background in kernel development for SIMD architectures. Familiarity with frameworks like llama.cpp, MLX, and MLC is a plus. Good knowledge of PyTorch, TFLite, and ONNX Runtime is preferred. Experience with parallel computing systems and languages like OpenCL and CUDA is a plus. Minimum Qualifications Bachelor's degree in Engineering, Information Systems, Computer Science, or related field and 2+ years of Software Engineering or related work experience. OR Master's degree in Engineering, Information Systems, Computer Science, or related field and 1+ year of Software Engineering or related work experience. OR PhD in Engineering, Information Systems, Computer Science, or related field. 2+ years of academic or work experience with Programming Language such as C, C++, Java, Python, etc. Applicants : Qualcomm is an equal opportunity employer. If you are an individual with a disability and need an accommodation during the application/hiring process, rest assured that Qualcomm is committed to providing an accessible process. You may e-mail disability-accomodations@qualcomm.com or call Qualcomm's toll-free number found here. Upon request, Qualcomm will provide reasonable accommodations to support individuals with disabilities to be able participate in the hiring process. Qualcomm is also committed to making our workplace accessible for individuals with disabilities. (Keep in mind that this email address is used to provide reasonable accommodations for individuals with disabilities. We will not respond here to requests for updates on applications or resume inquiries). Qualcomm expects its employees to abide by all applicable policies and procedures, including but not limited to security and other requirements regarding protection of Company confidential information and other confidential and/or proprietary information, to the extent those requirements are permissible under applicable law. To all Staffing and Recruiting Agencies : Our Careers Site is only for individuals seeking a job at Qualcomm. Staffing and recruiting agencies and individuals being represented by an agency are not authorized to use this site or to submit profiles, applications or resumes, and any such submissions will be considered unsolicited. Qualcomm does not accept unsolicited resumes or applications from agencies. Please do not forward resumes to our jobs alias, Qualcomm employees or any other company location. Qualcomm is not responsible for any fees related to unsolicited resumes/applications. If you would like more information about this role, please contact Qualcomm Careers. 3075196 Show more Show less

Posted 2 weeks ago

Apply

10.0 years

0 Lacs

Trivandrum, Kerala, India

On-site

Our Company Techvantage.ai is a next-generation technology and product engineering company at the forefront of innovation in Generative AI, Agentic AI, and autonomous intelligent systems. We design intelligent platforms that solve complex business problems and deliver measurable impact through cutting-edge AI Overview : We are seeking an experienced Solution Architect with a strong foundation in software architecture and a working knowledge of AI-based products or platforms. In this role, you will be responsible for designing robust, scalable, and secure architectures that support AI-driven applications and enterprise systems. You will work closely with cross-functional teamsincluding data scientists, product managers, and engineering leadsto bridge the gap between business needs, technical feasibility, and AI we are looking from an ideal candidate ? Architect end-to-end solutions for enterprise and product-driven platforms, including components such as data pipelines, APIs, AI model integration, cloud infrastructure, and user interfaces. Guide teams in selecting the right technologies, tools, and design patterns to build scalable systems. Collaborate with AI/ML teams to understand model requirements and ensure smooth deployment and integration into production. Define system architecture diagrams, data flow, service orchestration, and infrastructure provisioning using modern tools. Work closely with stakeholders to translate business needs into technical solutions, with a focus on scalability, performance, and security. Provide leadership on best practices for software development, DevOps, and cloud-native architecture. Conduct architecture reviews and ensure alignment with security, compliance, and performance Skills : What skills do you need ? Requirements 10+ years of experience in software architecture or solution design roles. Proven experience designing systems using microservices, RESTful APIs, event-driven architecture, and cloud-native technologies. Hands-on experience with at least one major cloud provider: AWS, GCP, or Azure. Familiarity with AI/ML platforms or components, such as integrating AI models, MLOps pipelines, or inference services. Understanding of data architectures, including data lakes, streaming, and ETL pipelines. Strong experience with containerization (Docker, Kubernetes) and DevOps principles. Ability to lead technical discussions, make design trade-offs, and communicate with both technical and non-technical Qualifications : Exposure to AI model lifecycle management, prompt engineering, or real-time inference workflows. Experience with infrastructure-as-code (Terraform, Pulumi). Knowledge of GraphQL, gRPC, or serverless architectures. Previous experience working in AI-driven product companies or digital transformation We Offer : High-impact role in designing intelligent systems that shape the future of AI adoption. Work with forward-thinking engineers, researchers, and innovators. Strong focus on career growth, learning, and technical leadership. Compensation is not a constraint for the right candidate. (ref:hirist.tech) Show more Show less

Posted 2 weeks ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies