Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
4.0 - 9.0 years
15 - 30 Lacs
Chennai
Hybrid
ACV Auctions is looking for an experienced Site Reliability Engineer III with a systems and software engineering background to focus on site reliability. We believe in taking a software engineers approach to operations by providing standards and software tools to all engineering projects. As a Site Reliability Engineer, you will split your time between developing software that improves overall reliability and providing operational support for production systems. What you will do: Maintain reliability and performance for your particular infrastructure area while working with software engineers to improve service quality and health. Develop, design, and review new software tools in Python & Java to improve infrastructure reliability and provide services with better monitoring, automation, and product delivery. Practice efficient incident response through on-call rotations alongside software engineers and document incidents through postmortems. Support service development with capacity plans, launch/deployment plans, scalable system design, and monitoring plans. What you will need: BS degree in Computer Science or a related technical discipline or equivalent practical experience. Experience building/managing infrastructure deployments on Google Cloud Platform 3+ years managing cloud infrastructure. Experience programming in at least one of the following: Python or Java You are experienced in Linux/Unix systems administration, configuration management, monitoring, and troubleshooting. You are comfortable with production systems including load balancing, distributed systems, microservice architecture, service meshes, and continuous delivery. Experience building and delivering software tools for monitoring, management, and automation that support production systems. Comfortable working with teams across multiple time -zones and working flexible hours as needed. Preferred Qualifications Experience maintaining and scaling Kubernetes clusters for production workloads is a plus
Posted 1 month ago
5.0 - 10.0 years
15 - 17 Lacs
Navi Mumbai
Work from Office
Strong experience with Apache Kafka and its ecosystem (Kafka Connect, Schema Registry, Kafka Streams). Proficient in Kafka cluster monitoring and performance tuning. Experience with tools such as Prometheus, Grafana, ELK stack. Required Candidate profile Solid knowledge of Linux/Unix system administration Experience with scripting languages like Bash, Python Familiarity with DevOps tools (Ansible, Jenkins) Experience with cloud-based Kafka deployments
Posted 1 month ago
2.0 - 5.0 years
10 - 20 Lacs
Bengaluru
Work from Office
Experience : 2+ years Expected Notice Period : 15 Days Shift : (GMT+05:30) Asia/Kolkata (IST) Must have skills required: Bash, Dynatrace, ELK, Grafana, Prometheus, Terraform, AWS, Kubernetes, ???Linux, Python Job Overview We are looking for a Site Reliability Engineer (SRE) with 2.5 to 5 years of experience to join our team. The ideal candidate will be responsible for ensuring the availability, scalability, and reliability of our distributed systems, improving observability, automating infrastructure, and enhancing system performance. This role provides an opportunity to work on high-scale, mission-critical environments and contribute to building a resilient infrastructure. Key Responsibilities Improve observability by implementing and managing monitoring, logging, and alerting solutions using Prometheus, ELK stack, and Grafana. Work with APMs like Dynatrace, New Relic to monitor performance metrics, define SLIs, SLOs, and error budgets. Participate in incident management, including on-call rotation, and Root Cause Analysis (RCA). Automate infrastructure provisioning using Terraform and Infrastructure as Code (IaC) principles. Ensure system scalability, reliability, and performance in a distributed environment. Strengthen security by applying cybersecurity best practices, vulnerability assessments, and compliance policies. Collaborate with cross-functional teams to establish SRE best practices, improve release pipelines, and minimize deployment risks. Maintain and improve disaster recovery plans to enhance resilience. Manage and optimize workflows using Apache Airflow to ensure efficient scheduling and execution of data pipelines. Support Snowflake data operations, ensuring high availability, performance optimization, and security compliance. Qualifications & Certifications Education: Bachelor's degree in Computer Science, Engineering, or related fields. Experience: 2.5 to 5 years of experience in Site Reliability Engineering, Observability, or Performance Monitoring. Hands-on experience in: Monitoring and observability using Prometheus, ELK, Grafana. Application Performance Monitoring (APM) tools like Dynatrace, New Relic, or Datadog. Incident response and on-call rotation management. Infrastructure automation using Terraform. Distributed systems operations and scaling. Load testing and performance analysis using tools like JMeter, k6, or Locust. Security at scale, including vulnerability scanning and compliance automation. Workflow automation and orchestration using Apache Airflow. Experience with Snowflake, including query optimization, data management, and security controls. Technical Skills: Strong knowledge of cloud platforms (AWS preferred). Experience with troubleshooting distributed systems and high-traffic environments. Hands-on knowledge of Linux, networking, and security fundamentals. Familiarity with container orchestration (Kubernetes, Docker). Ability to write automation scripts using Python, Bash, or Go. Preferred Certifications: AWS Certified DevOps Engineer Professional (or equivalent AWS certification). HashiCorp Certified: Terraform Associate. Certified Kubernetes Administrator (CKA). Google SRE Professional Certificate (preferred but not mandatory). Skills Bash, Dynatrace, ELK, Grafana, Prometheus, Terraform, AWS, Kubernetes, ???Linux, Python
Posted 1 month ago
6.0 - 10.0 years
8 - 12 Lacs
Mumbai
Work from Office
We are looking for an experienced DevOps Engineer (Level 2 & 3) to design, automate, and optimize cloud infrastructure. You will play a key role in CI/CD automation, cloud management, observability, and security, ensuring scalable and reliable systems. Key Responsibilities : Design and manage AWS environments using Terraform/Ansible. Build and optimize deployment pipelines (Jenkins, ArgoCD, AWS CodePipeline). Deploy and maintain EKS, ECS clusters. Implement OpenTelemetry, Prometheus, Grafana for logs, metrics, and tracing. Manage and scale cloud-native microservices efficiently. Required Skills : Proven experience in DevOps, system administration, or software development. Strong knowledge of AWS. Programming languages: Python, Go, Bash, are good to have Experience with IAC tools like Terraform, Ansible Solid understanding of CI/CD tools (Jenkins, ArgoCD , AWS CodePipeline). Experience in containers and orchestration tools like Kubernetes (EKS) Understanding of OpenTelemetry observability stack (logs, metrics, traces) Good to have : Experience with container orchestration platforms (e.g., EKS, ECS). Familiarity with serverless architecture and tools (e.g., AWS Lambda). Experience using monitoring tools like DataDog/ NewRelic, CloudWatch, Prometheus/Grafana Experience with managing more than 20+ cloud-native microservices. Previous experience of working in a startup Education & Experience : Bachelors degree in Computer Science, Information Technology, or a related field (or equivalent work experience). Years of relevant experience in DevOps or a similar role. About Kissht: Kissht, a Great Place to Work certified organization, is a consumer-first credit app that is transforming the landscape of consumer credit. As one of the fastest-growing and most respected FinTech companies, Kissht is a pioneer in data and machine-based lending. With over 15 million customers, including 40% from tier 2 cities and beyond, we offer both short and long-term loans for personal consumption, business needs, and recurring expenses. Founded by Ranvir and Krishnan, alumni of IIT and IIM, and backed by renowned investors like Endiya Partners, the Brunei Investment Authority, and the Singapore Government, Kissht is synonymous with excellence in the industry. Join us and be a part of a dynamic, innovative company that is changing the future of financial technology.
Posted 1 month ago
8.0 - 12.0 years
35 - 50 Lacs
Bengaluru
Work from Office
We are looking for an experienced DevOps Architect to lead the design, development, and maintenance of scalable DevOps solutions. The ideal candidate will have deep expertise in cloud platforms, CI/CD pipelines, automation frameworks, and infrastructure-as-code (IaC) principles. You will work closely with development, IT operations, and security teams to streamline processes and enhance the reliability and scalability of our applications. Key Responsibilities: Architect and Implement : Design and implement scalable, secure, and high-performance DevOps pipelines. Infrastructure as Code (IaC): Manage infrastructure using tools like Terraform, CloudFormation, or Ansible. CI/CD Management : Build and maintain robust CI/CD pipelines for automated testing, integration, and deployment. Cloud Management: Architect solutions on cloud platforms such as AWS, Azure, or GCP. Monitoring and Logging: Set up monitoring tools (Prometheus, Grafana, ELK Stack) to ensure application reliability and performance. Security and Compliance : Implement security best practices in CI/CD pipelines, infrastructure, and cloud environments. Collaboration : Work closely with development and operations teams to automate and optimize their workflows. Documentation : Maintain clear documentation of architecture, configurations, and processes. Skills and Qualifications: Technical Skills: Proficiency in cloud platforms: AWS, Azure, or GCP. Expertise in CI/CD pipelines: Jenkins, GitLab CI, CircleCI, or similar. Hands-on experience with Infrastructure as Code (IaC): Terraform, CloudFormation, Ansible. Strong scripting skills: Bash, Python, or PowerShell. Knowledge of containerization and orchestration: Docker, Kubernetes. Familiarity with monitoring and logging tools: Prometheus, Grafana, ELK Stack. Solid understanding of networking, security, and performance optimization. Soft Skills: Excellent problem-solving and troubleshooting abilities. Strong communication and collaboration skills. Ability to work in an agile environment and adapt to change.
Posted 1 month ago
2.0 - 4.0 years
4 - 9 Lacs
Bengaluru
Work from Office
Skills Required: Technical areas (hands-on experience in academic projects/internships) Experience with Kubernetes, Jenkins, Gitlab, Github, CI/CD, Terraform, Linux, Bash, Python, AWS, GCP, GKE, and EKSUnderstanding of Public/Private/Hybrid Cloud Solutions. Own the responsibility for platform management, supporting services, and all related tooling and automation. Proficient in cloud-native technologies, automation, and containerization. Experience in setting up and managing cloud infrastructure and services for a wide range of Applications. Some experience in ReactJS / NodeJS, PHP, Phyton and UNIX shell,so a background in system- oriented languages is important. Managing and deploying cloud-native applications on Kubernetes clusters, Setting CI/CD pipelines in (Jenkins, Gitlab, Github), Databases Migration (MySQL, Postgresql, Cassandra), Setting up Monitoring (Grafana, Loki, Prometheus, Mimir, ELK Stack). Certified in Kubernetes and Jenkins.Experienced in using Terraform to automate infrastructure provisioning. We are looking for bright, passionate, and dedicated people with clearly demonstrated initiative and a history of success in their past positions to join our growing team.
Posted 1 month ago
6.0 - 10.0 years
15 - 25 Lacs
Noida, Gurugram, Delhi
Work from Office
Mandatory Skills (Docker and Kubernetes) Should have good understanding of various components of Kubernetes cluster Should have hands on experience of provisioning of Kubernetes cluster Should have expertise on managing and upgradation Kubernetes Cluster / Redhat Openshift platform Should have good experience of Container storage Should have good experience on CICD workflow (Preferable Azure DevOps, Ansible and Jenkin) Should have hands on experience of linux operating system administration Should have understanding of Cloud Infrastructure preferably Vmware Cloud Should have good understanding of application life cycle management on container platform Should have basis understanding of cloud networks and container networks Should have good understanding of Helm and Helm Charts Should be good in performance optimization of container platform Should have good understanding of container monitoring tools like Prometheus, Grafana and ELK Should be able to handle Severity#1 and Severity#2 incidents Good communication skills Should have capability to provide the support Should have analytical and problem-solving capabilities, ability to work with teams Should have experience on 24*7 operation support framework) Should have knowledge of ITIL Process Preferred Skills/Knowledge Container Platforms - Docker, CRI/O, Kubernetes and OpenShift Automation Platforms - Shell Scripts, Ansible, Jenkin Cloud Platforms - GCP/AZURE/OpenStack Operating System - Linux/CentOS/Ubuntu Container Storage and Backup
Posted 1 month ago
7.0 - 11.0 years
4 - 7 Lacs
Bengaluru
Work from Office
Skill required: Delivery - Marketing Analytics and Reporting Designation: I&F Decision Sci Practitioner Specialist Qualifications: Any Graduation Years of Experience: 7 to 11 years About Accenture Combining unmatched experience and specialized skills across more than 40 industries, we offer Strategy and Consulting, Technology and Operations services, and Accenture Song all powered by the worlds largest network of Advanced Technology and Intelligent Operations centers. Our 699,000 people deliver on the promise of technology and human ingenuity every day, serving clients in more than 120 countries. Visit us at www.accenture.com What would you do Data & AIAnalytical processes and technologies applied to marketing-related data to help businesses understand and deliver relevant experiences for their audiences, understand their competition, measure and optimize marketing campaigns, and optimize their return on investment. What are we looking for Python (Programming Language)Structured Query Language (SQL)Machine LearningData ScienceWritten and verbal communicationAbility to manage multiple stakeholdersStrong analytical skillsDetail orientationExpertise in AWS, Azure, or Google Cloud for ML workflows.Hands-on experience with Kubernetes, Docker, Jenkins, or GitLab CI/CDFamiliarity with MLflow, TFX, Kubeflow, or SageMaker.Knowledge of Prometheus, Grafana, or similar tools for tracking system health and model performance.Understanding of ETL processes, data pipelines, and big data tools like Spark or Kafka.Proficiency in Git and model versioning best practices. Roles and Responsibilities: In this role you are required to do analysis and solving of moderately complex problems May create new solutions, leveraging and, where needed, adapting existing methods and procedures The person would require understanding of the strategic direction set by senior management as it relates to team goals Primary upward interaction is with direct supervisor May interact with peers and/or management levels at a client and/or within Accenture Guidance would be provided when determining methods and procedures on new assignments Decisions made by you will often impact the team in which they reside Individual would manage small teams and/or work efforts (if in an individual contributor role) at a client or within Accenture Work closely with data scientists, engineers, and DevOps teams to operationalize MLOptimize ML pipelines for performance, cost, and scalability in production.Automate deployment pipelines for ML models, ensuring fast and reliable transitions from development to production environments Set up and manage scalable cloud or on-premise environments for ML workflows. Qualification Any Graduation
Posted 1 month ago
2.0 - 6.0 years
4 - 8 Lacs
Bengaluru
Work from Office
React/Redux, HTML5, CSS3, JavaScript, Python, Django and REST APIs. BS or MS in Computer Science or related field. Strong foundation in Computer Science, with deep knowledge of data structures, algorithms, and software design. Experience with GIT, CI/CD tools, Sentry, Atlassian software and AWS CodeDeploy a plus Contribute with ideas to overall product strategy and roadmap. Improve codebase with continuous refactoring. Self-starter to take ownership of the platform engineering and application development. Work on multiple projects simultaneously and get things done. Take products from prototype to production. Collaborate with team in Sunnyvale, CA to lead 24x7 product development. Bonus: If you have worked on one or more below then highlight those projects when applying: Experience with Time Series DB - M3DB, Prometheus, InfluxDB, OpenTSDB, ELK Stack Experience with visualization tools like Tableau, KeplerGL etc. Experience with MQTT or other IoT communication protocols a plus
Posted 1 month ago
4.0 - 9.0 years
25 - 40 Lacs
Bengaluru
Hybrid
Key Responsibilities Design, develop, and maintain scalable microservices using Python, Go, or Node.js. Build and maintain RESTful APIs to support web and mobile applications. Develop event-driven and asynchronous systems using Apache Kafka. Deploy and manage services using Kubernetes in AWS cloud environments. Work with SQL and NoSQL databases to store and retrieve data efficiently. Write clean, maintainable, and well-tested code following Test-Driven Development (TDD) practices. Ensure systems are fault-tolerant, scalable, and performant under load. Collaborate cross-functionally with frontend engineers, DevOps, and product teams. Participate in code reviews, architecture discussions, and team ceremonies. Continuously improve system design and development workflows. Required Skills & Qualifications 4-7 years of professional backend development experience. Proficiency in Python, Go, or Node.js (at least one language required). Strong understanding and hands-on experience with Microservices architecture. Experience with Kafka or other messaging systems (e.g., RabbitMQ). Solid understanding of both SQL (e.g., PostgreSQL, MySQL) and NoSQL (e.g., MongoDB, DynamoDB) databases. Working experience with AWS services (e.g., EC2, S3, RDS, Lambda). Hands-on experience with Kubernetes and containerized application deployment. Proven experience in writing and maintaining RESTful APIs. Commitment to Test-Driven Development (TDD) and clean coding practices. Strong debugging, problem-solving, and analytical skills. Ability to thrive in a fast-paced, dynamic startup environment. Familiarity with observability and monitoring tools (e.g., Prometheus, Grafana, ELK Stack). Nice to Have Exposure to gRPC APIs. Familiarity with Large Language Models (LLMs) and their integration into applications. Experience with Voice Technologies, such as speech recognition, text-to-speech (TTS), conversational AI Understanding of real-time streaming and event-driven systems.
Posted 1 month ago
7.0 - 9.0 years
27 - 35 Lacs
Bengaluru
Hybrid
Expected Notice Period: 15 Days Shift: (GMT+05:30) Asia/Kolkata (IST) What do you need for this opportunity Must have skills required: DICOM), healthcare interoperability standards (FHIR, HL7, Contract Testing, JMeter, Automation Testing, AWS, Docker, JavaScript, Kubernetes A Funded healthcare data and application integration platform is Looking for: As a Software Test Engineer - Automation, you will be a critical part of our quality assurance process. You will design, implement, and maintain robust testing frameworks with a strong emphasis on automation, contract testing, and performance validation. Your contributions will ensure that our platform delivers secure, reliable, and scalable solutions for healthcare interoperability. Things You Could Be Working On Develop and maintain automated test suites on unit, integration and end-to-end level. Implement and manage contract testing frameworks (e.g., Pact) to validate service interactions and ensure seamless integration. Validate compliance with healthcare interoperability standards such as FHIR, HL7, and DICOM. Perform performance and load testing to ensure the platform handles high data volumes effectively. Collaborate with developers to establish and optimize CI/CD pipelines for automated testing. Enhance the testing process by integrating the tests into CI/CD pipelines for real-time validation. Execute exploratory, security, and regression testing for critical releases. Test role-based access control (RBAC) mechanisms to ensure secure access to sensitive healthcare data. In Our Engineering Team We value collaboration to solve complex problems and build impactful solutions. We actively contribute to and own the roadmap. We perform thorough code reviews and ensure comprehensive testing at every stage. We prioritize quality over quantity and aim for engineering excellence. We promote an inclusive and collaborative culture where everyone can contribute across projects. We are committed to professional growth through mentorship and learning opportunities. We balance remote-first flexibility with in-person collaboration twice a week in our office. Our Technology Stack Testing Frameworks and Tools: Cypress, Jest, Pact (for contract testing), Postman, JMeter, Playwright. Programming Languages: JavaScript, TypeScript, Java, GoLang Infrastructure and CI/CD: Docker, Kubernetes, AWS, GitLab CI/CD. Healthcare Standards: FHIR, HL7, DICOM. Monitoring and Logging: Prometheus, Elastic, Sentry, Unleash. Working Methodology: Shape Up. Our Ideal Candidate Experience: At least 5 years in software testing, with hands-on experience in automated (mandatory) and manual testing. Contract Testing: Experience implementing and managing contract testing frameworks like Pact. Automation Skills: Proficiency in designing and maintaining test automation frameworks using tools like Cypress, Selenium, or Postman. Programming Knowledge: JavaScript, TypeScript, Python, Golang, Java. Performance Testing: Exposure to performance testing tools like JMeter or equivalent for testing high-scale systems. Cloud Infrastructure: Familiarity with AWS, Docker, and Kubernetes. Healthcare Standards: Working knowledge of testing products related to healthcare interoperability standards (FHIR, HL7, DICOM) is mandatory. Soft Skills: Strong analytical skills, attention to detail, and ability to work collaboratively with cross-functional teams. Mindset: A proactive approach to learning new technologies and continuously improving testing processes.
Posted 1 month ago
8.0 - 13.0 years
85 - 90 Lacs
Noida
Work from Office
About the Role We are looking for a Staff EngineerReal-time Data Processing to design and develop highly scalable, low-latency data streaming platforms and processing engines. This role is ideal for engineers who enjoy building core systems and infrastructure that enable mission-critical analytics at scale. Youll work on solving some of the toughest data engineering challenges in healthcare. A Day in the Life Architect, build, and maintain a large-scale real-time data processing platform. Collaborate with data scientists, product managers, and engineering teams to define system architecture and design. Optimize systems for scalability, reliability, and low-latency performance. Implement robust monitoring, alerting, and failover mechanisms to ensure high availability. Evaluate and integrate open-source and third-party streaming frameworks. Contribute to the overall engineering strategy and promote best practices for stream and event processing. Mentor junior engineers and lead technical initiatives. What You Need 8+ years of experience in backend or data engineering roles, with a strong focus on building real-time systems or platforms. Hands-on experience with stream processing frameworks like Apache Flink, Apache Kafka Streams, or Apache Spark Streaming. Proficiency in Java, Scala, or Python or Go for building high-performance services. Strong understanding of distributed systems, event-driven architecture, and microservices. Experience with Kafka, Pulsar, or other distributed messaging systems. Working knowledge of containerization tools like Docker and orchestration tools like Kubernetes. Proficiency in observability tools such as Prometheus, Grafana, OpenTelemetry. Experience with cloud-native architectures and services (AWS, GCP, or Azure). Bachelor's or Masters degree in Computer Science, Engineering, or a related field.
Posted 1 month ago
10.0 - 15.0 years
20 - 35 Lacs
Gurugram
Work from Office
Key Responsibilities Design and implement scalable, secure, and highly available infrastructure solutions. Architect and maintain CI/CD pipelines for efficient, reliable software delivery. Drive adoption of DevOps tools, practices, and automation across engineering teams. Lead cloud infrastructure strategy (AWS/GCP/Azure), cost optimization, and security controls. Implement Infrastructure as Code (IaC) using tools like Terraform, Cloud Formation, or Pulumi. Ensure monitoring, alerting, and observability best practices (Prometheus, ELK, Datadog, etc.). Guide container orchestration using Docker, Kubernetes, EKS, or AKS. Collaborate with development, QA, and security teams to ensure high-quality delivery. Mentor and support a team of DevOps engineers and promote a DevSecOps culture. Participate in architecture discussions and contribute to system design decisions. Required Skills & Qualifications 10+ years of experience in DevOps, Site Reliability Engineering, or Infrastructure Engineering. Expertise in at least one cloud platform (AWS, Azure, or GCP). Strong hands-on experience with CI/CD tools (Jenkins, GitLab CI/CD, CircleCI, etc.). Proficiency with Docker, Kubernetes, Helm. Infrastructure as Code (Terraform, Cloud Formation, Ansible). Solid scripting skills in Bash, Python, or Go. Deep understanding of networking, security, and cloud architecture patterns. Strong knowledge of monitoring/logging tools like Prometheus, Grafana, ELK, or Splunk. Experience with version control systems (Git, GitHub, Bitbucket). Excellent problem-solving and communication skills.
Posted 1 month ago
8.0 - 13.0 years
85 - 90 Lacs
Noida
Work from Office
About the Role We are seeking a highly skilled Staff Engineer to lead the architecture, development, and scaling of our Marketplace platform - including portals & core services such as Identity & Access Management (IAM), Audit, and Tenant Management services. This is a hands-on technical leadership role where you will drive engineering excellence, mentor teams, and ensure our platforms are secure, compliant, and built for scale. A Day in the Life Design and implement scalable, high-performance backend systems for all the platform capabilities Lead the development and integration of IAM, audit logging, and compliance frameworks, ensuring secure access, traceability, and regulatory adherence. Champion best practices for reliability, availability, and performance across all marketplace and core service components. Mentor engineers, conduct code/design reviews, and establish engineering standards and best practices. Work closely with product, security, compliance, and platform teams to translate business and regulatory requirements into technical solutions. Evaluate and integrate new technologies, tools, and processes to enhance platform efficiency, developer experience, and compliance posture. Take end-to-end responsibility for the full software development lifecycle, from requirements and design through deployment, monitoring, and operational health. What You Need 8+ years of experience in backend or infrastructure engineering, with a focus on distributed systems, cloud platforms, and security. Proven expertise in building and scaling marketplace platforms and developer/admin/API portals. Deep hands-on experience with IAM, audit logging, and compliance tooling. Strong programming skills in languages such as Python or Go. Experience with cloud infrastructure (AWS, Azure), containerization (Docker, Kubernetes), and service mesh architectures. Understanding of security protocols (OAuth, SAML, TLS), authentication/authorization, and regulatory compliance. Demonstrated ability to lead technical projects and mentor engineering teams & excellent problem-solving, communication, and collaboration skills. Proficiency in observability tools such as Prometheus, Grafana, OpenTelemetry. Prior experience with Marketplace & Portals Bachelor's or Masters degree in Computer Science, Engineering, or a related field
Posted 1 month ago
8.0 - 12.0 years
15 - 27 Lacs
Pune
Hybrid
*****GenAI DevOps Engineer AWS Bedrock***** *****Pune Hinjewadi***** *****Immediate Joiners Preferred***** *****Minimum 4 Days WFO***** Job Description: We are seeking a highly experienced GenAI DevOps Engineer to join our dynamic team in Pune. The ideal candidate will have a strong background in building, deploying, and optimizing Generative AI applications on AWS Bedrock, along with expertise in DevOps practices. You will be responsible for automating infrastructure, managing CI/CD pipelines, and ensuring high performance and reliability of AI models. Key Responsibilities: Design, develop, and deploy Generative AI applications leveraging AWS Bedrock and SageMaker. Automate infrastructure provisioning and deployment processes. Build and maintain robust CI/CD pipelines using CodePipeline and CodeBuild. Monitor application and model performance using CloudWatch, Prometheus, and Grafana. Optimize AI models for performance, scalability, and cost-efficiency. Work with RAG (Retrieval-Augmented Generation) tools such as LangChain, Haystack, and LlamaIndex for building advanced AI solutions. Collaborate with data scientists and developers to streamline model deployment and monitoring. Required Skills: Extensive hands-on experience with AWS Bedrock and SageMaker. Strong expertise in CI/CD tools: CodePipeline, CodeBuild. Proficiency with monitoring tools: CloudWatch, Prometheus, Grafana. Experience with RAG frameworks like LangChain, Haystack, and LlamaIndex. Solid understanding of DevOps best practices and automation. Ability to troubleshoot and optimize AI deployment pipelines. Excellent problem-solving and communication skills. Preferred Skills: Knowledge of containerization (Docker, Kubernetes). Familiarity with scripting languages (Python, Bash). Experience with cloud security best practices. Understanding of machine learning lifecycle management. Mandatory Skills: AWS Bedrock and SageMaker expertise. CI/CD pipeline automation. Monitoring and performance optimization. RAG-based application development.
Posted 1 month ago
3.0 - 8.0 years
3 - 7 Lacs
Noida
Hybrid
Job Title: DevOps Engineer (Kubernetes & Terraform) Location: Noida Experience: 3 to 8 years Type: Full-time About the Role: We are looking for a DevOps Engineer with 38 years of experience who specializes in Kubernetes and Terraform. This role is ideal for someone passionate about automation, infrastructure scalability, and cloud-native technologies. You will be responsible for designing and maintaining infrastructure platforms that support continuous delivery and scalability across our development and production environments. Key Responsibilities: Design, deploy, and manage scalable and secure Kubernetes clusters in production. Develop and manage Infrastructure as Code (IaC) using Terraform to provision cloud infrastructure. Build and maintain CI/CD pipelines to automate build, test, and deployment workflows. Ensure system availability, performance, and security across all environments. Work closely with development and QA teams to enable efficient DevOps practices. Automate system provisioning, configuration, and application deployments. Monitor infrastructure using tools like Prometheus, Grafana, ELK, or similar. Implement security best practices in container orchestration and infrastructure management. Must-Have Qualifications: 3-8 years of experience in DevOps, SRE, or infrastructure engineering roles. Hands-on experience with Kubernetes (deployment patterns, Helm, RBAC, ingress controllers, etc.). Proficiency in Terraform, including module creation and state management. Strong background in at least one public cloud provider (AWS, Azure, or GCP). Experience with CI/CD tools such as Jenkins, GitLab CI, GitHub Actions, or ArgoCD. Solid Linux administration skills. Experience with containerization using Docker. Scripting skills in Bash, Python, or Go. What You'll Get: Competitive compensation and benefits. Exposure to cutting-edge DevOps tools and practices. A collaborative, remote-friendly engineering culture. Opportunities for upskilling and certifications. Involvement in end-to-end infrastructure design and decisions.
Posted 1 month ago
3.0 - 8.0 years
10 - 20 Lacs
Pune
Hybrid
Lead Site Reliability Engineer Lead Site Reliability Engineers at UKG are critical team members that have a breadth of knowledge encompassing all aspects of service delivery. They develop software solutions to enhance, harden and support our service delivery processes. This can include building and managing CI/CD deployment pipelines, automated testing, capacity planning, performance analysis, monitoring, alerting, chaos engineering and auto remediation. Lead Site Reliability Engineers must be passionate about learning and evolving with current technology trends. They strive to innovate and are relentless in pursuing a flawless customer experience. They have an automate everything mindset, helping us bring value to our customers by deploying services with incredible speed, consistency, and availability. Job Responsibilities: Engage in and improve the lifecycle of services from conception to EOL, including system design consulting, and capacity planning Define and implement standards and best practices related to: System Architecture, Service delivery, metrics and the automation of operational tasks Support services, product & engineering teams by providing common tooling and frameworks to deliver increased availability and improved incident response. Improve system performance, application delivery and efficiency through automation, process refinement, postmortem reviews, and in-depth configuration analysis Collaborate closely with engineering professionals within the organization to deliver reliable services Increase operational efficiency, effectiveness, and quality of services by treating operational challenges as a software engineering problem (reduce toil) Guide junior team members and serve as a champion for Site Reliability Engineering Actively participate in incident response, including on-call responsibilities Partner with stakeholders to influence and help drive the best possible technical and business outcomes Required Qualifications Engineering degree, or a related technical discipline, or equivalent work experience Experience coding in higher-level languages (e.g., Python, JavaScript, C++, or Java) Knowledge of Cloud based applications & Containerization Technologies Demonstrated understanding of best practices in metric generation and collection, log aggregation pipelines, time-series databases, and distributed tracing Working experience with industry standards like Terraform, Ansible Demonstrable fundamentals in 2 of the following: Computer Science, Cloud architecture, Security or Network Design fundamentals Demonstrable fundamentals in 2 of the following: Computer Science, Cloud architecture, Security, or Network Design fundamentals (Experience, Education, Certification, License and Training) Must have at least 5 years of hands-on experience working in Engineering or Cloud Minimum 5 years' experience with public cloud platforms (e.g. GCP, AWS, Azure) Minimum 3 years' Experience in configuration and maintenance of applications and/or systems infrastructure for large scale customer facing company Experience with distributed system design and architecture
Posted 1 month ago
5.0 - 7.0 years
10 - 19 Lacs
Bengaluru
Hybrid
Looking for Immediate Joiners Perform Incident Management and Change Management to maintain the continuous availability of all Cloud Infrastructure services. Ensure all SRE and operating procedures are maintained and executed. Maintain a 24x7 production environment with a high level of service availability and perform quality reviews, manage operational issues. Perform root cause analysis for major incidents and drive the process by involving required stakeholders. Perform problem management by analyzing metrics, alarms and dashboards to troubleshoot problem areas, report issues to assist in performance tuning and fault finding. Implementation of proactive monitoring, alerting, trend analysis, and self-healing solutions. Explore and innovate new technologies, features, and tools to improve the platform and automate operational tasks using Bash, Python or any other programming language. Manage and maintain Runbooks and Standard Operating procedures Manage, coordinate, and document all types of maintenance activities and outages. Perform patching and upgrades for vulnerability management. Work closely with the teams to initiate the development of new ideas into internal tools. Understand the existing architecture and work with various Engineering teams to develop and execute strategies to provide a high-quality production service. Capable of working a flexible work schedule in a 24 x 7 environment with rotational shifts Qualifications: Bachelors degree in computer science, electrical engineering or a related area, with 7+ years of SRE experience in a large enterprise organization System admin experience on Linux environments. Experience with end-to-end monitoring setup for infra and applications Experience with Prometheus, Grafana, ELK, Opensearch, Cloudwatch, PagerDuty and other monitoring tools. Solid experience with Cloud Technologies such as AWS and OCI. Good experience with containerized workloads tools like Kubernetes. Network knowledge (TCP/IP, UDP, DNS, Load balancing) and prior network administration experience is required. Experience with BGP, NAT, TCP/IP, iBGP, Proxies, Cross connects. Experience with L2/L3 switching, knowledge of Juniper and Cisco routing devices. Experience understanding and managing web servers (Apache, Tomcat, Nginx) Ability to script/program with one or more high level languages, such as Python, Go, etc. Experience with any configuration management tools like Salt or Puppet or Ansible or similar. Experience with source control tools such as Github and SVN. Experience with deployment tools Jenkins, Harness etc. Experience with SQL and NoSQL databases like Redis, CouchBase, Cassandra, Crate, Elasticsearch. Experience in performing and writing Root Cause Analysis documents. Strong communication and analytical/problem-solving skills. Systematic approach and to drive problems to resolution. Good to have experience/knowledge of GCP, Azure Experience in Security domain will be added advantageRole & responsibilities Preferred candidate profile
Posted 1 month ago
3.0 - 8.0 years
15 - 30 Lacs
Bengaluru
Remote
Hiring for USA based big Multinational Company (MNC) The Cloud Engineer is responsible for designing, implementing, and managing cloud-based infrastructure and services. This role involves working with cloud platforms such as AWS, Microsoft Azure, or Google Cloud to ensure scalable, secure, and efficient cloud environments that meet the needs of the organization. Design, deploy, and manage cloud infrastructure in AWS, Azure, GCP, or hybrid environments. Automate cloud infrastructure provisioning and configuration using tools like Terraform, Ansible, or CloudFormation. Ensure cloud systems are secure, scalable, and reliable through best practices in architecture and monitoring. Work closely with development, operations, and security teams to support cloud-native applications and services. Monitor system performance and troubleshoot issues to ensure availability and reliability. Manage CI/CD pipelines and assist in DevOps practices to streamline software delivery. Implement and maintain disaster recovery and backup procedures. Optimize cloud costs and manage billing/reporting for cloud resources. Ensure compliance with data security standards and regulatory requirements. Stay current with new cloud technologies and make recommendations for continuous improvement. Bachelors degree in Computer Science, Information Technology, Engineering, or a related field. 3+ years of experience working with cloud platforms such as AWS, Azure, or Google Cloud. Proficiency in infrastructure as code (IaC) tools (e.g., Terraform, CloudFormation). Experience with CI/CD tools (e.g., Jenkins, GitLab CI, Azure DevOps). Familiarity with containerization and orchestration (e.g., Docker, Kubernetes). Strong scripting skills (e.g., Python, Bash, PowerShell). Solid understanding of networking, security, and identity management in the cloud. Excellent problem-solving and communication skills. Ability to work independently and as part of a collaborative team.
Posted 1 month ago
8.0 - 12.0 years
30 - 45 Lacs
Hyderabad
Work from Office
Responsibilities: Design, implement, and maintain scalable cloud infrastructure primarily on AWS, with some exposure to Azure. Manage and optimize CI/CD pipelines using Jenkins and Git-based version control systems (GitHub/GitLab). Build and maintain containerized applications using Docker, Kubernetes (including AWS EKS), and Helm. Automate infrastructure provisioning and configuration using Terraform and Ansible. Implement GitOps-style deployment processes using ArgoCD and similar tools. Ensure observability through monitoring and logging with Prometheus, Grafana, Datadog, Splunk, and Kibana. Develop automation scripts using Python, Shell, and GoLang Implement and enforce security best practices in CI/CD pipelines and container orchestration environments using tools like Trivy, OWASP, SonarQube, Aqua Security, Cosign, and HashiCorp Vault. Support blue/green deployments and other advanced deployment strategies. Required Qualifications: 8-12 years of professional experience in a DevOps, SRE, or related role. Strong hands-on experience with AWS (EC2, S3, IAM, EKS, RDS, Lambda, Secrets Manager). Solid experience with CI/CD tools (Jenkins, GitHub/GitLab, Maven). Proficient with containerization and orchestration tools: Docker, Kubernetes, Helm. Experience with Infrastructure as Code tools: Terraform and Ansible. Proficiency in scripting languages: Python, Shell; GoLang Strong understanding of observability, monitoring, and logging frameworks. Familiarity with security practices and tools integrated into DevOps workflows. Excellent problem-solving and troubleshooting skills. Certifications (good to have): AWS Certified DevOps Engineer Certified Kubernetes Administrator (CKA) Azure Administrator/Developer Certifications
Posted 1 month ago
5.0 - 10.0 years
12 - 15 Lacs
Bengaluru
Hybrid
Job Description We are seeking a skilled and proactive AWS DevOps Engineer to join our growing team. You will be responsible for managing scalable infrastructure, automating deployments, monitoring environments, and ensuring optimal performance and security across cloud-based systems. If you're passionate about automation, cloud technologies, and system reliability wed love to hear from you! Key Responsibilities Design, manage, and optimize AWS infrastructure components (EC2, S3, RDS, IAM, VPC, Lambda, etc.). Develop and maintain automation scripts using Bash , Python , or PowerShell for operations, deployments, and monitoring. Implement monitoring and alerting systems using CloudWatch , Datadog , Prometheus , or similar tools. Automate infrastructure provisioning through Infrastructure as Code (IaC) tools like Terraform , CloudFormation , or AWS CDK . Enforce security best practices (IAM policies, encryption, logging, patch management). Manage incident response, conduct root cause analysis, and resolve production issues efficiently. Support and enhance CI/CD pipelines using tools like Jenkins , AWS CodePipeline , GitHub Actions , etc. Monitor and optimize cost, performance, and resource utilization across environments. Ensure robust backup and disaster recovery strategies for cloud workloads. Participate in on-call rotations and respond to high-priority alerts when necessary. Nice to Have AWS Certifications : AWS Certified SysOps Administrator or Solutions Architect. Experience with Kubernetes , ECS , or EKS . Familiarity with Ansible , Chef , or other configuration management tools. Exposure to multi-cloud or hybrid-cloud environments. Experience working in regulated environments (e.g., healthcare, finance, government). Why Join Us? Opportunity to work with a high-performing, collaborative DevOps team. Exposure to cutting-edge cloud technologies. Dynamic work culture with a strong emphasis on innovation and continuous learning. Interested candidates can apply here or send your resume to srinivas.appana@relevancelab.com
Posted 1 month ago
2.0 - 5.0 years
20 - 22 Lacs
Bengaluru
Work from Office
As a Systems Engineer, you are a key member of the Infrastructure Development and Optimization team, which provides support for server, database, network, and storage solutions. In this role, you will be responsible for delivering quality technical solutions to clients, as you continue to build your technical skillset in multiple domains. As an established performer of the practice, you will also be responsible for providing guidance to Associate Systems Engineers Role & responsibilities Assist in the design and implementation of infrastructure solutions, including servers, networks, storage, and virtualization platforms. Monitor system performance, analyze trends, and identify opportunities for optimization and efficiency improvements. Work to deliver tasks within the assigned timeframe, and proactively seeks new work within projects (may require supervision on work). Provide detailed and accurate information, and often owns the task of providing documentation on work performed that can be used by others. Implement performance tuning measures to enhance system reliability, scalability, and responsiveness. Implement and maintain security controls, policies, and procedures to protect client infrastructure and data. Conduct security assessments, audits, and risk assessments, and recommend remediation measures to address vulnerabilities and compliance gaps. Communicate technical concepts and solutions effectively, manage client expectations, and address concerns in a timely and professional manner. • Build and maintain strong relationships with clients, serving as a trusted advisor on infrastructure matters. Serve as a subject matter expert in troubleshooting complex infrastructure issues, providing guidance and mentorship to junior engineers. Investigate root causes of incidents, implement solutions, and document resolution steps for future reference. Stay abreast of industry trends, emerging technologies, and best practices in infrastructure management. Identify opportunities for innovation, process improvement, and automation to drive operational excellence and enhance service delivery. Takes ownership of tasks within their skillset. Proactively provides any updates to task progress, risks, and roadblocks prior to Project or Standup Meetings. Provides guidance to Associate Systems Engineers on project tasks in a project assigned by a more senior Engineer or Architect. Contributes towards best practices within their SME skillsets. May occasionally be involved in engagement development as a career growth opportunity. May be expected to achieve certifications in their area of skillset for Vervint certification needs (partner level required). Working to achieve mastery in multiple skillsets. Preferred candidate profile • Demonstrated ability to perform tasks with general guidance, performing tasks that are generally laid out and defined. • Strong knowledge of cloud computing platforms, such as AWS, Azure, or Google Cloud, and experience with infrastructure-as-code (IaC) tools. • Demonstrated proficiency in virtualization technologies and containerization platforms. • Strong understanding of security principles, protocols, and technologies, including firewalls, intrusion detection/prevention systems, and encryption. • Ability to work independently and collaboratively in a fast-paced, dynamic environment, managing multiple priorities and deadlines effectively. • Strong analytical and problem-solving abilities, with a focus on delivering innovative and sustainable solutions to complex technical challenges. • Excellent communication and interpersonal skills, with the ability to interact effectively with clients, colleagues, and vendors at all levels of the organization. • Ability to provide solutions to technical problems of moderate scope and complexity. • Desire to expand technical skillset through training, certifications, and/or exposure to a variety of assessments. • Time management skills, delivering project tasks on time. • Collaborative mindset with demonstrated ability to train, mentor and coach other engineers.
Posted 1 month ago
1.0 - 3.0 years
2 - 4 Lacs
Kolkata
Hybrid
Required Skills Strong proficiency in Python (3.x) and Django (2.x/3.x/4.x) Hands-on experience with Django REST Framework (DRF) Expertise in relational databases like PostgreSQL or MySQL Proficiency with Git and Bitbucket Solid understanding of RESTful API design and integration Experience in domain pointing and hosting setup on AWS or GCP Deployment knowledge on EC2 , GCP Compute Engine , etc. SSL certificate installation and configuration Familiarity with CI/CD pipelines (GitHub Actions, Bitbucket Pipelines, GitLab CI) Basic usage of Docker for development and containerization Ability to independently troubleshoot server/deployment issues Experience managing cloud resources like S3 , Load Balancers , and IAM roles Preferred Skills Experience with Celery and Redis / RabbitMQ for asynchronous task handling Familiarity with front-end frameworks like React or Vue.js Exposure to Cloudflare or similar CDN/DNS tools Experience with monitoring tools: Prometheus , Grafana , Sentry , or CloudWatch Why Join Us? Work on impactful and modern web solutions Growth opportunities across technologies and cloud platforms Collaborative, inclusive, and innovation-friendly work environment Exposure to challenging and rewarding projects
Posted 1 month ago
7.0 - 11.0 years
12 - 20 Lacs
Mumbai
Hybrid
Job Title: DevOps Engineer Location: Mumbai Work mode: Onsite Notice Period: Immediate iSource Services is hiring for one of their client for the position of DevOps Engineer. Skills: DevOps (AWS, Jenkins, k8, Prometheus, Splunk, Grafana), PHP Framework OOP Responsibilities: Security patches QCR/Compliance Bug fixes. L3 escalations including PDP dropin (FE) BF/Holiday. Special events preparation Testing (e2e, performance) Release validation. Deployment on pre-prod environment. Monitoring and alerting changes Monitor AWS resources, K8 clusters On-call duties.
Posted 1 month ago
7.0 - 12.0 years
8 - 12 Lacs
Mumbai
Work from Office
iSource Services is hiring for one of their client for the position of DevOps Engineer. Skills: DevOps (AWS, Jenkins, k8, Prometheus, Splunk, Grafana), PHP Framework OOP Responsibilities: Security patches QCR/Compliance Bug fixes. L3 escalations including PDP dropin (FE) BF/Holiday. Special events preparation Testing (e2e, performance) Release validation. Deployment on pre-prod environment. Monitoring and alerting changes Monitor AWS resources, K8 clusters On-call duties.
Posted 1 month ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
32455 Jobs | Dublin
Wipro
16590 Jobs | Bengaluru
EY
11025 Jobs | London
Accenture in India
10991 Jobs | Dublin 2
Amazon
8878 Jobs | Seattle,WA
Uplers
8715 Jobs | Ahmedabad
IBM
8204 Jobs | Armonk
Oracle
7750 Jobs | Redwood City
Capgemini
6181 Jobs | Paris,France
Muthoot FinCorp (MFL)
6170 Jobs | New Delhi