Jobs
Interviews

900 Prometheus Jobs - Page 11

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

12.0 - 16.0 years

40 - 45 Lacs

Pune

Remote

What You'll Do Job Summary We are looking for a Senior Technical Lead with expertise in designing SAAS applications and integrations for scale and expertise with full stack technologies development to join our globally distributed Electronic Invoicing & Live Reporting (ELR) team and assist us to become a global leader in the e-invoicing market, and part of every transaction in the world! We have a phenomenal team working in an open, collaborative environment that makes taxes and compliance less taxing to deal with. It will be up to you and the team to convert the product vision and requirements into a finished product. You will be Reporting to Senior Engineering Manager. You will work as individual contributor and you will not have managerial Responsibilities. You will work Remotely in India. What Your Responsibilities Will Be Job Responsibilities Avalara e-Invoicing and Platforms: Dig into our multi-patented, cloud-native Avalara product suite. We are building a flexible, platform that can handle any opportunity to create and submit electronic invoices and live reporting processes for any industry in any geography. Work with your team to create that while maximizing performance, scalability, and reliability while making it 'oh-so-simple' to operate. Design With The Vision In Mind Code, Review, Commit Create Industry-Leading Products Automation vs. People Power. Computers are great for process automation, but there's a limit to what they can do. You and the team will the unique challenges at the intersection of software Provide Technical guidance and mentoring the engineers in the team. What Youll Need To Be Successful Qualifications You have experience delivering high-quality features to production with expertise in service-oriented architectures, microservices, and web application development. You understand system performance trade-offs, load balancing, and high availability engineering. We're looking for a full-stack developer with expertise in Java, Node.js, and Python, so adaptability is valued. Experience with Java, React, microservices, web services, and REST APIs. We also use MySQL and PostgresDB as our primary transactional RDBMS We're expanding our cloud tech stack with Redis, Snowflake, Prometheus, Kafka, Kinesis, and Grafana. We use Docker for containerization, Kubernetes for orchestration, and AWS, though Azure and GCP backgrounds are welcomed. Collaborate with other teams to solve challenges and improve code to improve application efficiency. Prior experience working in e Invoicing. A Bachelor in Computer Science, Engineering, or related field is desirable. 12+ Years work experience Required.

Posted 1 week ago

Apply

11.0 - 17.0 years

45 - 50 Lacs

Pune

Work from Office

: Job Title: Fintech Product Engineering Lead Corporate Title: VP Location: Pune, India Role Description Engineer is responsible for managing or performing work across multiple areas of the bank's overall IT Platform/Infrastructure including analysis, development, and administration. It may also involve taking functional oversight of engineering delivery for specific departments. Work includes: Planning and developing entire engineering solutions to accomplish business goals Building reliability and resiliency into solutions with appropriate testing and reviewing throughout the delivery lifecycle Ensuring maintainability and reusability of engineering solutions Ensuring solutions are well architected and can be integrated successfully into the end-to-end business process flow Reviewing engineering plans and quality to drive re-use and improve engineering capability Participating in industry forums to drive adoption of innovative technologies, tools and solutions in the Bank. Deutsche Banks Corporate Bank division is a leading provider of cash management, trade finance and securities finance. We complete green-field projects that deliver the best Corporate Bank - Securities Services products in the world. Our team is diverse, international, and driven by shared focus on clean code and valued delivery. At every level, agile minds are rewarded with competitive pay, support, and opportunities to excel. You will work as part of a cross-functional agile delivery team. You will bring an innovative approach to software development, focusing on using the latest technologies and practices, as part of a relentless focus on business value. You will be someone who sees engineering as team activity, with a predisposition to open code, open discussion and creating a supportive, collaborative environment. You will be ready to contribute to all stages of software delivery, from initial analysis right through to production support. What well offer you 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Your Key Responsibilities: Ability to navigate a strong sense of urgency while maintaining focus and clarity. Skilled at solving complex design challenges independently , without needing oversight. Proven track record of quickly delivering high-quality code and features . Able to inspire and energise teams through urgency, ownership, and technical excellence. Willing to do whatever it takes to ensure product success , from strategy to hands-on execution. Deep experience in architecting scalable systems (HLD & LLD) in fast-paced environments. Comfortable leading through ambiguity, change, and high-growth pressure . Known for balancing speed with engineering quality and operational readiness . Strong communicator can align teams, resolve conflicts, and drive decisions fast. A true builder mindset acts with ownership, speed, and high accountability . Your skills and experience Hands-on experience in building responsive UIs withReact and Javascript. Hands-on knowledge ofGo(Golang)/ Java and GIN/ SpringBoot framework for backend development. Proficient in HTML, CSS and styling tools like Tailwind. Proficient inRESTful, GraphQLandgRPCfor building scalable and high-performance APIs. Experience with GCP/AWS, for building scalable, resilient micro-service based architectures. Experience with relational and NoSQL databases (e.g.,PostgreSQL,MySQL,Firestore,BigTable). Experience with logging, monitoring and alerting using ( egGrafana, Prometheus, ELK ) Familiarity with CI/CD pipelines, automated testing and deployment strategies with detailed knowledge on Terrafom. Knowledge of best practices for building secure applications (e.g., mTLS, Encryption, OAuth, JWT and Data Compliance). Knowledge of disaster recovery, zero-downtime deploys, and backup strategies How well support you

Posted 1 week ago

Apply

10.0 - 15.0 years

35 - 40 Lacs

Pune

Work from Office

: Job TitleLead Engineer, VP LocationPune, India Role Description Vice President Core Engineering (Technical Leadership Role) We are seeking a highly skilled and experienced Vice President of Engineering to lead the design, development, and maintenance of our core software systems and infrastructure. This is a purely technical leadership role ideal for someone who thrives on solving complex engineering challenges, stays ahead of modern technology trends, and is passionate about software craftsmanship. You will play a pivotal role in shaping our architecture, contributing directly to the codebase, and mentoring engineers across the organization. This role does not involve people management responsibilities, but requires strong collaboration and technical influence. Deutsche Banks Corporate Bank division is a leading provider of cash management, trade finance and securities finance. We complete green-field projects that deliver the best Corporate Bank - Securities Services products in the world. Our team is diverse, international, and driven by shared focus on clean code and valued delivery. At every level, agile minds are rewarded with competitive pay, support, and opportunities to excel. You will work as part of a cross-functional agile delivery team. You will bring an innovative approach to software development, focusing on using the latest technologies and practices, as part of a relentless focus on business value. You will be someone who sees engineering as team activity, with a predisposition to open code, open discussion and creating a supportive, collaborative environment. You will be ready to contribute to all stages of software delivery, from initial analysis right through to production support. What well offer you 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Your key responsibilities System Design & Development Architect, develop, and maintain high-performance, scalable software systems using Java. Code Contribution Actively contribute to the codebase, ensuring high standards of quality, performance, and reliability. Database Engineering Design and optimize data-intensive applications using MongoDB, including indexing and query optimization. Microservices & Cloud Implement microservices architecture following established guidelines, deployed on Google Kubernetes Engine (GKE) . Security & Compliance Ensure systems comply with security regulations and internal policies. Infrastructure Oversight Review and update policies related to internal systems and equipment. Mentorship Guide and mentor engineers, setting a high bar for technical excellence and best practices. Cross-functional Collaboration Work closely with product managers, architects, and other stakeholders to translate business requirements into scalable technical solutions, including HLD and LLD documentation. Process Improvement Drive best practices in software development, deployment, and operations. Your skills and experience Deep expertise in software architecture, cloud infrastructure, and modern development practices. Strong coding skills and a passion for hands-on development. Excellent communication and leadership abilities. 10+ years of professional software development experience, with deep expertise in Java . Strong experience with MongoDB and building data-intensive applications. Proficiency in Kubernetes and deploying systems at scale in cloud environments , preferably Google Cloud Platform (GCP) . Hands-on experience with CI/CD pipelines , monitoring, logging, and alerting tools (e.g., Prometheus, Grafana, ELK ). Solid understanding of reactive or event-driven architectures . Familiarity with Infrastructure as Code (IaC) tools such as Terraform . Experience with modern software engineering practices , including TDD, CI/CD, and Agile methodologies. Front-end knowledge is a plus. How well support you About us and our teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.

Posted 1 week ago

Apply

9.0 - 14.0 years

30 - 35 Lacs

Pune

Work from Office

: Job TitleProduction Specialist, AVP LocationPune, India Role Description Our organization within Deutsche Bank is AFC Production Services. We are responsible for providing technical L2 application support for business applications. The AFC (Anti-Financial Crime) line of business has a current portfolio of 25+ applications. The organization is in process of transforming itself using Google Cloud and many new technology offerings. As an Assistant Vice President, your role will include hands-on production support and be actively involved in technical issues resolution across multiple applications. You will also be working as application lead and will be responsible for technical & operational processes for all application you support. Deutsche Banks Corporate Bank division is a leading provider of cash management, trade finance and securities finance. We complete green-field projects that deliver the best Corporate Bank - Securities Services products in the world. Our team is diverse, international, and driven by shared focus on clean code and valued delivery. At every level, agile minds are rewarded with competitive pay, support, and opportunities to excel. You will work as part of a cross-functional agile delivery team. You will bring an innovative approach to software development, focusing on using the latest technologies and practices, as part of a relentless focus on business value. You will be someone who sees engineering as team activity, with a predisposition to open code, open discussion and creating a supportive, collaborative environment. You will be ready to contribute to all stages of software delivery, from initial analysis right through to production support." What well offer you , 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Your key responsibilities Provide technical support by handling and consulting on BAU, Incidents/emails/alerts for the respective applications. Perform post-mortem, root cause analysis using ITIL standards of Incident Management, Service Request fulfillment, Change Management, Knowledge Management, and Problem Management. Manage regional L2 team and vendor teams supporting the application. Ensure the team is up to speed and picks up the support duties. Build up technical subject matter expertise on the applications being supported including business flows, application architecture, and hardware configuration. Define and track KPIs, SLAs and operational metrics to measure and improve application stability and performance. Conduct real time monitoring to ensure application SLAs are achieved and maximum application availability (up time) using an array of monitoring tools. Build and maintain effective and productive relationships with the stakeholders in business, development, infrastructure, and third-party systems / data providers & vendors. Assist in the process to approve application code releases as well as tasks assigned to support to perform. Keep key stakeholders informed using communication templates. Approach support with a proactive attitude, desire to seek root cause, in-depth analysis, and strive to reduce inefficiencies and manual efforts. Mentor and guide junior team members, fostering technical upskill and knowledge sharing. Provide strategic input into disaster recovery planning, failover strategies and business continuity procedures Collaborate and deliver on initiatives and install these initiatives to drive stability in the environment. Perform reviews of all open production items with the development team and push for updates and resolutions to outstanding tasks and reoccurring issues. Drive service resilience by implementing SRE(site reliability engineering) principles, ensuring proactive monitoring, automation and operational efficiency. Ensure regulatory and compliance adherence, managing audits,access reviews, and security controls in line with organizational policies. The candidate will have to work in shifts as part of a Rota covering APAC and EMEA hours between 07:00 IST and 09:00 PM IST (2 shifts). In the event of major outages or issues we may ask for flexibility to help provide appropriate cover. Weekend on-call coverage needs to be provided on rotational/need basis. Your skills and experience 9-15 years of experience in providing hands on IT application support. Experience in managing vendor teams providing 24x7 support. Preferred Team lead role experience, Experience in an investment bank, financial institution. Bachelors degree from an accredited college or university with a concentration in Computer Science or IT-related discipline (or equivalent work experience/diploma/certification). Preferred ITIL v3 foundation certification or higher. Knowledgeable in cloud products like Google Cloud Platform (GCP) and hybrid applications. Strong understanding of ITIL /SRE/ DEVOPS best practices for supporting a production environment. Understanding of KPIs, SLO, SLA and SLI Monitoring ToolsKnowledge of Elastic Search, Control M, Grafana, Geneos, OpenShift, Prometheus, Google Cloud Monitoring, Airflow,Splunk. Working Knowledge of creation of Dashboards and reports for senior management Red Hat Enterprise Linux (RHEL) professional skill in searching logs, process commands, start/stop processes, use of OS commands to aid in tasks needed to resolve or investigate issues. Shell scripting knowledge a plus. Understanding of database concepts and exposure in working with Oracle, MS SQL, Big Query etc. databases. Ability to work across countries, regions, and time zones with a broad range of cultures and technical capability. Skills That Will Help You Excel Strong written and oral communication skills, including the ability to communicate technical information to a non-technical audience and good analytical and problem-solving skills. Proven experience in leading L2 support teams, including managing vendor teams and offshore resources. Able to train, coach, and mentor and know where each technique is best applied. Experience with GCP or another public cloud provider to build applications. Experience in an investment bank, financial institution or large corporation using enterprise hardware and software. Knowledge of Actimize, Mantas, and case management software is good to have. Working knowledge of Big Data Hadoop/Secure Data Lake is a plus. Prior experience in automation projects is great to have. Exposure to python, shell, Ansible or other scripting language for automation and process improvement Strong stakeholder management skills ensuring seamless coordination between business, development, and infrastructure teams. Ability to manage high-pressure issues, coordinating across teams to drive swift resolution. Strong negotiation skills with interface teams to drive process improvements and efficiency gains. How well support you . . . About us and our teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.

Posted 1 week ago

Apply

3.0 - 5.0 years

6 - 10 Lacs

Coimbatore

Work from Office

Educational Bachelor of Engineering Service Line Equinox Responsibilities A day in the life of an Infosys Equinox employee As part of the Infosys Equinox delivery team, your primary role would be to ensure effective Design, Development, Validation and Support activities, to assure that our clients are satisfied with the high levels of service in the technology domain. You will gather the requirements and specifications to understand the client requirements in a detailed manner and translate the same into system requirements. A Clear understanding of HTTP / Network protocol concepts, designs & operations - TCP dump, Cookies, Sessions, Headers, Client Server Architecture. Core strength in Linux and Azure infrastructure provisioning including VNet, Subnet, Gateway, VM, Security groups, MySQL, Blob Storage, Azure Cache, AKS Cluster etc. Expertise with automating Infrastructure as a code using Terraform, Packer, Ansible, Shell Scripting and Azure DevOps. Expertise with patch management, APM tools like AppDynamics, Instana for monitoring and alerting. Knowledge in technologies including Apache Solr, MySQL, Mongo, Zookeeper, RabbitMQ, Pentaho etc. Knowledge with Cloud platform including AWS and GCP are added advantage. Ability to identify and automate recurring tasks for better productivity. Ability to understand, implement industry standard security solutions. Experience in implementing Auto scaling, DR, HA, Multi-region with best practices is added advantage. Ability to work under pressure, managing expectations from various key stakeholders. You will gather the requirements and specifications to understand the client requirements in a detailed manner and translate the same into system requirements. You will play a key role in the overall estimation of work requirements to provide the right information on project estimations to Technology Leads and Project Managers. You would be a key contributor to building efficient prog Additional Responsibilities: Ability to identify bottlenecks, debugging hotspots in optimizing performanceContinuously learning with the latest trends in performance engineering, frameworks and methodologies. Technical and Professional : Ability to grasp cloud platforms [AWS, Azure, GCP] Kubernetes and containerization for scalable deploymentsBasic knowledge with Performance testing tools like JMeter, LoadRunner or any other related tool.Good / Expertise in any of the programming languages like Java, Python, C or C++Ability to analyze system metrics using profiling / monitoring tools like Instana, Dynatrace, Prometheus and Grafana Preferred Skills: Technology-Analytics - Packages-Python - Big Data Technology-Java-Java - ALL Technology-Performance Testing-Performance Engineering-Apache Jmeter Technology-Performance Testing-Performance Testing - ALL Technology-Infra_ToolAdministration-Others-Loadrunner

Posted 1 week ago

Apply

3.0 - 5.0 years

0 - 0 Lacs

Bengaluru

Work from Office

What You Will Do DevOps and Platform Engineering Build and maintain CI/CD pipelines using tools like GitHub Actions, GitLab CI, or Jenkins Manage infrastructure as code using Terraform, Ansible, or similar tools Deploy, monitor, and scale services using Docker and Kubernetes Operate cloud-native environments on AWS, Azure, or GCP with a focus on cost, performance, and security Automate routine tasks and build reliable deployment workflows Set up observability tools such as Prometheus, Grafana, or ELK stack Contribute to incident handling, recovery, and root cause analysis Development and Engineering Mindset Strong hands-on background in full-stack or backend development using Node.js, Python, Java, or Go Familiarity with modern frontend frameworks such as React, Vue, or Angular Solid understanding of APIs, microservices, and database design (SQL and NoSQL) Apply software engineering best practices to infrastructure and operations Proficiency in Bash and shell scripting is a must What We Expect Bachelors degree in Computer Science, Engineering, or related field At least 2 years of experience in full-stack or backend development 1 to 3 years of hands-on DevOps experience including container orchestration and CI/CD Strong Git skills and experience with Git-based workflows Exposure to Kubernetes (EKS, GKE, or AKS) in production environments Experience in automating deployments and managing secure, cloud-native infrastructure Bonus Points Certifications such as AWS DevOps Engineer, Azure DevOps Expert, or Kubernetes Administrator Experience with serverless platforms or edge computing Exposure to DevSecOps and secure-by-default infrastructure design What We Look For Engineering depth with a problem-solving mindset Strong scripting and automation habits Clear communicator who collaborates well across teams Ownership-driven, with a focus on reliability and performance This role is ideal for engineers who started as developers and grew into infrastructure and automation. If you think in code, automate by instinct, and care about building reliable systems, you will thrive here.

Posted 1 week ago

Apply

6.0 - 10.0 years

40 - 45 Lacs

Gurugram

Work from Office

About the Role Were looking for a skilled Site Reliability Engineer (SRE) with a strong foundation in Java or Python development, infrastructure automation, and application monitoring. You’ll be embedded within engineering teams to drive reliability, scalability, and performance across our systems. If you have a product-first mindset, enjoy solving real-world problems at scale, and love diving into code and systems alike — we’d love to talk to you. What You’ll Work On Enhancing service reliability & availability by implementing robust SLI/SLO-based monitoring and alerting systems Collaborating with developers to optimize service performance and reliability in Java/Spring Boot applications Building infrastructure as code with Terraform and automating provisioning pipelines Conducting chaos testing, capacity planning, and failure analysis Working with cloud-native observability stacks (e.g., CloudWatch, Prometheus, Victoria Metrics) Reporting with Snowflake and Sigma for operational insights Supporting scalable and resilient database operations across RDS and NoSQL systems What We’re Looking For 6–10 years of experience Strong backend coding skills – Java (preferred) or Python (not just scripting) Experience with monitoring tools: CloudWatch, Prometheus, Victoria Metrics Familiarity with Snowflake and Sigma reporting (preferred) Terraform experience for IaC Strong database skills: RDS and any major NoSQL platform Deep understanding of SLI/SLOs, alerting, capacity planning, chaos testing Application/service-oriented mindset, aligned with an embedded SRE approach

Posted 2 weeks ago

Apply

3.0 - 7.0 years

15 - 25 Lacs

Bengaluru

Hybrid

Job Summary: We are seeking a highly motivated and experienced DevOps Engineer to join our team and champion a Devops culture. This role will be focused on designing, building, and maintaining our CI/CD pipelines, with a specific focus on secure Java application builds, branching strategies, and automated deployments. The ideal candidate will be proficient in scripting (Python, Bash, and Groovy), have a strong understanding of Jenkins pipeline development, including the use of shared libraries, infrastructure as code, modern DevOps practices, and a security-first mindset. Additionally, a solid understanding of QA automation testing tools is beneficial, and proficiency with Selenium is considered an added advantage. You will play a critical role in enabling our development and QA teams to deliver high-quality, secure software rapidly and reliably. Key Responsibilities: Secure CI/CD Pipeline Development: Design, implement, and maintain robust, automated, and secure CI/CD pipelines using Jenkins and other relevant tools. This includes integrating security scanning and testing tools into the pipeline. Secure Java Build Expertise: Optimize and troubleshoot Java application builds, ensuring security best practices (including dependency management, vulnerability scanning), artifact repository management (Nexus, Artifactory) and build performance. Secure Branching Strategy Management: Implement and maintain effective and secure branching strategies (e.g., Gitflow, Trunk-Based Development), ensuring secure collaboration between development teams and preventing unauthorized code merges. Infrastructure as Code (IaC) with Security: Leverage IaC tools (e.g., Terraform, CloudFormation, Ansible) to automate secure infrastructure provisioning, configuration, and management for our CI/CD environment. This includes implementing security policies and access controls in code. Secure Jenkins Pipeline Development with Groovy: Develop complex, efficient, maintainable, and secure Jenkins pipelines using Groovy scripting, that integrate testing, security code quality checks, and automated deployments. This includes the use of shared libraries for reusable pipeline logic. Shared Library Development: Develop and maintain reusable Jenkins shared libraries using Groovy to encapsulate common CI/CD logic, promoting consistency and maintainability across pipelines. Scripting and Automation for Security: Utilize scripting (Python, Bash, and Groovy) to automate repetitive tasks, enhance CI/CD workflows, integrate security tools, and create custom tools to improve development and security productivity. Monitoring and Logging with Security Focus: Implement secure monitoring and logging solutions to proactively identify and address potential security vulnerabilities, performance issues, and malicious behavior in the CI/CD pipelines. Security Collaboration: Collaborate closely with security teams, development teams, QA, and other stakeholders to understand their needs, deliver secure solutions, and participate in security reviews. Continuous Security Improvement: Continuously identify areas for improvement within our CI/CD processes and infrastructure from a security perspective, and implement security best practices to improve efficiency and reliability. Documentation with Security Focus: Create and maintain clear, comprehensive, and security focused documentation for all CI/CD processes, infrastructure, and security controls. Vulnerability and Security Assessment: Perform regular vulnerability assessments of the CI/CD infrastructure and application codebase to identify and fix potential weaknesses. Threat Modeling: Conduct threat modeling exercises to identify potential security risks in the CI/CD process. Troubleshooting with Security Awareness: Troubleshoot issues with the CI/CD pipeline, build processes, deployments, and other areas related to the build system, always with security as a key concern. Integrate QA automation tools and frameworks within the CI/CD pipeline, enabling automated testing of new builds. Stay up to date with Security: Stay abreast of industry trends and emerging DevOps and security technologies and practices. Required Qualifications: Bachelor's degree in Computer Science, Engineering, or a related field, or equivalent practical experience. 3 to 7 years of experience as a DevOps Engineer, specializing in CI/CD and automation. Extensive experience developing and maintaining complex Jenkins pipelines with built-in security checks, including declarative and scripted pipelines using Groovy. Proven experience developing and using Jenkins shared libraries using Groovy. Proven experience with infrastructure as code (Terraform, CloudFormation, Ansible) with a security focus. Excellent scripting skills in Python, Bash, and Groovy, with the ability to automate security tasks and integrate security tools into the workflows. Strong understanding of containerization technologies (Docker) and container orchestration (Kubernetes), including secure configurations. Experience with cloud platforms (AWS, Azure, GCP), including their security features. Excellent problem-solving, analytical, and troubleshooting skills, with a focus on security. Strong communication and collaboration skills. Experience working with security teams. Solid understanding of QA automation testing tools and frameworks.

Posted 2 weeks ago

Apply

7.0 - 12.0 years

9 - 14 Lacs

Bengaluru

Work from Office

FICO (NYSEFICO) is a leading global analytics software company, helping businesses in 100+ countries make better decisions. Join our world-class team today and fulfill your career potential! The Opportunity Join our dynamic and forward-thinking Platform Engineering team at a world-class analytics company. Our mission is to accelerate innovation by delivering a cohesive internal developer platform that combines an enterprise-grade Spotify Backstage portal, Buf Schema Registry, GitOps automation, and cloud-native tooling. As a Lead Platform Engineer, youll architect and own the services, plugins, and pipelines that power a world-class developer experience for thousands of engineers building fraud, risk, marketing, and customer-management solutions. Sr. Director, 1ES Engineering What Youll Contribute Operate and scale Backstage as the single pane of glass for developers. Design and publish custom Backstage plugins, templates, and software catalog integrations that reduce cognitive load and surface business context. Define governance & RBAC models for Backstage groups, entities, and APIs. Establish and maintain BSR as the system of record for Protobuf and gRPC APIs. Automate linting, breaking-change detection, versioning, and dependency insights in CI/CD. Integrate BSR metadata into Backstage to provide full API lineage and documentation. Collaborate with product and infrastructure teams to deliver resilient, self-service platform building blocks. Own GitHub Actions, Argo CD, Crossplane, and policy-as-code workflows that enable secure, audit-ready deployments. Continuously experiment with new ideashack days, proofs-of-concept, and brown-bag sessionsto push the envelope of DevEx. Champion data-driven improvements using DORA/SPACE metrics and developer feedback loops. Instrument, monitor, and tune platform components for scale (Prometheus/Grafana, Splunk, Cribl, CloudWatch). Embed security controls (SCA, SAST, OPA/Kyverno) early in the SDLC. Guide engineers across domains, codify best practices, and foster a culture of psychological safety, creativity, and ownership. What Were Seeking Deep Backstage ExpertiseProven experience deploying, customizing, and scaling Backstage in production, including authoring plugins (React/Node), scaffolder templates, and catalog processors. Buf Schema Registry MasteryHands-on knowledge of managing API contracts in BSR, enforcing semantic versioning, and integrating breaking-change gates into CI/CD. Cloud-Native & GitOps ProficiencyKubernetes (EKS/GKE/AKS), Argo CD, Crossplane, Docker, Helm; expert-level GitHub Actions workflow design. Programming Skills: Strong in TypeScript/JavaScript (for Backstage), plus one or more of Go, Python, or NodeJS for platform services. Infrastructure as Code & AutomationTerraform, Pulumi, or Ansible to codify cloud resources and policies. Observability & Incident ManagementPrometheus, Grafana, Datadog, PagerDuty; ability to design SLOs/SLA dashboards. Creative Problem-Solving & Growth MindsetDemonstrated ability to think big, prototype quickly, and iterate based on data and feedback. Excellent Communication & CollaborationClear written and verbal skills; ability to translate technical details to diverse stakeholders. Education / ExperienceBachelors in Computer Science or equivalent experience; 7 + years in platform, DevOps, or developer-experience roles, with 2 + years focused on Backstage and/or BSR. Our Offer to You An inclusive culture strongly reflectingourcore valuesAct Like an Owner, DelightOurCustomers and Earn the Respect of Others. The opportunitytomake an impact and develop professionally by leveraging your unique strengths and participating in valuable learning experiences. Highly competitive compensation, benefits and rewards programs that encourageyoutobring yourbest every day and be recognized for doing so. An engaging, people-first work environmentoffering work/life balance, employee resource groups, and social eventstopromote interaction and camaraderie. Why Make a Move to FICO At FICO, you can develop your career with a leading organization in one of the fastest-growing fields in technology today Big Data analytics. Youll play a part in our commitment to help businesses use data to improve every choice they make, using advances in artificial intelligence, machine learning, optimization, and much more. FICO makes a real difference in the way businesses operate worldwide Credit Scoring FICO Scores are used by 90 of the top 100 US lenders. Fraud Detection and Security 4 billion payment cards globally are protected by FICO fraud systems. Lending 3/4 of US mortgages are approved using the FICO Score. Global trends toward digital transformation have created tremendous demand for FICOs solutions, placing us among the worlds top 100 software companies by revenue. We help many of the worlds largest banks, insurers, retailers, telecommunications providers and other firms reach a new level of success. Our success is dependent on really talented people just like you who thrive on the collaboration and innovation thats nurtured by a diverse and inclusive environment. Well provide the support you need, while ensuring you have the freedom to develop your skills and grow your career. Join FICO and help change the way business thinks! Learn more about how you can fulfil your potential at www.fico.com/Careers FICO promotes a culture of inclusion and seeks to attract a diverse set of candidates for each job opportunity. We are an equal employment opportunity employer and were proud to offer employment and advancement opportunities to all candidates without regard to race, color, ancestry, religion, sex, national origin, pregnancy, sexual orientation, age, citizenship, marital status, disability, gender identity or Veteran status. Research has shown that women and candidates from underrepresented communities may not apply for an opportunity if they dont meet all stated qualifications. While our qualifications are clearly related to role success, each candidates profile is unique and strengths in certain skill and/or experience areas can be equally effective. If you believe you have many, but not necessarily all, of the stated qualifications we encourage you to apply. Information submitted with your application is subject to theFICO Privacy policy at https://www.fico.com/en/privacy-policy

Posted 2 weeks ago

Apply

6.0 - 8.0 years

8 - 10 Lacs

Hyderabad, Pune, Bengaluru

Work from Office

Choosing Capgemini means choosing a company where you will be empowered to shape your career in the way youd like, where youll be supported and inspired by a collaborative community of colleagues around the world, and where you ll be able to reimagine what s possible. Join us and help the world s leading organizations unlock the value of technology and build a more sustainable, more inclusive world. Your Role Build and manage CI/CD pipelines (Jenkins, GitLab CI, Azure DevOps). Automate infrastructure with Terraform, Ansible, or CloudFormation. Set up monitoring with Prometheus, Grafana, and ELK. Manage containers with Docker and orchestrate via Kubernetes. Collaborate with developers to embed DevOps into the SDLC. Your Profile 612 years in DevOps, CI/CD, and IaC. Expertise in Docker, Kubernetes, and cloud (AWS/Azure/GCP). Experience with monitoring tools (Prometheus, Grafana, ELK). Knowledge of security, compliance, and performance. On-call readiness and production issue handling. What you'll love about working with us Flexible work option: Hybrid Competitive salary and benefits package Career growth with SAP and cloud certifications Inclusive and collaborative work environment Capgemini is a global business and technology transformation partner, helping organizations to accelerate their dual transition to a digital and sustainable world, while creating tangible impact for enterprises and society. It is a responsible and diverse group of 340,000 team members in more than 50 countries. With its strong over 55-year heritage, Capgemini is trusted by its clients to unlock the value of technology to address the entire breadth of their business needs. It delivers end-to-end services and solutions leveraging strengths from strategy and design to engineering, all fueled by its market leading capabilities in AI, generative AI, cloud and data, combined with its deep industry expertise and partner ecosystem.

Posted 2 weeks ago

Apply

3.0 - 5.0 years

3 - 6 Lacs

Mumbai, Navi Mumbai, Mumbai (All Areas)

Work from Office

Job Title : ELK Stack Developer (Elasticsearch, Logstash, Kibana) Experience : 3+ Years Location : Mumbai / Navi Mumbai (Work from Office) Job Type : Full-time Job Summary : We are seeking a skilled ELK Stack Developer with at least 3 years of hands-on experience in implementing and managing Elasticsearch, Logstash, and Kibana. The ideal candidate will be responsible for developing scalable logging and monitoring solutions, integrating log sources, designing visual dashboards, and optimizing search and analytics capabilities. Key Responsibilities : Develop and maintain scalable ELK (Elasticsearch, Logstash, Kibana) solutions. Design and implement centralized logging and monitoring solutions. Integrate various data sources into Elasticsearch using Logstash or Beats. Create Kibana dashboards and visualizations to monitor systems, applications, and business KPIs. Write efficient Elasticsearch queries and tune clusters for performance and reliability. Handle data parsing, indexing, and transformation using Logstash filters and pipelines. Implement alerting and reporting solutions using Kibana or third-party integrations. Collaborate with DevOps, security, and development teams to define logging requirements. Troubleshoot issues across the ELK stack and ensure high availability. Required Skills : Strong experience in Elasticsearch , Logstash , and Kibana (minimum 3 years). Experience with Beats (Filebeat, Metricbeat) for data ingestion. Strong understanding of Elasticsearch indexing, mapping, query DSL, and performance tuning. Proficiency in creating custom Kibana dashboards and visualizations . Hands-on experience with Logstash pipelines , grok filters, and data enrichment. Understanding of REST APIs and JSON-based data formats. Familiarity with Linux , Bash , and scripting . Knowledge of monitoring tools (e.g., Prometheus, Grafana) is a plus. Experience with cloud platforms (AWS, Azure, GCP) is desirable. Please Share & Refer resume at ajay.kurlekar@cloverinfotech.com

Posted 2 weeks ago

Apply

6.0 - 11.0 years

12 - 22 Lacs

Hyderabad

Work from Office

ideyaLabs is seeking a highly skilled and motivated DevOps Lead to drive our CI/CD pipelines, infrastructure automation, and cloud operations. The ideal candidate will lead a team of DevOps engineers and collaborate closely with development, QA, and IT teams to enhance our deployment processes, scalability, and system reliability. Key Responsibilities- Lead and mentor a team of DevOps engineers to deliver scalable and secure infrastructure. Design, build, and maintain CI/CD pipelines using tools like Jenkins, GitLab CI, GitHub Actions, etc. Architect and automate infrastructure provisioning using Infrastructure as Code (IaC) tools such as Terraform, CloudFormation, or Pulumi. Manage and optimize cloud infrastructure (AWS/Azure/GCP) to ensure high availability and performance. Implement monitoring, logging, and alerting systems using tools like Prometheus, Grafana, ELK, Datadog, or New Relic. Champion DevSecOps practices and integrate security at every stage of the DevOps lifecycle. Collaborate with software development teams to ensure smooth deployments and fast recovery from incidents. Define and enforce SRE/DevOps best practices, SLAs, and disaster recovery procedures Continuously evaluate emerging tools and technologies to improve system efficiency and team productivity. Required Skills & Qualifications Bachelor's or Master's in Computer Science, Engineering, or a related field. Proven experience leading DevOps/SRE/Platform teams. Strong expertise in cloud platforms (AWS, Azure, or GCP). Proficient with containerization and orchestration (Docker, Kubernetes, Helm). Hands-on experience with IaC and configuration management tools (Terraform, Ansible, Chef, Puppet). Excellent scripting skills (Python, Bash, or Go preferred). Familiarity with GitOps and modern release strategies (Blue/Green, Canary deployments) Deep understanding of network, system, and security principles. Having an understanding of PCI DSS and SOC 2 procedures. Strong communication and stakeholder management skills. Preferred Qualifications Certifications like AWS Certified DevOps Engineer, CKAD/CKA, and Azure DevOps Expert. Experience with service mesh technologies (Istio, Linkerd).- Exposure to FinOps or cost optimization practices in the cloud

Posted 2 weeks ago

Apply

2.0 - 4.0 years

4 - 6 Lacs

Bengaluru

Work from Office

Job Title: Jr. DevOps Engineer Location: Bangalore (Work from Office 5 days a week) Experience: 2+ Years Joining: Immediate Joiners Preferred Key Responsibilities: Design, implement, and maintain CI/CD pipelines using tools like Jenkins, GitHub Actions, or GitLab CI. Manage cloud infrastructure (preferably AWS) with a focus on scalability, reliability, and security. Deploy and manage containerized applications using Docker and Kubernetes. Automate infrastructure provisioning using tools like Terraform or Ansible. Monitor system performance and troubleshoot issues using tools like Prometheus, Grafana, ELK, etc. Collaborate with development, QA, and operations teams to ensure seamless deployments. Technical Skills Required: CI/CD: Jenkins, Git, GitHub/GitLab Cloud: AWS (EC2, S3, IAM, CloudWatch) Containers: Docker, Kubernetes IaC: Terraform / Ansible Scripting: Bash / Python Monitoring: Prometheus, Grafana, ELK Stack Eligibility: Minimum 2 years of DevOps experience. Strong troubleshooting and communication skills. Willing to work full-time from our Bangalore office.

Posted 2 weeks ago

Apply

4.0 - 6.0 years

6 - 8 Lacs

Bengaluru

Work from Office

We are seeking a skilled DevOps Engineer with strong experience in Google Cloud Platform (GCP) to support AI/ML project infrastructure. The ideal candidate will work closely with data scientists, ML engineers, and developers to build and manage scalable, secure, and automated pipelines for AI/ML model training, testing, and deployment. Responsibilities: Design and manage cloud infrastructure to support AI/ML workloads on GCP. Develop and maintain CI/CD pipelines for ML models and applications. Automate model training, validation, deployment, and monitoring processes using tools like Kubeflow, Vertex AI, Cloud Composer, Airflow, etc. Set up and manage infrastructure as code (IaC) using tools such as Terraform or Deployment Manager. Implement robust security, monitoring, logging, and alerting systems using Cloud Monitoring, Cloud Logging, Prometheus, Grafana, etc. Collaborate with ML engineers and data scientists to optimize compute environments (e.g., GPU/TPU instances, notebooks). Manage and maintain containerized environments using Docker and Kubernetes (GKE). Ensure cost-efficient cloud resource utilization and governance. Required Skills Bachelor's degree in engineering or relevant field Must have 4 years of proven experience as DevOps Engineer with at least 1 year on GCP Strong experience with DevOps tools and methodologies in production environments Proficiency in scripting with Python, Bash, or Shell Experience with Terraform, Ansible, or other IaC tools. Deep understanding of Docker, Kubernetes, and container orchestration Knowledge of CI/CD pipelines, automated testing, and model deployment best practices. Familiarity with ML lifecycle tools such as MLflow, Kubeflow Pipelines, or TensorFlow Extended (TFX). Experience in designing conversational flows for AI Agents/chatbot

Posted 2 weeks ago

Apply

7.0 - 12.0 years

0 - 1 Lacs

Dhule

Work from Office

Key Responsibilities AI Model Deployment & Integration: Deploy and manage AI/ML models, including traditional machine learning and GenAI solutions (e.g., LLMs, RAG systems). Implement automated CI/CD pipelines for seamless deployment and scaling of AI models. Ensure efficient model integration into existing enterprise applications and workflows in collaboration with AI Engineers. Optimize AI infrastructure for performance and cost efficiency in cloud environments (AWS, Azure, GCP). Monitoring & Performance Management: Develop and implement monitoring solutions to track model performance, latency, drift, and cost metrics. Set up alerts and automated workflows to manage performance degradation and retraining triggers. Ensure responsible AI by monitoring for issues such as bias, hallucinations, and security vulnerabilities in GenAI outputs. Collaborate with Data Scientists to establish feedback loops for continuous model improvement. Automation & MLOps Best Practices: Establish scalable MLOps practices to support the continuous deployment and maintenance of AI models. Automate model retraining, versioning, and rollback strategies to ensure reliability and compliance. Utilize infrastructure-as-code (Terraform, CloudFormation) to manage AI pipelines. Security & Compliance: Implement security measures to prevent prompt injections, data leakage, and unauthorized model access. Work closely with compliance teams to ensure AI solutions adhere to privacy and regulatory standards (HIPAA, GDPR). Regularly audit AI pipelines for ethical AI practices and data governance. Collaboration & Process Improvement: Work closely with AI Engineers, Product Managers, and IT teams to align AI operational processes with business needs. Contribute to the development of AI Ops documentation, playbooks, and best practices. Continuously evaluate emerging GenAI operational tools and processes to drive innovation. Qualifications & Skills Education: Bachelors or Masters degree in Computer Science, Data Engineering, AI, or a related field. Relevant certifications in cloud platforms (AWS, Azure, GCP) or MLOps frameworks are a plus. Experience: 3+ years of experience in AI/ML operations, MLOps, or DevOps for AI-driven solutions. Hands-on experience deploying and managing AI models, including LLMs and GenAI solutions, in production environments. Experience working with cloud AI platforms such as Azure AI, AWS SageMaker, or Google Vertex AI. Technical Skills: Proficiency in MLOps tools and frameworks such as MLflow, Kubeflow, or Airflow. Hands-on experience with monitoring tools (Prometheus, Grafana, ELK Stack) for AI performance tracking. Experience with containerization and orchestration tools (Docker, Kubernetes) to support AI workloads. Familiarity with automation scripting using Python, Bash, or PowerShell. Understanding of GenAI-specific operational challenges such as response monitoring, token management, and prompt optimization. Knowledge of CI/CD pipelines (Jenkins, GitHub Actions) for AI model deployment. Strong understanding of AI security principles, including data privacy and governance considerations.

Posted 2 weeks ago

Apply

5.0 - 8.0 years

12 - 18 Lacs

Mumbai, Hyderabad, Chennai

Work from Office

We are seeking an experienced AWS Platform Engineer Developer to architect and manage secure, scalable AWS environments in compliance with industry regulations such as GDPR, FCA, and PRA. The role involves deploying and maintaining EKS clusters, Istio service mesh, and Kong API Gateway, implementing robust security measures using Dynatrace, Fortigate, and AWS-native security services (Security Hub, GuardDuty, WAF), and automating infrastructure provisioning with Terraform and CloudFormation. Responsibilities also include enforcing Privileged Access Management (PAM) policies, integrating observability tools (Dynatrace, Grafana, Prometheus), and collaborating with teams on container orchestration using Kubernetes and Docker. Experience in serverless technologies like AWS Lambda and API Gateway, as well as container security scanning tools such as Trivy and Aqua Security, is preferred.

Posted 2 weeks ago

Apply

5.0 - 7.0 years

7 - 9 Lacs

Bengaluru

Work from Office

A skilled DevOps Engineer to manage and optimize both on-premises and AWS cloud infrastructure. The ideal candidate will have expertise in DevOps tools, automation, system administration, and CI/CD pipeline management while ensuring security, scalability, and reliability. Key Responsibilities: 1. AWS & On-Premises Solution Architecture: o Design, deploy, and manage scalable, fault-tolerant infrastructure across both on-premises and AWS cloud environments. o Work with AWS services like EC2, IAM, VPC, CloudWatch, GuardDuty, AWS Security Hub, Amazon Inspector, AWS WAF, and Amazon RDS with Multi-AZ. o Configure ASG and implement load balancing techniques such as ALB and NLB. o Optimize cost and performance leveraging Elastic Load Balancing and EFS. o Implement logging and monitoring with CloudWatch, CloudTrail, and on-premises monitoring solutions. 2. DevOps Automation & CI/CD: o Develop and maintain CI/CD pipelines using Jenkins and GitLab for seamless code deployment across cloud and on-premises environments. o Automate infrastructure provisioning using Ansible, and CloudFormation. o Implement CI/CD pipeline setups using GitLab, Maven, Gradle, and deploy on Nginx and Tomcat. o Ensure code quality and coverage using SonarQube. o Monitor and troubleshoot pipelines and infrastructure using Prometheus, Grafana, Nagios, and New Relic. 3. System Administration & Infrastructure Management: o Manage and maintain Linux and Windows systems across cloud and on-premises environments, ensuring timely updates and security patches. o Configure and maintain web/application servers like Apache Tomcat and web servers like Nginx and Node.js. o Implement robust security measures, SSL/TLS configurations, and secure communications. o Configure DNS and SSL certificates. o Maintain and optimize on-premises storage, networking, and compute resources. 4. Collaboration & Documentation: o Collaborate with development, security, and operations teams to optimize deployment and infrastructure processes. o Provide best practices and recommendations for hybrid cloud and on-premises architecture, DevOps, and security. o Document infrastructure designs, security configurations, and disaster recovery plans for both environments. Required Skills & Qualifications: Cloud & On-Premises Expertise: Extensive knowledge of AWS services (EC2, IAM, VPC, RDS, etc.) and experience managing on-premises infrastructure. DevOps Tools: Proficiency in SCM tools (Git, GitLab), CI/CD (Jenkins, GitLab CI/CD), and containerization. Code Quality & Monitoring: Experience with SonarQube, Prometheus, Grafana, Nagios, and New Relic. Operating Systems: Experience managing Linux/Windows servers and working with CentOS, Fedora, Debian, and Windows platforms. Application & Web Servers: Hands-on experience with Apache Tomcat, Nginx, and Node.js. Security & Networking: Expertise in DNS configuration, SSL/TLS implementation, and AWS security services. Soft Skills: Strong problem-solving abilities, effective communication, and proactive learning. Preferred Qualifications: AWS certifications (Solutions Architect, DevOps Engineer) and a bachelors degree in Computer Science or related field. Experience with hybrid cloud environments and on-premises infrastructure automation.

Posted 2 weeks ago

Apply

3.0 - 8.0 years

5 - 10 Lacs

Bengaluru

Work from Office

• Primary Skills: Prometheus, Grafana, Datadog ,Alerting Techniques, Alert Triage and Incident Management, Application Issues RCA/Debugging, SQL. • Proven L3 level experience in managing large-scale, distributed systems in production environments. Required Candidate profile Drive SRE transformations by building frameworks and migrating traditional IT support to modern SRE practices. Collaborate closely with development and operations teams to improve system observability

Posted 2 weeks ago

Apply

1.0 - 3.0 years

3 - 5 Lacs

Chennai

Work from Office

Design and develop backend components and RESTful APIs using Java (11+) and Spring Boot Build and maintain scalable microservices with strong emphasis on clean architecture Write reliable and efficient SQL queries; work with relational and optionally NoSQL (MongoDB) databases Apply DSA fundamentals in solving problems, optimizing code, and building performant features Follow and advocate for SOLID principles, clean code, and test-driven development Collaborate across product, design, and QA to build meaningful, high-quality features Contribute to internal tools or AI-powered enhancements to accelerate workflows Participate in code reviews, peer discussions, and technical design sessions What Were Looking For : 1-2 years of backend development experience using Java and Spring Boot Solid understanding and application of Data Structures and Algorithms in real-world scenarios Strong foundation in Object-Oriented Programming and adherence to SOLID principles Hands-on experience with SQL databases and understanding of performance tuning Familiarity with MongoDB or other NoSQL databases (good to have) Curiosity or exposure to AI/ML, generative APIs, or automation use cases Good communication skills, debugging ability, and a mindset for continuous learning Bonus Points For: Familiarity with cloud environments (AWS) Experience with Git, CI/CD pipelines (e g , GitHub Actions, Jenkins) Exposure to monitoring/logging tools like Prometheus, Grafana, or ELK Past experience in competitive programming, hackathons, or personal projects

Posted 2 weeks ago

Apply

2.0 - 7.0 years

4 - 9 Lacs

Pune, Coimbatore

Work from Office

Job Summary : We are seeking a skilled Erlang Developer to join our backend engineering team. The ideal candidate will have a strong background in Erlang, with working experience in Elixir and RabbitMQ. You will play a key role in designing, building, and maintaining scalable, fault-tolerant systems used in high-availability environments. Key Responsibilities : - Design, develop, test, and maintain scalable Erlang-based backend applications. - Collaborate with cross-functional teams to understand requirements and deliver efficient solutions. - Integrate messaging systems such as RabbitMQ to ensure smooth communication between services. - Write reusable, testable, and efficient code in Erlang and Elixir. - Monitor system performance and troubleshoot issues in production. - Ensure high availability and responsiveness of services. - Participate in code reviews and contribute to best practices in functional programming. Required Skills : - Proficiency in Erlang with hands-on development experience. - Working knowledge of Elixir and the Phoenix framework. - Strong experience with RabbitMQ and messaging systems. - Good understanding of distributed systems and concurrency. - Experience with version control systems like Git. - Familiarity with CI/CD pipelines and containerization (Docker is a plus). Preferred Qualifications : - Experience working in telecom, fintech, or real-time systems. - Knowledge of OTP (Open Telecom Platform) and BEAM VM internals. - Familiarity with monitoring tools like Prometheus, Grafana, etc.

Posted 2 weeks ago

Apply

3.0 - 8.0 years

30 - 35 Lacs

Bengaluru

Work from Office

The IT AI Application Platform team is seeking a Senior Site Reliability Engineer (SRE) to develop, scale, and operate our AI Application Platform based on Red Hat technologies, including OpenShift AI (RHOAI) and Red Hat Enterprise Linux AI (RHEL AI). As an SRE you will contribute to running core AI services at scale by enabling customer self-service, making our monitoring system more sustainable, and eliminating toil through automation. On the IT AI Application Platform team, you will have the opportunity to influence the complex challenges of scale which are unique to Red Hat IT managed AI platform services, while using your skills in coding, operations, and large-scale distributed system design. We develop, deploy, and maintain Red Hats next-generation Ai application deployment environment for custom applications and services across a range of hybrid cloud infrastructures. We are a global team operating on-premise and in the public cloud, using the latest technologies from Red Hat and beyond. Red Hat relies on teamwork and openness for its success. We are a global team and strive to cultivate a transparent environment that makes room for different voices. We learn from our failures in a blameless environment to support the continuous improvement of the team. At Red Hat, your individual contributions have more visibility than most large companies, and visibility means career opportunities and growth. What you will do The day-to-day responsibilities of an SRE involve working with live systems and coding automation. As an SRE you will be expected to Build and manage our large scale infrastructure and platform services, including public cloud, private cloud, and datacenter-based Automate cloud infrastructure through use of technologies (e.g. auto scaling, load balancing, etc.), scripting (bash, python and golang), monitoring and alerting solutions (e.g. Splunk, Splunk IM, Prometheus, Grafana, Catchpoint etc) Design, develop, and become expert in AI capabilities leveraging emerging industry standards Participate in the design and development of software like Kubernetes operators, webhooks, cli-tools.. Implement and maintain intelligent infrastructure and application monitoring designed to enable application engineering teams Ensure the production environment is operating in accordance with established procedures and best practices Provide escalation support for high severity and critical platform-impacting events Provide feedback around bugs and feature improvements to the various Red Hat Product Engineering teams Contribute software tests and participate in peer review to increase the quality of our codebase Help and develop peers capabilities through knowledge sharing, mentoring, and collaboration Participate in a regular on-call schedule, supporting the operation needs of our tenants Practice sustainable incident response and blameless postmortems Work within a small agile team to develop and improve SRE methodologies, support your peers, plan and self-improve What you will bring A bachelor's degree in Computer Science or a related technical field involving software or systems engineering is required. However, hands-on experience that demonstrates your ability and interest in Site Reliability Engineering are valuable to us, and may be considered in lieu of degree requirements. You must have some experience programming in at least one of these languagesPython, Golang, Java, C, C++ or another object-oriented language. You must have experience working with public clouds such as AWS, GCP, or Azure. You must also have the ability to collaboratively troubleshoot and solve problems in a team setting. As an SRE you will be most successful if you have some experience troubleshooting an as-a-service offering (SaaS, PaaS, etc.) and some experience working with complex distributed systems. We like to see a demonstrated ability to debug, optimize code and automate routine tasks. We are Red Hat, so you need a basic understanding of Unix/Linux operating systems. Desired skills 3+ years of experience of using cloud providers and technologies (Google, Azure, Amazon, OpenStack etc) 1+ years of experience administering a kubernetes based production environment 2+ years of experience with enterprise systems monitoring 2+ years of experience with enterprise configuration management software like Ansible by Red Hat, Puppet, or Chef 2+ years of experience programming with at least one object-oriented language; Golang, Java, or Python are preferred 2+ years of experience delivering a hosted service Demonstrated ability to quickly and accurately troubleshoot system issues Solid understanding of standard TCP/IP networking and common protocols like DNS and HTTP Demonstrated comfort with collaboration, open communication and reaching across functional boundaries Passion for understanding users needs and delivering outstanding user experiences Independent problem-solving and self-direction Works well alone and as part of a global team Experience working with Agile development methodologies #LI-SH4 About Red Hat Red Hat is the worlds leading provider of enterprise open source software solutions, using a community-powered approach to deliver high-performing Linux, cloud, container, and Kubernetes technologies. Spread across 40+ countries, our associates work flexibly across work environments, from in-office, to office-flex, to fully remote, depending on the requirements of their role. Red Hatters are encouraged to bring their best ideas, no matter their title or tenure. We're a leader in open source because of our open and inclusive environment. We hire creative, passionate people ready to contribute their ideas, help solve complex problems, and make an impact. Inclusion at Red Hat Red Hats culture is built on the open source principles of transparency, collaboration, and inclusion, where the best ideas can come from anywhere and anyone. When this is realized, it empowers people from different backgrounds, perspectives, and experiences to come together to share ideas, challenge the status quo, and drive innovation. Our aspiration is that everyone experiences this culture with equal opportunity and access, and that all voices are not only heard but also celebrated. We hope you will join our celebration, and we welcome and encourage applicants from all the beautiful dimensions that compose our global village. Equal Opportunity Policy (EEO) Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, medical condition, marital status, or any other basis prohibited by law. Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance completing our online job application, email application-assistance@redhat.com . General inquiries, such as those regarding the status of a job application, will not receive a reply.

Posted 2 weeks ago

Apply

4.0 - 8.0 years

15 - 19 Lacs

Bengaluru

Work from Office

We are seeking a DevOps & Odoo Tech Lead (India) to spearhead the rollout and support of the Services QT tool within Odoo, ensuring robust infrastructure, configuration, and operational excellence. Youll design and implement APIs for seamless integration with Nokias service automation platforms and external systems, architect and manage the DevOps environmentincluding CI/CD pipelines, containerization, infrastructure as code, high-availability Odoo deployments, monitoring, and automationand resolve complex performance and integration issues. As team leader, you will coach and mentor DevOps staff, manage agile release cycles, and drive best practices for operational stability, scalability, and security. You have: Deep understanding of Odoo architecture (frontend, backend, database structure) and proficiency in Linux (Ubuntu, Debian, CentOS) as Odoo primarily runs on Linux-based environments Experience in installing, configuring, and optimizing Odoo (Community and Enterprise editions) and system monitoring using tools like Prometheus, Grafana, or ELK stack Knowledge of Odoo modules, customization, and development (Python, XML, JavaScript) and ability to manage Odoo scaling (multi-instance, multi-database) Expertise in Odoo performance tuning (load balancing, caching, database optimization) Experience with Git, GitHub/GitLab CI/CD for version control and deployment automation Experience in setting up and managing virtual machines (VMs), bare-metal servers, and containers and automation of deployments using Ansible, Terraform, or shell scripting It would be nice if you also had: Expertise in PostgreSQL (Odoos database) Experience in AWS, Google Cloud, Azure, or DigitalOcean for cloud-based Odoo hosting Expertise in network security, firewalls, and VPNs Define, design, and oversee the development of APIs required from Nokia products (and other new-tech vendors) to enable seamless integration with Nokias service automation platforms. Act as the primary technical liaison for both internal and external service software teams, guiding effective integration with Nokias service automation components. Diagnose and resolve complex performance and reliability issues within service operations automation using deep expertise in DevOps, infrastructure, and Odoo tuning. Use in-depth business domain knowledge to align architectural and DevOps strategies with service automation goals and customer objectives. Provide structured mentoring, best practices, and real-time guidance to Managed Services DevOps staff, taskforces, and workteams. Coordinate task allocation, monitor progress, and coach team members, contributing feedback for formal performance evaluations. Lead release management within Scrum/Agile cycles, including planning, execution, regression testing, and post-release reviews to meet customer requirements. Administer and optimize Odoo deployments on Linux or cloud platformshandling installation, configuration, performance tuning, HA, backupswhile implementing CI/CD pipelines, containerization, infrastructure automation, monitoring, and security best practices.

Posted 2 weeks ago

Apply

4.0 - 8.0 years

20 - 25 Lacs

Mumbai

Work from Office

Required Qualification: BE/ B Tech/ MCA Skill, Knowledge &Trainings: Own and manage the CI/CD pipelines for automated build, test, and deployment. Design and implement robust deployment strategies for microservices and web applications. Set up and maintain monitoring, alerting, and logging frameworks (e.g., Prometheus, Grafana, ELK) Build automations which will help optimize software delivery. Improve reliability, quality, and time-to-market of our suite of software solutions. Will be responsible for availability, latency, performance efficiency, change management, monitoring, emergency response and capacity planning. Will create services that will do automatic provisioning of test environments, automation of release management process, setting up pre-emptive monitoring of logs and creating dashboards for metrics visualisations Partner with development teams to improve services through rigorous testing and release procedures. Run our infrastructure with Gitlab CI/CD, Kubernetes, Kafka, NGINX and ELK stack. Co-ordinate with infra teams and developers to improvise the incident management process. Responsible for L1 support as well. Good Communication and Presentation skills Core Competencies(Must Have): Elastic, Logstash, Kibana or AppDynamics CI/CD Gitlab/Jenkins Other KeySkills SSO technologies Ansible Python Linux Administration Additional Competencies (Nice to have): Kubernetes Kafka, MQ NGINX or APIGEE Redis Experience in working with outsourced vendor teams for application development Appreciation of Enterprise Functional Architecture in Capital Markets Job Purpose: We are looking for a skilled and proactive Site Reliability Engineer (SRE) with strong expertise in deployment automation, monitoring, and infrastructure reliability. The ideal candidate will be responsible for managing the end-to-end deployment lifecycle, ensuring the availability, scalability, and performance of our production and non-production environments. Area of Operations Key Responsibility Deployment & Release Management Own and manage the CI/CD pipelines for automated build, test, and deployment. Design and implement robust deployment strategies for microservices and web applications. Monitor and troubleshoot deployment issues and rollbacks, ensuring zero-downtime deployment where possible System Reliability & Performance Set up and maintain monitoring, alerting, and logging frameworks (e.g., Prometheus, Grafana, ELK) Any Other Requirement: Should be a good team player. Would be required to work with multiple projects / teams concurrently

Posted 2 weeks ago

Apply

8.0 - 10.0 years

6 - 10 Lacs

Bengaluru

Work from Office

Company Overview: Maximus is a leading innovator in the government space, providing transformative solutions in the management and service delivery of government health and human services programs. We pride ourselves in our commitment to excellence, innovation, and a customer-first approach, driven by our core values. This has fostered our continual support of public programs and improving access to government services for citizens. Maximus continues to grow its Digital Solutions organization to better serve the needs of our organization, our customers in the government, health, and human services space, while improving access to government services for citizens. We use an approach grounded in design thinking, lean, and agile to help solve complicated problems and turn bold ideas into delightful solutions. Job Description: We are seeking a hands-on and strategic Lead DevOps Engineer to architect, implement, and lead the automation and CI/CD practices across our cloud infrastructure. This role demands deep expertise in cloud-native technologies and modern DevOps tooling with a strong emphasis on AWS, Kubernetes, ArgoCD, and Infrastructure Code. The ideal candidate is also expected to be a motivated self-starter with a proactive approach to resolving problems and issues with minimal supervision Key Responsibilities: Design and manage scalable infrastructure across AWS and Azure using Terraform (IaC) Define and maintain reusable Terraform modules to enforce infrastructure standards and best practices Implement secrets management , configuration management, and automated environment provisioning Architect and maintain robust CI/CD pipelines using Jenkins and ArgoCD Implement GitOps workflows for continuous delivery and environment promotion Automate testing, security scanning, and deployment processes across multiple environments Design and manage containerized applications with Docker Deploy and manage scalable, secure workloads using Kubernetes (EKS/ECS/GKE/AKS/self-managed) Create and maintain Helm charts , C ustomize configs, or other manifest templating tools Manage Git repositories, branching strategies, and code review workflows Promote version control best practices including commit hygiene and semantic release tagging Set up and operate observability stacks: Prometheus , Grafana , ELK , Loki , Alertmanager any of those. Define SLAs, SLOs, and SLIs for critical services Lead incident response , perform root cause analysis, and publish post-mortems documentations Integrate security tools and checks directly into CI/CD workflows Manage access control, secrets, and ensure compliance with standards such as FedRamp Mentor and guide DevOps engineers to build a high-performing team Collaborate closely with software engineers, QA, product managers, and security teams Promote a culture of automation , reliability , and continuous improvement Roles and Responsibilities Qualifications: Bachelor's degree in computer science, Information Security, or a related field (or equivalent experience). 8+ years of experience in DevOps or a similar role, with a strong security focus. Preferred AWS Certified Cloud Practitioner certification or AWS Certified Devops Engineer – Professional or AWS Certified Solution Architect or similar. Knowledge of cloud platforms (AWS) (Azure – Good to have) and containerization technologies (Docker, Kubernetes) with a key focus on AWS and EKS, ECS. Experience with infrastructure such as code (IaC) tools such as Terraform. Proficiency in CI/CD tools like AWS CodePipeline, Jenkins, Azure DevOps Server Familiarity with programming and scripting languages (e.g., Python, Bash, Go, Bash). Excellent problem-solving skills and the ability to work in a fast-paced, collaborative environment. Strong communication skills, with the ability to convey complex security concepts to technical and non-technical stakeholders. Preferred Qualifications: Strong understanding and working experience with enterprise applications and containerized application workloads. Knowledge of networking concepts Knowledge of network security principles and technologies (e.g., Firewalls, VPNs, IDS/IPS).

Posted 2 weeks ago

Apply

6.0 - 10.0 years

7 - 11 Lacs

Mumbai

Work from Office

We are looking for an experienced DevOps Engineer (Level 2 3) to design, automate, and optimize cloud infrastructure. You will play a key role in CI/CD automation, cloud management, observability, and security, ensuring scalable and reliable systems. Key Responsibilities: Design and manage AWS environments using Terraform/Ansible. Build and optimize deployment pipelines (Jenkins, ArgoCD, AWS CodePipeline). Deploy and maintain EKS, ECS clusters. Implement OpenTelemetry, Prometheus, Grafana for logs, metrics, and tracing. Manage and scale cloud-native microservices efficiently. Required Skills: Proven experience in DevOps, system administration, or software development. Strong knowledge of AWS. Programming languages: Python, Go, Bash, are good to have Experience with IAC tools like Terraform, Ansible Solid understanding of CI/CD tools (Jenkins, ArgoCD , AWS CodePipeline). Experience in containers and orchestration tools like Kubernetes (EKS) Understanding of OpenTelemetry observability stack (logs, metrics, traces Good to have: Experience with container orchestration platforms (e.g., EKS, ECS). Familiarity with serverless architecture and tools (e.g., AWS Lambda). Experience using monitoring tools like DataDog/ NewRelic, CloudWatch, Prometheus/Grafana Experience with managing more than 20+ cloud-native microservices. Previous experience of working in a startup Education Experience: Bachelor’s degree in Computer Science, Information Technology, or a related field (or equivalent work experience). Years of relevant experience in DevOps or a similar role.

Posted 2 weeks ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies