Home
Jobs

577 Prometheus Jobs - Page 6

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

1.0 - 4.0 years

5 - 9 Lacs

Mumbai

Work from Office

AI Opportunities with Soul AIs Expert Community!. Are you an MLOps Engineer ready to take your expertise to the next levelSoul AI (by Deccan AI) is building an elite network of AI professionals, connecting top-tier talent with cutting-edge projects. Why Join. Above market-standard compensation. Contract-based or freelance opportunities (2"“12 months). Work with industry leaders solving real AI challenges. Flexible work locations- Remote | Onsite | Hyderabad/Bangalore. Your Role:. Architect and optimize ML infrastructure with Kubeflow, MLflow, SageMaker Pipelines. Build CI/CD pipelines (GitHub Actions, Jenkins, GitLab CI/CD). Automate ML workflows (feature engineering, retraining, deployment). Scale ML models with Docker, Kubernetes, Airflow. Ensure model observability, security, and cost optimization in cloud (AWS/GCP/Azure). Must-Have Skills: . Proficiency in Python, TensorFlow, PyTorch, CI/CD pipelines. Hands-on experience with cloud ML platforms (AWS SageMaker, GCP Vertex AI, Azure ML). Expertise in monitoring tools (MLflow, Prometheus, Grafana). Knowledge of distributed data processing (Spark, Kafka). (BonusExperience in A/B testing, canary deployments, serverless ML). Next Steps:. Register on Soul AIs website. Get shortlisted & complete screening rounds. Join our Expert Community and get matched with top AI projects. Dont just find a job. Build your future in AI with Soul AI!.

Posted 6 days ago

Apply

1.0 - 5.0 years

9 - 13 Lacs

Mumbai

Work from Office

Step into the world of AI innovation with the Experts Community of Soul AI (By Deccan AI). We are looking for Indias top 1% Platform Engineers for a unique job opportunity to work with the industry leaders. Who can be a part of the community. We are looking for Platform Engineers focusing on building scalable and high-performance AI/ML platforms. Strong background in cloud architecture, distributed systems, Kubernetes, and infrastructure automation is expected. If you have experience in this field then this is your chance to collaborate with industry leaders. Whats in it for you. Pay above market standards. The role is going to be contract based with project timelines from 2 12 months, or freelancing. Be a part of an Elite Community of professionals who can solve complex AI challenges. Work location could be:. Remote (Highly likely). Onsite on client location. Deccan AIs OfficeHyderabad or Bangalore. Responsibilities:. Architect and maintain scalable cloud infrastructure on AWS, GCP, or Azure using tools like Terraform and Cloud Formation. Design and implement Kubernetes clusters with Helm, Kustomize, and Service Mesh (Istio, Linkerd). Develop CI/CD pipelines using GitHub Actions, GitLab CI/CD, Jenkins, and Argo CD for automated deployments. Implement observability solutions (Prometheus, Grafana, ELK stack) for logging, monitoring, and tracing & automate infrastructure provisioning with tools like Ansible, Chef, Puppet, and optimize cloud costs and security. Required Skills: . Expertise in cloud platforms (AWS, GCP, Azure) and infrastructure as code (Terraform, Pulumi) with strong knowledge of Kubernetes, Docker, CI/CD pipelines, and scripting (Bash, Python). Experience with observability tools (Prometheus, Grafana, ELK stack) and security practices (RBAC, IAM). Familiarity with networking (VPC, Load Balancers, DNS) and performance optimization. Nice to Have:. Experience with Chaos Engineering (Gremlin, LitmusChaos), Canary or Blue-Green deployments. Knowledge of multi-cloud environments, FinOps, and cost optimization strategies. What are the next steps. Register on our Soul AI website. Our team will review your profile. Clear all the screening roundsClear the assessments once you are shortlisted. Profile matching and Project AllocationBe patient while we align your skills and preferences with the available project. Skip the Noise. Focus on Opportunities Built for You!.

Posted 6 days ago

Apply

2.0 - 5.0 years

7 - 11 Lacs

Mumbai

Work from Office

Step into the world of AI innovation with the Experts Community of Soul AI (By Deccan AI). We are looking for Indias top 1% DevOps Engineers for a unique job opportunity to work with the industry leaders. Who can be a part of the community. We are looking for professionals skilled in infrastructure automation, CI/CD pipelines, cloud computing, and monitoring tools. Proficiency in Terraform, Kubernetes, Docker, and cloud platforms is required. If you have experience in this field then this is your chance to collaborate with industry leaders. Whats in it for you. Pay above market standards. The role is going to be contract based with project timelines from 2 12 months, or freelancing. Be a part of an Elite Community of professionals who can solve complex AI challenges. Work location could be:. Remote (Highly likely). Onsite on client location. Deccan AIs OfficeHyderabad or Bangalore. Responsibilities:. Implement and manage CI/CD pipelines for efficient software deployment. Automate infrastructure provisioning using tools like Terraform, Cloud Formation, or Ansible. Monitor system performance, troubleshoot issues, and ensure high availability and manage cloud environments (AWS, GCP, Azure) for scalability and security. Collaborate with development teams to ensure smooth integration and deployment processes. Required Skills: . Proficiency with CI/CD tools (Jenkins, GitLab CI, Circle CI) and infrastructure automation (Terraform, Ansible). Strong experience with cloud platforms (AWS, GCP, Azure) and containerization (Docker, Kubernetes). Familiarity with version control (Git), system monitoring tools (Prometheus, Grafana) and scripting languages (Python, Bash) for automation. Nice to Have:. Experience with server less architectures and microservices. Knowledge of security best practices and compliance (IAM, encryption). What are the next steps. Register on our Soul AI website. Our team will review your profile. Clear all the screening roundsClear the assessments once you are shortlisted. Profile matching and Project AllocationBe patient while we align your skills and preferences with the available project. Skip the Noise. Focus on Opportunities Built for You!.

Posted 6 days ago

Apply

1.0 - 4.0 years

5 - 9 Lacs

Kolkata

Work from Office

AI Opportunities with Soul AIs Expert Community!. Are you an MLOps Engineer ready to take your expertise to the next levelSoul AI (by Deccan AI) is building an elite network of AI professionals, connecting top-tier talent with cutting-edge projects. Why Join. Above market-standard compensation. Contract-based or freelance opportunities (2"“12 months). Work with industry leaders solving real AI challenges. Flexible work locations- Remote | Onsite | Hyderabad/Bangalore. Your Role:. Architect and optimize ML infrastructure with Kubeflow, MLflow, SageMaker Pipelines. Build CI/CD pipelines (GitHub Actions, Jenkins, GitLab CI/CD). Automate ML workflows (feature engineering, retraining, deployment). Scale ML models with Docker, Kubernetes, Airflow. Ensure model observability, security, and cost optimization in cloud (AWS/GCP/Azure). Must-Have Skills: . Proficiency in Python, TensorFlow, PyTorch, CI/CD pipelines. Hands-on experience with cloud ML platforms (AWS SageMaker, GCP Vertex AI, Azure ML). Expertise in monitoring tools (MLflow, Prometheus, Grafana). Knowledge of distributed data processing (Spark, Kafka). (BonusExperience in A/B testing, canary deployments, serverless ML). Next Steps:. Register on Soul AIs website. Get shortlisted & complete screening rounds. Join our Expert Community and get matched with top AI projects. Dont just find a job. Build your future in AI with Soul AI!.

Posted 6 days ago

Apply

2.0 - 5.0 years

7 - 11 Lacs

Kolkata

Work from Office

Step into the world of AI innovation with the Experts Community of Soul AI (By Deccan AI). We are looking for Indias top 1% DevOps Engineers for a unique job opportunity to work with the industry leaders. Who can be a part of the community. We are looking for professionals skilled in infrastructure automation, CI/CD pipelines, cloud computing, and monitoring tools. Proficiency in Terraform, Kubernetes, Docker, and cloud platforms is required. If you have experience in this field then this is your chance to collaborate with industry leaders. Whats in it for you. Pay above market standards. The role is going to be contract based with project timelines from 2 12 months, or freelancing. Be a part of an Elite Community of professionals who can solve complex AI challenges. Work location could be:. Remote (Highly likely). Onsite on client location. Deccan AIs OfficeHyderabad or Bangalore. Responsibilities:. Implement and manage CI/CD pipelines for efficient software deployment. Automate infrastructure provisioning using tools like Terraform, Cloud Formation, or Ansible. Monitor system performance, troubleshoot issues, and ensure high availability and manage cloud environments (AWS, GCP, Azure) for scalability and security. Collaborate with development teams to ensure smooth integration and deployment processes. Required Skills: . Proficiency with CI/CD tools (Jenkins, GitLab CI, Circle CI) and infrastructure automation (Terraform, Ansible). Strong experience with cloud platforms (AWS, GCP, Azure) and containerization (Docker, Kubernetes). Familiarity with version control (Git), system monitoring tools (Prometheus, Grafana) and scripting languages (Python, Bash) for automation. Nice to Have:. Experience with server less architectures and microservices. Knowledge of security best practices and compliance (IAM, encryption). What are the next steps. Register on our Soul AI website. Our team will review your profile. Clear all the screening roundsClear the assessments once you are shortlisted. Profile matching and Project AllocationBe patient while we align your skills and preferences with the available project. Skip the Noise. Focus on Opportunities Built for You!.

Posted 6 days ago

Apply

1.0 - 5.0 years

9 - 13 Lacs

Kolkata

Work from Office

Step into the world of AI innovation with the Experts Community of Soul AI (By Deccan AI). We are looking for Indias top 1% Platform Engineers for a unique job opportunity to work with the industry leaders. Who can be a part of the community. We are looking for Platform Engineers focusing on building scalable and high-performance AI/ML platforms. Strong background in cloud architecture, distributed systems, Kubernetes, and infrastructure automation is expected. If you have experience in this field then this is your chance to collaborate with industry leaders. Whats in it for you. Pay above market standards. The role is going to be contract based with project timelines from 2 12 months, or freelancing. Be a part of an Elite Community of professionals who can solve complex AI challenges. Work location could be:. Remote (Highly likely). Onsite on client location. Deccan AIs OfficeHyderabad or Bangalore. Responsibilities:. Architect and maintain scalable cloud infrastructure on AWS, GCP, or Azure using tools like Terraform and Cloud Formation. Design and implement Kubernetes clusters with Helm, Kustomize, and Service Mesh (Istio, Linkerd). Develop CI/CD pipelines using GitHub Actions, GitLab CI/CD, Jenkins, and Argo CD for automated deployments. Implement observability solutions (Prometheus, Grafana, ELK stack) for logging, monitoring, and tracing & automate infrastructure provisioning with tools like Ansible, Chef, Puppet, and optimize cloud costs and security. Required Skills: . Expertise in cloud platforms (AWS, GCP, Azure) and infrastructure as code (Terraform, Pulumi) with strong knowledge of Kubernetes, Docker, CI/CD pipelines, and scripting (Bash, Python). Experience with observability tools (Prometheus, Grafana, ELK stack) and security practices (RBAC, IAM). Familiarity with networking (VPC, Load Balancers, DNS) and performance optimization. Nice to Have:. Experience with Chaos Engineering (Gremlin, LitmusChaos), Canary or Blue-Green deployments. Knowledge of multi-cloud environments, FinOps, and cost optimization strategies. What are the next steps. Register on our Soul AI website. Our team will review your profile. Clear all the screening roundsClear the assessments once you are shortlisted. Profile matching and Project AllocationBe patient while we align your skills and preferences with the available project. Skip the Noise. Focus on Opportunities Built for You!.

Posted 6 days ago

Apply

5.0 - 10.0 years

10 - 20 Lacs

Pune, Chennai, Mumbai (All Areas)

Hybrid

Job Title: Backend Engineer Python + AI Integration Location: India (Hybrid - Preferably Chennai/ Mumbai / Pune) Experience Required: Minimum 5 Years Joining Timeline: Immediate to 30 Days Preferred Role Overview: We are seeking a seasoned Backend Engineer with strong proficiency in Python and a solid understanding of AI/ML model integration . This role is ideal for someone who thrives at the intersection of backend engineering and intelligent systems building scalable APIs, handling data workflows, and integrating machine learning models in production environments. Key Responsibilities: 1. Backend & API Development Develop RESTful and GraphQL APIs using Django, FastAPI, or Flask . Implement async tasks with tools like Celery, RabbitMQ , and webhooks. Design clean, scalable, and secure architecture following SOLID principles . 2. Database Design & Optimization Design, maintain, and optimize PostgreSQL databases JSONB, partitioning, window functions, materialized views, etc. Write complex SQL queries , design ER diagrams, manage schema migrations (e.g., Alembic, Flyway). Troubleshoot performance issues using EXPLAIN plans , handle connection pooling, and resolve deadlocks. 3. AI/ML System Integration Work closely with data scientists to deploy ML models as APIs. Use tools like scikit-learn, PyTorch, Hugging Face, TensorFlow for integrating AI capabilities. Deploy RAG systems using FAISS, Weaviate, Qdrant , and integrate with OpenAI APIs, LangChain, LlamaIndex . Build pipelines with Airflow, Prefect for continuous training and deployment. 4. Infrastructure & DevOps Containerize applications using Docker & Docker Compose . Implement CI/CD pipelines using GitHub Actions, GitLab CI, or Jenkins . Monitor systems using Prometheus, Grafana, ELK, or Sentry . Familiarity with cloud platforms ( AWS, GCP, Azure ) working with S3, Lambda, Cloud SQL, SageMaker , etc. Must-Have Skills: Proficiency in Python and OOP principles. Deep knowledge of PostgreSQL and general RDBMS optimization. Experience with RESTful APIs , async processing, and microservice design. Exposure to AI/ML workflows , including model deployment and monitoring. Knowledge of authentication/authorization standards (OAuth2, JWT). Good-to-Have (Bonus Skills): Experience with LLMs , embedding-based search , and RAG systems . Familiarity with Streamlit/Dash for internal tools and dashboards. Understanding of data governance , PII protection, and anonymization. Exposure to event-driven systems (Kafka, AWS SNS/SQS). Open-source contributions or technical blog writing. Interested Candidates can drop your resume to subashini.gopalan@kiya.ai

Posted 6 days ago

Apply

2.0 - 7.0 years

11 - 15 Lacs

Pune

Work from Office

: Job TitleL2 Lead Technical Application Support, Associate LocationPune, India Role Description Our organization within Deutsche Bank is Compliance Production Services. We are responsible for providing technical L2 application support for business applications. The Compliance line of business has a current portfolio of 20 applications. The organization is in process of transforming itself using Google Cloud and many new technology offerings. As a L2 Lead Technical Application Support You will provide technical hands-on oversight to several support teams and be actively involved in technical issues resolution across multiple applications. You will also be application working as application lead and will be responsible for technical & operational processes for all application you support. What well offer you , 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Your key responsibilities Act as application lead , You need to own the responsibilities related technical, process, operational, and people for all applications supported. Provide technical hands-on oversight to several support teams and be actively involved in technical issues across multiple applications. Build up technical subject matter expertise on the applications being supported including business flows, application architecture, and hardware configuration. Maintain documentation, knowledge articles, and runbooks. Assist in the process to approve application code releases change tickets as well as tasks assigned to support to perform. Build and maintain effective and productive relationships with the stakeholders in business, development, infrastructure, and third-party systems / data providers & vendors. Assist in special projects and view them as opportunities to enhance your skillset and develop your growth. These projects can include coding using shell scripting, Python and YAML language for support functions. Your skills and experience Minimum 2 years of experience in providing the hands-on IT support and interacting with applications and end users. Engineering Degree/Post graduation from an accredited college or university with a concentration in Computer Science or IT-related discipline. knowledgeable in cloud products like Google Cloud Platform (GCP) and hybrid applications. Strong understanding of ITIL /SRE/ DEVOPS best practices for supporting a production environment. Working knowledge of Elastic Search, WebLogic, Tomcat, OpenShift, Grafana, and Prometheus, Google Cloud Monitoring. Understanding of Java (J2SE), spring, Hibernate, micro services. Red Hat Enterprise Linux (RHEL) professional skill in searching logs, process commands, start/stop processes, use of OS commands to aid in tasks needed to resolve or investigate issues. Shell scripting knowledge a plus. Understanding of database concepts and exposure in working with oracle and SQL databases. Skills That Will Help You Excel Strong written and oral communication skills, including the ability to communicate technical information to a non-technical audience and good analytical and problem-solving skills. Able to train, coach, and mentor and know where each technique is best applied. Confident working with several programming languages, tools, and technologies, including Infrastructure as Code, with the ability to guide colleagues as to the context where each is useful (preferably Python and Terraform) . Experience with GCP or another public cloud provider to build applications. Experience in an investment bank, financial institution or large corporation using enterprise hardware and software. How well support you . . . . About us and our teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.

Posted 6 days ago

Apply

5.0 - 10.0 years

35 - 40 Lacs

Pune

Work from Office

: Job TitleServiceNow SRE Support Consultant LocationPune, India Corporate TitleAVP Role Description We are seeking a ServiceNow SRE Support Consultant with 5+ years of experience to ensure the stability, scalability, and reliability of our ServiceNow platform . This role will focus on monitoring, troubleshooting, automation, and performance optimization of the ServiceNow environment while applying SRE and DevOps best practices . The ideal candidate should have a strong background in ServiceNow administration, UNIX/Linux, Windows Server, and cloud infrastructure . What well offer you 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Your key responsibilities ServiceNow Platform Support: Monitor and maintain the health, availability, and performance of the ServiceNow platform. Troubleshoot ServiceNow infrastructure issues , including instance performance, integrations, and database bottlenecks. Collaborate with developers and architects to enhance system stability and optimize workflows. Perform ServiceNow instance upgrades, patching, cloning, and configuration tuning . Incident Management & Troubleshooting: Handle incident response and escalations to resolve ServiceNow-related issues efficiently. Conduct root cause analysis (RCA) and implement long-term fixes. Define and maintain SLOs/SLIs to measure and improve system reliability. Work closely with the ServiceNow support team and infrastructure teams to minimize downtime. System Administration & Automation: Manage ServiceNow MID Servers and integrations with third-party systems. Administer UNIX/Linux, and Windows Server environments supporting ServiceNow. Automate routine administration tasks using PowerShell, Bash, Python, or Ansible . Monitoring & Performance Optimization: Implement and maintain proactive monitoring solutions using Splunk, ELK, Prometheus, Grafana, or ServiceNow Event Management . Optimize ServiceNow database performance, query execution, and API integrations . Perform capacity planning and performance tuning to ensure seamless scalability. Your skills and experience Technical Expertise: 5+ years of experience in ServiceNow administration, support, or SRE roles . Strong knowledge of ServiceNow architecture, modules, and instance management . Experience in UNIX/Linux, and Windows Server administration . Hands-on expertise in scripting (Bash, PowerShell, Python) and automation tools (Ansible, Terraform, Jenkins, Git) . Proficiency in networking (DNS, TCP/IP, Load Balancing, Firewalls) related to ServiceNow. ServiceNow Knowledge: Experience with ServiceNow ITOM (Discovery, Event Management, CMDB, Performance Analytics) . Understanding of ServiceNow upgrades, patching, and instance cloning . Basic knowledge of ServiceNow scripting (JavaScript, REST API, Web Services) . Soft Skills & Collaboration: Strong problem-solving and analytical skills . Ability to work in a fast-paced environment and manage multiple tasks. Excellent communication skills to collaborate with cross-functional teams. Proactive mindset with a focus on automation and continuous improvement . Preferred Qualifications: ServiceNow Certified System Administrator (CSA) or ITOM certifications. Linux (RHCSA, RHCE) or Windows (MCSA, MCSE) certifications. Cloud certifications (AWS, Azure, GCP) are a plus. How well support you About us and our teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.

Posted 6 days ago

Apply

10.0 - 15.0 years

14 - 19 Lacs

Pune

Work from Office

: Job Title Principal Engineer Location Pune As a principal software engineer, you will be responsible for designing, developing, and maintaining core parts our software and infrastructure, contributing heavily to the codebase and collaboration with engineers at all levels. You will play a pivotal role in the shaping of our architecture, ensuring robustness of our systems, and mentoring junior engineers to help them elevate their skills. This role is ideal for someone who enjoys working on challenging technical problems, has a deep understanding of modern technology trends, and is passionate about software craftsmanship. This is purely a technical position with no people management responsibilities. What well offer you 100% reimbursement under child care assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Key responsibilities Design, develop and maintain high-performance, scalable software in Java, Kotlin, Contribute actively to the codebase, ensuring quality, performance, and reliability, Develop solutions using MongoDB and work on optimization, indexing and queries, Architect and implement micro-services deployed in GKE, Ensure compliance with security regulations Review and update policies relevant to internal systems and equipment Mentor and guide engineers across multiple teams, setting standard for technical excellence, Collaborate with product managers, architects, and cross-functional teams to translate business requirements into technical solutions Qualification: 10+ years of professional software development experience e, with expertise in java, Strong experience with MongoDB and working with Data-intensive applications Experience with modern software engineering practices, including test-driven development, continuous integration, and agile methodologies. Solid hands-on experience with Kubernetes, Experience designing and running systems at scale in cloud environments, preferably GCP, Familiarity with CI/CD tools, monitoring logging and alerting stacks (e.g. Prometheus, Grafana, ELK) Strong experience with reactive or event-driven architectures, Experience with infra as code tooling, e.g. Terraform. How well support you About us and our teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.

Posted 6 days ago

Apply

4.0 - 9.0 years

14 - 19 Lacs

Pune

Work from Office

: Job Title Principal Engineer Location Pune As a principal software engineer, you will be responsible for designing, developing, and maintaining core parts our software and infrastructure, contributing heavily to the codebase and collaboration with engineers at all levels. You will play a pivotal role in the shaping of our architecture, ensuring robustness of our systems, and mentoring junior engineers to help them elevate their skills. This role is ideal for someone who enjoys working on challenging technical problems, has a deep understanding of modern technology trends, and is passionate about software craftsmanship. This is purely a technical position with no people management responsibilities. What well offer you 100% reimbursement under child care assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Key responsibilities: Design, develop and maintain high-performance, scalable software in Java, Kotlin, Contribute actively to the codebase, ensuring quality, performance, and reliability, Develop solutions using MongoDB and work on optimization, indexing and queries, Architect and implement micro-services deployed in GKE, Ensure compliance with security regulations Review and update policies relevant to internal systems and equipment Mentor and guide engineers across multiple teams, setting standard for technical excellence, Collaborate with product managers, architects, and cross-functional teams to translate business requirements into technical solutions Qualification: 10+ years of professional software development experience e, with expertise in java, Strong experience with MongoDB and working with Data-intensive applications Experience with modern software engineering practices, including test-driven development, continuous integration, and agile methodologies. Solid hands-on experience with Kubernetes, Experience designing and running systems at scale in cloud environments, preferably GCP, Familiarity with CI/CD tools, monitoring logging and alerting stacks (e.g. Prometheus, Grafana, ELK) Strong experience with reactive or event-driven architectures, Experience with infra as code tooling, e.g. Terraform. How well support you About us and our teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.

Posted 6 days ago

Apply

15.0 - 20.0 years

45 - 50 Lacs

Pune

Work from Office

: Job Title Solution Architect, VP LocationPune, India Corporate TitleVP Role Description Your role will be an Individual contributor in the team. You will be closely working with team comprising of engineers, Lead, functional analysts, and test lead. The team is responsible for developing and implementing micro-services, Front end Application development & enhancements, integrating another partner and client integrations. As a Solution architect you are expected to give the team be hands on with software development, contribute towards good software design and test developed software. You will also be engaged in peer code reviews, document design decisions, and components APIs. You will be participating in daily stand up meetings, analysing software defects and fixing them in a timely manner, and working closely with the Functional Analysis and Quality Assurance teams. As/when required to, you are also expected to train other team members to bring them up to speed. Banks Corporate Bank division is a leading provider of cash management, trade finance and securities finance. We complete green-field projects that deliver the best Corporate Bank - Securities Services products in the world. Our team is diverse, international, and driven by shared focus on clean code and valued delivery. At every level, agile minds are rewarded with competitive pay, support, and opportunities to excel. You will work as part of a cross-functional agile delivery team. You will bring an innovative approach to software development, focusing on using the latest technologies and practices, as part of a relentless focus on business value. You will be someone who sees engineering as team activity, with a predisposition to open code, open discussion and creating a supportive, collaborative environment. You will be ready to contribute to all stages of software delivery, from initial analysis right through to production support. What well offer you 100% reimbursement under childcare assistance benefit (gender neutral) Sponsorship for Industry relevant certifications and education Accident and Term life Insurance Your key responsibilities As a Solution Architect, you will be responsible for creating the Architecture and Design of the applications on the Cloud Platform. You are expected to have practical architecture depth to meet the business and technical requirements and expected to have hands on Engineering Experience. You are expected to create the Design and Architecture blueprint for the multi-region, highly available applications on Google Cloud Platform. You will be responsible for the design and implementation of various non-functional requirements like Security, Scalability, Observability, Disaster Recovery, Data Protection. You will provide the technical leadership to the Engineering teams and deliver the application releases. You will work with the SRE and support teams to help bring Architectural improvements. Your skills and experience Expert level Cloud Architecture experience at the Solution Design and implementation level. Overall 15+ years of hands-on coding and engineering experience with at least 5 years of experience in designing and building applications on Cloud Platforms. Cloud Certifications for GCP (Preferred) or AWS or Azure. Well versed with the Well Architected Framework pillars like Security, Availability, Reliability, Operational Excellence, Cost Optimization. Hands on experience in Cloud Services like Kubernetes, API gateways, Load Balancers, Cloud Storage Services, VPCs, NAT Gateways, Cloud SQL databases, VMs and compute services like Cloud Run. Hands on development experience in developing applications using Core Java, Sprint Boot, REST APIs, Databased like Oracle, MongoDB, Apache Kafka. Good knowledge about Frontend technologies like JavaScript, React.js, TypeScript. Experience in designing multi-region Disaster Recovery (DR) solutions and achieving the Recovery Point Objectives (RPO) and Recovery Time Objectives (RTO). Experience in building highly available, low latency and high-volume applications, performance testing and tuning. Good knowledge about Microservices architecture. Working knowledge DevOps tools like Jenkins/ GitHub actions, Terraform, Helm Chart. Experience in building application Observability using tools like Prometheus/Grafana, New Relic and creating SLO dashboards. Good understanding of security principle like encryption techniques, handling security vulnerabilities, and building solutions to prevent DDoS attacks. Nice to have skills. FunctionalPayment Industry overview, Payment processing, Real-time payments processing Shell Scripting is nice to have Change management process exposure Software and infra production promotion experience Test Automation Frameworks Moderate coding skills on Python. Experience in distributed system development. Cross-platform development in several CPU/operating system environments and network protocols. Demonstrated expertise in problem-solving and technical innovation Data Structures, Algorithms and Design Patterns Data stores, persistence, caching (Oracle, MongoDB, Cassandra, and Hadoop tools, memcache etc) How well support you About us and our teams Please visit our company website for further information: https://www.db.com/company/company.htm We strive for a culture in which we are empowered to excel together every day. This includes acting responsibly, thinking commercially, taking initiative and working collaboratively. Together we share and celebrate the successes of our people. Together we are Deutsche Bank Group. We welcome applications from all people and promote a positive, fair and inclusive work environment.

Posted 6 days ago

Apply

6.0 - 11.0 years

5 - 15 Lacs

Pune

Work from Office

Job Description: Mandatory Skills: Design, develop, and maintain Grafana dashboards to visualize data from applications developed with Go lang, Flutter, and Python, etc Integrate Grafana with various data sources, including Prometheus, InfluxDB, Elasticsearch, and other relevant systems Good to have Skills : Database Knowledge: Strong understanding of Elasticsearch and other databases. Core Java Knowledge: Basic knowledge of Core Java is a plus. CI/CD Processes: Experience with Continuous Integration/Continuous Deployment (CI/CD) processes is beneficial. Detailed JD Position Overview: We are seeking a skilled Grafana Developer to join our team. The ideal candidate will be responsible for designing, developing, and maintaining Grafana dashboards to visualize operational and business data. This role requires a deep understanding of data integration, performance optimization, and user-centric design . Key Responsibilities: Design, develop, and maintain Grafana dashboards to visualize data from applications developed with Go lang, Flutter, and Python. Integrate Grafana with various data sources, including Prometheus, InfluxDB, Elasticsearch, and other relevant systems Performance Optimization: Optimize dashboards for performance, scalability, and real-time insights. Stakeholder Collaboration: Work closely with stakeholders to understand their data visualization requirements and ensure dashboards meet their needs. User-Friendly Design: Ensure dashboards are user-friendly, intuitive, and aligned with organizational goals. Required Skills: Grafana Expertise: Proven experience with Grafana and other data visualization tools. Data Integration: Proficiency in integrating various data sources into Grafana. Database Knowledge: Strong understanding of Elasticsearch and other databases. Core Java Knowledge: Basic knowledge of Core Java is a plus. CI/CD Processes: Experience with Continuous Integration/Continuous Deployment (CI/CD) processes is beneficial. Preferred Qualifications: Experience: 6-8 years of experience in software development. Real-Time Monitoring: Familiarity with real-time monitoring solutions. Team Collaboration: Ability to work effectively as part of a cross-functional team

Posted 6 days ago

Apply

2.0 - 5.0 years

4 - 9 Lacs

Chennai

Work from Office

Job Title: Dashboard Developer Location: Chennai/INDIA Requirements: Ability to work between 2pm to 11pm IST supporting client base in the U.S.A Looking for immediate joiners About the Role: This role is to supplement the IT organization in developing dashboards from various sources that generates system related data, precisely from monitoring systems with the expertise into basic operations of monitoring. Additionally able to script for minimal automations that are required in this domain. Responsibilities: Must be able to develop own dashboards using tools like Power BI, Graphane and Perses etc. Should understand monitoring systems like LogicMonitor, Prometheus or any other open-source monitoring system Able to correlate the data streams and make dashboards to drive business efficiency Candidate Requirements: Must have a good working knowledge in Power BI and Graphane (Other open-source dashboard solutions are added advantage) Need to have a good knowledge about Prometheus, an open source monitoring tool Good working knowledge in managing dockers and Linux DevOps is added advantage however not mandatory Strong understanding of infrastructure is required Good communication skills is required Ability and experience in working with stakeholders of different geographies Willing to learn and scale up as needed

Posted 6 days ago

Apply

4.0 - 9.0 years

6 - 11 Lacs

Bengaluru

Work from Office

About us: As a Fortune 50 company with more than 400,000 team members worldwide, Target is an iconic brand and one of America's leading retailers. Joining Target means promoting a culture of mutual care and respect and striving to make the most meaningful and positive impact. Becoming a Target team member means joining a community that values different voices and lifts each other up. Here, we believe your unique perspective is important, and you'll build relationships by being authentic and respectful. Overview about TII At Target, we have a timeless purpose and a proven strategy. And that hasnt happened by accident. Some of the best minds from different backgrounds come together at Target to redefine retail in an inclusive learning environment that values people and delivers world-class outcomes. That winning formula is especially apparent in Bengaluru, where Target in India operates as a fully integrated part of Targets global team and has more than 4,000 team members supporting the companys global strategy and operations. (Pyramid overview) Network Security Monitoring (NSM) Position About Network Security Monitoring: Target's Network Security Monitoring (NSM) team builds and maintains a fleet of over 2000 network sensors across the globe, providing network visibility and advanced monitoring capabilities to our Cyber Defense organization. We build scalable and maintainable infrastructure with full end-to-end ownership of both the hardware and software lifecycle. Our work enables timely detection and response of adversaries by delivering reliable network visibility through a resilient sensor grid and advanced monitoring capability. Team Overview NSM team members regularly: - Collaborate with Networking partners on network design and network sensor placement - Build, deploy, and upgrade network sensors (servers) globally - Design and implement network traffic analysis solutions using engines like Zeek and Suricata - Leverage Salt for configuration management, deployment automation, and infrastructure-as-code implementation - Partner with Cyber Defense to build network-based detections and consult in response scenarios - Develop performance monitoring solutions to track data quality and sensor health to ensure grid health and data fidelity Position Overview Expect to: - Configure, troubleshoot, and optimize network sensors across diverse environments - Debug complex networking issues and perform packet-level analysis to ensure proper traffic visibility. - Build and maintain Salt-based automation for configuration management and deployment. - Analyze monitoring data to identify system improvements and validate detection coverage. - Develop and automate testing to ensure results and outcomes are as expected. - Participate in on-call rotations to support the global sensor grid and respond to critical issues. - Collaborate cross-functionally with teams throughout Cyber Defense and IT - Document operational procedures for sensor management best practices - Research new network security monitoring technologies and evaluate their potential implementation. - Contribute to capacity planning and architectural design of monitoring infrastructure. - Manage and maintain Linux/Unix-based systems that host Zeek sensors, ensuring high availability, performance, and security. - Perform OS-level troubleshooting, patching, and hardening of sensor infrastructure. - Automate server provisioning and configuration using tools like Salt, shell scripting, and Python. - Monitor system logs and metrics to proactively identify and resolve issues affecting sensor performance. About you: - Bachelor's degree in Networking, Computer Science, or related field (or equivalent experience). - 4+ years of experience in network administration, network security, or related roles, with a deep knowledge of network protocols and packet analysis. - Experience with network security monitoring tools, including Zeek and Suricata. - Strong foundation in automation and infrastructure as code, Salt experience preferred. - You understand CI/CD principles and can implement pipelines for testing and deploying code and configuration changes. - Proficient in Linux/Unix systems administration, including shell scripting, system tuning, and troubleshooting. - Hands-on experience managing server infrastructure in production environments, including patching, upgrades, and performance tuning. - Practical experience with packet capture technologies and traffic analysis tools. - Proven ability to troubleshoot complex distributed systems and methodically diagnose network issues. - You appreciate the importance of dev/prod parity and can design for consistent environments across dev and prod. - Experience writing custom detection rules and understanding their performance implications. - Familiarity with technologies such as Zabbix, Prometheus, Nagios, Grafana, Elastic, Kibana

Posted 6 days ago

Apply

4.0 - 8.0 years

8 - 18 Lacs

Bengaluru

Hybrid

Job Title: DevOps L2 Support Engineer Experience: 46 Years Work Type: Full-time | Rotational Shift (24x7) Shift Window: Between 09:30 AM 10:30 PM Job Description: We are looking for a DevOps L2 Support Engineer with strong debugging capabilities and a solid foundation in Linux and scripting. The ideal candidate should be able to work in a fast-paced production support environment and proactively resolve issues related to infrastructure, deployment, and monitoring systems. Key Responsibilities: Perform root cause analysis and debug Java-based applications in a production environment Monitor and support systems hosted in Linux environments Write and manage shell scripts for automation and operational tasks Manage and troubleshoot Docker containers and commands Analyze logs and build insightful dashboards using Splunk Write effective Splunk queries to identify and resolve issues based on error signatures Collaborate with development and infrastructure teams to ensure seamless deployments and operations Participate in a 24x7 rotational shift , including two daily shifts between 9:30 AM to 10:30 PM Required Skills: Strong debugging skills in Java-based environments Hands-on experience with Linux OS and command-line tools Experience with Shell Scripting Working knowledge of Docker and related commands Proficiency in Splunk query writing and dashboard creation Excellent communication skills (both written and verbal) Good to Have: Exposure to Prometheus for system monitoring Understanding of CI/CD pipelines and basic DevOps concepts

Posted 6 days ago

Apply

1.0 - 4.0 years

1 - 5 Lacs

Mumbai

Work from Office

DevOps Engineer (2–8 yrs) – Mumbai. Experience in CI/CD (Jenkins/GitLab), Docker, Kubernetes, Terraform/Ansible, AWS/Azure, Git, and monitoring tools (Prometheus/Grafana). Strong scripting & DevOps practices.

Posted 6 days ago

Apply

7.0 - 9.0 years

19 - 25 Lacs

Pune

Work from Office

Location: Pune Experience: 7 - 9 Years Notice Period: Immediate to 15 Days Overview We are looking for an experienced IT Operations (Monitoring & Observability) Consultant to design, implement, and optimize end-to-end observability solutions. The ideal candidate will have a strong background in monitoring frameworks, ITSM integrations, and AIOps tools to drive system reliability, performance, and proactive incident management. Key Responsibilities Design and deploy comprehensive monitoring and observability architectures for infrastructure, applications, and networks. Implement tools like Prometheus, Grafana, OpsRamp, Dynatrace, New Relic for system performance monitoring. Integrate monitoring systems with ITSM platforms (e.g., ServiceNow, BMC Remedy). Develop dashboards, alerts, and reports to enable real-time performance insights. Architect solutions for hybrid and multi-cloud environments. Automate alerting, remediation, and reporting to streamline operations. Apply AIOps and ML for anomaly detection and predictive insights. Collaborate with DevOps, infra, and app teams to embed monitoring into CI/CD. Document architectures, procedures, and operational playbooks. Required Skills Hands-on experience with observability tools: Prometheus, Grafana, ELK Stack, Fluentd, Dynatrace, New Relic, OpsRamp . Strong scripting knowledge in Python, Ansible . Familiar with tracing tools (e.g., Jaeger, Zipkin ) and REST API integrations . Working knowledge of AIOps concepts and predictive monitoring. Solid understanding of ITIL processes and service management frameworks . Familiarity with security monitoring and compliance considerations. Excellent analytical, troubleshooting, and documentation skills.

Posted 6 days ago

Apply

6.0 - 10.0 years

10 - 20 Lacs

Bengaluru

Work from Office

Who We Are At Kyndryl, we design, build, manage and modernize the mission-critical technology systems that the world depends on every day. So why work at Kyndryl? We are always moving forward – always pushing ourselves to go further in our efforts to build a more equitable, inclusive world for our employees, our customers and our communities. The Role Are you ready to join the team of software engineering experts at Kyndryl? We are seeking a talented Software Engineering Technical Specialist to contribute to our software engineering space and provide critical skills required for the development of cutting-edge products. As a Software Engineering Technical Specialist, you will develop solutions in specific domains such as Security, Systems, Databases, Networking Solutions, and more. You will be a leader – contributing knowledge, guidance, technical expertise, and team leadership skills. Your leadership will be demonstrated in your work, to your customers, and within your teams. At Kyndryl, we value effective communication and collaboration skills. When you recognise opportunities for business change, you will have the ability to clearly and persuasively communicate complex technical and business concepts to both customers and team members. You’ll be the go-to person for problem-solving of customers’ business and technical issues. You have a knack for effectively identifying and framing problems, leading the collection of elements of information, and integrating this information to produce timely and thoughtful decisions. Your aim throughout, is to improve the effectiveness, efficiency and delivery of services through the use of technology and technical methods and methodologies. Driving the design, development, integration, delivery, and evolution of highly scalable distributed software you will integrate with other layers and offerings. You will provide deeper functionality and solutions to address customer needs. You will work closely with software engineers, architects, product managers, and partner teams to get high-quality products and features through the agile software development lifecycle. Your continuous grooming of features/user stories to estimate, identify technical risks/dependencies and clearly communicate them to project stakeholders will ensure the features are delivered with the right quality and within timeline. You will maintain and drive the clearing of technical debt, vulnerabilities, and currency of the 3rd party components within the product. As a Software Engineering Technical Specialist, you will also coach and mentor engineers to design and implement highly available, secure, distributed software in a scalable architecture. This is an opportunity to make a real impact and contribute to the success of Kyndryl's innovative software products. Join us and become a key player in our team of software engineering experts! Your Future at Kyndryl Every position at Kyndryl offers a way forward to grow your career. We have opportunities that you won’t find anywhere else, including hands-on experience, learning opportunities, and the chance to certify in all four major platforms. Whether you want to broaden your knowledge base or narrow your scope and specialize in a specific sector, you can find your opportunity here. Who You Are You’re good at what you do and possess the required experience to prove it. However, equally as important – you have a growth mindset; keen to drive your own personal and professional development. You are customer-focused – someone who prioritizes customer success in their work. And finally, you’re open and borderless – naturally inclusive in how you work with others. Required Technical and Professional Expertise 5–8 years of experience in infrastructure monitoring, logging, or performance engineering. Strong experience designing and implementing observability solutions using: Azure Monitor / Application Insights / Log Analytics Prometheus / Grafana / Loki / ELK / Splunk OpenTelemetry, Fluentd/Fluent Bit, Jaeger Familiarity with microservices, Kubernetes (AKS), and cloud-native patterns Experience working with CI/CD tools (GitHub Actions, Azure DevOps) and automation (Terraform, ARM, Bicep). Knowledge of ITSM and incident management workflows Preferred Technical and Professional Experience Strong communication and documentation skills. Ability to balance multiple priorities across delivery teams. Collaborative mindset with a focus on solution quality and operational reliability Being You Diversity is a whole lot more than what we look like or where we come from, it’s how we think and who we are. We welcome people of all cultures, backgrounds, and experiences. But we’re not doing it single-handily: Our Kyndryl Inclusion Networks are only one of many ways we create a workplace where all Kyndryls can find and provide support and advice. This dedication to welcoming everyone into our company means that Kyndryl gives you – and everyone next to you – the ability to bring your whole self to work, individually and collectively, and support the activation of our equitable culture. That’s the Kyndryl Way. What You Can Expect With state-of-the-art resources and Fortune 100 clients, every day is an opportunity to innovate, build new capabilities, new relationships, new processes, and new value. Kyndryl cares about your well-being and prides itself on offering benefits that give you choice, reflect the diversity of our employees and support you and your family through the moments that matter – wherever you are in your life journey. Our employee learning programs give you access to the best learning in the industry to receive certifications, including Microsoft, Google, Amazon, Skillsoft, and many more. Through our company-wide volunteering and giving platform, you can donate, start fundraisers, volunteer, and search over 2 million non-profit organizations. At Kyndryl, we invest heavily in you, we want you to succeed so that together, we will all succeed. Get Referred! If you know someone that works at Kyndryl, when asked ‘How Did You Hear About Us’ during the application process, select ‘Employee Referral’ and enter your contact's Kyndryl email address.

Posted 1 week ago

Apply

10.0 - 12.0 years

16 - 20 Lacs

Bengaluru

Work from Office

About the Job The Data & AI team is a highly focused effort to lead digital-first execution and transformation at Red Hat leveraging data strategically for our customers, partners, and associates. Radical CollaborationThere is no work done in isolation, as such, each team has to strive to collaborate with teams within the group, cross-group, and the communities. You will strive to make these collaborations as seamless as possible using tools, processes, best practices, and your own brand of creative problem-solving. Continuous LearningThis is a fast paced team and you are expected to be continuously curious, have a can do attitude and be proficient in understanding multiple aspects of the business, continuous improving your skill sets (technical and business) as the industry progresses Data and AI team is looking for a Engineering Manager to lead the Platform practice for the next generation SaaS based data and AI products. You will interact with product managers, Red Hat Sales, Marketing, Finance teams and data platform and product engineers to deliver a sophisticated data as-a-service platform. You'll coach and develop software engineers as they build the Platform, Infrastructure-as-code components, platform observability, agentic AI capabilities and other software to autonomously manage the environment, and guide problem management resolution (PMR) analysis when things go wrong. Youll work in a fast-paced globally distributed team while quickly learning new skills and creating ways to consistently meet service-level agreements (SLAs) for our data products.This role requires a leader with a proven record of navigating the complexities of working across multiple organizations, helping define and gain consensus on strategy and direction, and aligning the team(s) toward those end goals. What you will do Support engineering team to foster and deliver in an inner-source manner Develop, and retain a team of engineers developing and operating Red Hats data-as-service platform Coach engineers on good engineering principleswriting good code, automation, observability, toil reduction, and root cause analysis Manage high-visibility project delivery, including estimation, schedule, risks, and dependencies Design processes and communication norms that facilitate coordination across a fast-moving, fast-growing, diverse global team Lead your team through frequent changes in organization, process, and technology commensurate with a high growth cloud service in a competitive market Participate in a periodic 24x7 management escalation on-call rotation What you will bring 10-12 years of hands on developing and maintaining software. 5+ years experience managing high performing engineering teams Previous software engineering experience delivering data products, applications, or services on cloud native or hybrid platforms Experience with Agile methodologies and working in a DevOps culture with continuous integration / continuous deliveries Ability to lead distributed, remote teams working across multiple time zones Ability to discuss complex technical issues with engineers, product managers, and less-technical stakeholders including customers and senior leaders Understand and collaborate with compliance teams to make sure Platform and Products are compliant as per regulation. Experience hiring and developing engineers Experience in communication with stakeholder and leadership The following are considered a plus: Experience with platforms like Kuberentes/OpenShift and Kubernetes Operators, Prometheus, Graphana etc Experience with Go and Python for developing scaling backend software. Experience with building full stack applications Knowledge of SaaS technologies like Snowflake, Fivetran, Astro etc. About Red Hat Red Hat is the worlds leading provider of enterprise open source software solutions, using a community-powered approach to deliver high-performing Linux, cloud, container, and Kubernetes technologies. Spread across 40+ countries, our associates work flexibly across work environments, from in-office, to office-flex, to fully remote, depending on the requirements of their role. Red Hatters are encouraged to bring their best ideas, no matter their title or tenure. We're a leader in open source because of our open and inclusive environment. We hire creative, passionate people ready to contribute their ideas, help solve complex problems, and make an impact. Inclusion at Red Hat Red Hats culture is built on the open source principles of transparency, collaboration, and inclusion, where the best ideas can come from anywhere and anyone. When this is realized, it empowers people from different backgrounds, perspectives, and experiences to come together to share ideas, challenge the status quo, and drive innovation. Our aspiration is that everyone experiences this culture with equal opportunity and access, and that all voices are not only heard but also celebrated. We hope you will join our celebration, and we welcome and encourage applicants from all the beautiful dimensions that compose our global village. Equal Opportunity Policy (EEO) Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, medical condition, marital status, or any other basis prohibited by law. Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance completing our online job application, email application-assistance@redhat.com . General inquiries, such as those regarding the status of a job application, will not receive a reply.

Posted 1 week ago

Apply

3.0 - 5.0 years

10 - 12 Lacs

Chennai, Delhi / NCR, Bengaluru

Work from Office

Responsibilities Work with development teams to ideate software solutions Building and setting up new development tools and infrastructure Working on ways to automate and improve development and release processes Ensuring that systems are safe and secure against cybersecurity threats Deploy updates and fixes Perform root cause analysis for production errors Develop scripts to automate infrastructure provision Working with software developers and software engineers to ensure that development follows established processes and works as intended Requirements At least 2+ years of professional experience as a software developer / DevOps engineer or equivalent Professional experience with Golang Experience with test-driven development and the use of testing frameworks Strong communication skills Technologies we use GitOps GitHub, GitLab, BitBucket Language Golang CI/CD Jenkins, Circle CI, Travis CI, TeamCity, Azure DevOps Containerization Docker, Swarm, Kubernetes Provisioning Terraform CloudOps Azure, AWS, GCP Observability Prometheus, Grafana, GrayLog, ELK Location: Delhi NCR,Bangalore,Chennai,Pune,Kolkata,Ahmedabad,Mumbai,Hyderabad

Posted 1 week ago

Apply

3.0 - 8.0 years

30 - 35 Lacs

Mumbai, Delhi / NCR, Bengaluru

Work from Office

Skill required- Azure Data Factory, Kubernetes, Azure DevOps Must-Have:- Working experience on Azure DevOps (4+ years) Working experience on Kubernetes - scripting, deployment Data Factory Terraform scripting Ansible Powershell Python, Cloud Formation, Good knowledge of ITIL process (good to have) Must have: Strong knowledge of Kubernetes, Istio Service mesh Linux - CLI and Basic knowledge of the OS Scripting (Bash and YAML) Containerization and Docker essentials Jenkins Pipeline creation and execution SCM Management such as GitHub and SVN Cloud Platform Knowledge Azure Monitoring tools like Grafana, Prometheus, ELK stack Certifications (Good to have): 1. Solutions architect associate 2. Certified Kubernetes Administrator (CKA) Location: Remote, Anywhere in- Delhi / NCR,Bangalore/Bengaluru ,Hyderabad/Secunderabad,Chennai,Pune,Kolkata,Ahmedabad,Mumbai

Posted 1 week ago

Apply

1.0 - 3.0 years

8 - 13 Lacs

Pune

Work from Office

Overview We are seeking a DevOps Engineer to join the Critical Start Technologies Private Ltd. team, operating under the Critical Start umbrella, for our India operations. The ideal candidate brings 1–3 years of experience, a strong background in AWS and Terraform, and a passion for infrastructure as code. Candidates should be skilled at writing well-structured Terraform modules, proficient in AWS service provisioning, and familiar with best practices for managing IaaS and PaaS environments. Additional experience with Linux administration, GitHub Actions, container orchestration, and monitoring solutions such as CloudWatch or Prometheus is a plus. Your experience includes writing production code and proficiency in understanding and structuring large projects using Terraform modules. You possess a deep understanding of provisioners and are well-versed in remote state management. We value individuals who are proactive, detail-oriented, and passionate about infrastructure as code. Critical Start is committed to building an inclusive, equitable, and respectful workplace, and we welcome candidates from all backgrounds to apply. Responsibilities As a DevOps Engineer, you will play a key role in maintaining, evolving, and enhancing our existing Terraform-based infrastructure. You'll work across a diverse infrastructure stack to support the delivery of new projects and services to our customers. A core part of your responsibilities will be using Terraform to build modular, maintainable, and scalable infrastructure solutions. You will also take initiative in identifying opportunities to improve performance—focusing on responsiveness, availability, and scalability. Establishing effective monitoring and alerting systems will be essential, as will troubleshooting issues within distributed systems, including throughput, resource utilization, and configuration. Our infrastructure stack includes the following components: Terraform: Used for comprehensive infrastructure management. AWS Fargate: Primary platform for hosting most of our applications and services, along with select EC2 instances for specific use cases. Monitoring and alerts: AWS CloudWatch, SNS, New Relic, and Sentry.io support effective monitoring and timely alerting. Storage and databases: S3, Postgres (RDS), Memcached, RabbitMQ, and AWS Elasticsearch Service handle our storage and data processing needs. Networking and security: VPC, Route 53, IAM, ALB/NLB, Security Groups, and Secrets Manager support a secure and resilient networking environment. CI/CD pipeline: Built using EC2 Image Builder, CodeBuild, and GitHub to streamline software delivery and deployment. Qualifications Required Qualifications: 1-3 years of professional experience in a DevOps, Site Reliability Engineer, or Systems Engineering role. Ability to work through ambiguity and uncertainty. You have a solid understanding of CI/CD pipelines, including their purpose and implementation, and hands-on experience setting them up in real-world environments. You bring experience working with Terraform for provisioning using modular approaches. Strong troubleshooting, problem-solving, and collaborative mindset . You hold a Bachelor's degree from a recognized institution or possess equivalent practical experience that demonstrates your technical capabilities. Preferred Qualifications: Shell scripting experience is a strong plus. Strong knowledge of Linux/Unix systems. Familiarity with source control tools, such as Git. Experience with observability tools such as CloudWatch, New Relic, or Sentry.io Proficiency with Docker and practical experience running containers in AWS environments such as EC2 and Fargate.

Posted 1 week ago

Apply

5.0 - 10.0 years

13 - 15 Lacs

Pune, Chennai, Bengaluru

Work from Office

Grafana specialist to lead the creation of robust dashboards for comprehensive end-to-end monitoring. Strong background in production support monitoring, with a keen understanding of the metrics that matter to both technology teams and management. Required Candidate profile 5y-Build Grafana dashboards for monitoring Use Prometheus&exporters for real-time data Integrate multi-source data&alerts Create Unix/Python scripts for log automation Manage Jira/ServiceNow dashboard

Posted 1 week ago

Apply

2.0 - 7.0 years

13 - 17 Lacs

Chennai

Work from Office

Job Area: Engineering Group, Engineering Group > Software Engineering General Summary: As a leading technology innovator, Qualcomm pushes the boundaries of what's possible to enable next-generation experiences and drives digital transformation to help create a smarter, connected future for all. As a Qualcomm Software Engineer, you will design, develop, create, modify, and validate embedded and cloud edge software, applications, and/or specialized utility programs that launch cutting-edge, world class products that meet and exceed customer needs. Qualcomm Software Engineers collaborate with systems, hardware, architecture, test engineers, and other teams to design system-level software solutions and obtain information on performance requirements and interfaces. Minimum Qualifications: Bachelor's degree in Engineering, Information Systems, Computer Science, or related field and 2+ years of Software Engineering or related work experience. OR Master's degree in Engineering, Information Systems, Computer Science, or related field and 1+ year of Software Engineering or related work experience. OR PhD in Engineering, Information Systems, Computer Science, or related field. 2+ years of academic or work experience with Programming Language such as C, C++, Java, Python, etc. Job Title: MLOps Engineer - ML Platform Hiring Title: Flexible based on candidate experience – about Staff Engineer preferred : We are seeking a highly skilled and experienced MLOps Engineer to join our team and contribute to the development and maintenance of our ML platform both on premises and AWS Cloud. As a MLOps Engineer, you will be responsible for architecting, deploying, and optimizing the ML & Data platform that supports training of Machine Learning Models using NVIDIA DGX clusters and the Kubernetes platform, including technologies like Helm, ArgoCD, Argo Workflow, Prometheus, and Grafana. Your expertise in AWS services such as EKS, EC2, VPC, IAM, S3, and EFS will be crucial in ensuring the smooth operation and scalability of our ML infrastructure. You will work closely with cross-functional teams, including data scientists, software engineers, and infrastructure specialists, to ensure the smooth operation and scalability of our ML infrastructure. Your expertise in MLOps, DevOps, and knowledge of GPU clusters will be vital in enabling efficient training and deployment of ML models. Responsibilities will include: Architect, develop, and maintain the ML platform to support training and inference of ML models. Design and implement scalable and reliable infrastructure solutions for NVIDIA clusters both on premises and AWS Cloud. Collaborate with data scientists and software engineers to define requirements and ensure seamless integration of ML and Data workflows into the platform. Optimize the platform’s performance and scalability, considering factors such as GPU resource utilization, data ingestion, model training, and deployment. Monitor and troubleshoot system performance, identifying and resolving issues to ensure the availability and reliability of the ML platform. Implement and maintain CI/CD pipelines for automated model training, evaluation, and deployment using technologies like ArgoCD and Argo Workflow. Implement and maintain monitoring stack using Prometheus and Grafana to ensure the health and performance of the platform. Manage AWS services including EKS, EC2, VPC, IAM, S3, and EFS to support the platform. Implement logging and monitoring solutions using AWS CloudWatch and other relevant tools. Stay updated with the latest advancements in MLOps, distributed computing, and GPU acceleration technologies, and proactively propose improvements to enhance the ML platform. What are we looking for: Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field. Proven experience as an MLOps Engineer or similar role, with a focus on large-scale ML and/or Data infrastructure and GPU clusters. Strong expertise in configuring and optimizing NVIDIA DGX clusters for deep learning workloads. Proficient in using the Kubernetes platform, including technologies like Helm, ArgoCD, Argo Workflow, Prometheus , and Grafana . Solid programming skills in languages like Python, Go and experience with relevant ML frameworks (e.g., TensorFlow, PyTorch ). In-depth understanding of distributed computing, parallel computing, and GPU acceleration techniques. Familiarity with containerization technologies such as Docker and orchestration tools. Experience with CI/CD pipelines and automation tools for ML workflows (e.g., Jenkins, GitHub, ArgoCD). Experience with AWS services such as EKS , EC2, VPC, IAM, S3, and EFS. Experience with AWS logging and monitoring tools. Strong problem-solving skills and the ability to troubleshoot complex technical issues. Excellent communication and collaboration skills to work effectively within a cross-functional team. We would love to see: Experience with training and deploying models. Knowledge of ML model optimization techniques and memory management on GPUs. Familiarity with ML-specific data storage and retrieval systems. Understanding of security and compliance requirements in ML infrastructure.

Posted 1 week ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies