Get alerts for new jobs matching your selected skills, preferred locations, and experience range. Manage Job Alerts
4.0 - 9.0 years
6 - 11 Lacs
Ahmedabad
Work from Office
Role Senior Databricks Engineer As a Mid Databricks Engineer, you will play a pivotal role in designing, implementing, and optimizing data processing pipelines and analytics solutions on the Databricks platform. You will collaborate closely with cross-functional teams to understand business requirements, architect scalable solutions, and ensure the reliability and performance of our data infrastructure. This role requires deep expertise in Databricks, strong programming skills, and a passion for solving complex engineering challenges. What you'll do : - Design and develop data processing pipelines and analytics solutions using Databricks. - Architect scalable and efficient data models and storage solutions on the Databricks platform. - Collaborate with architects and other teams to migrate current solution to use Databricks. - Optimize performance and reliability of Databricks clusters and jobs to meet SLAs and business requirements. - Use best practices for data governance, security, and compliance on the Databricks platform. - Mentor junior engineers and provide technical guidance. - Stay current with emerging technologies and trends in data engineering and analytics to drive continuous improvement. You'll be expected to have : - Bachelor's or master's degree in computer science, Engineering, or a related field. - 5 to 8 years of overall experience and 2+ years of experience designing and implementing data solutions on the Databricks platform. - Proficiency in programming languages such as Python, Scala, or SQL. - Strong understanding of distributed computing principles and experience with big data technologies such as Apache Spark. - Experience with cloud platforms such as AWS, Azure, or GCP, and their associated data services. - Proven track record of delivering scalable and reliable data solutions in a fast-paced environment. - Excellent problem-solving skills and attention to detail. - Strong communication and collaboration skills with the ability to work effectively in cross-functional teams. - Good to have experience with containerization technologies such as Docker and Kubernetes. - Knowledge of DevOps practices for automated deployment and monitoring of data pipelines.
Posted 1 week ago
5.0 - 10.0 years
10 - 14 Lacs
Hyderabad
Work from Office
Overview DataOps L3 The role will leverage & enhance existing technologies in the area of data and analytics solutions like Power BI, Azure data engineering technologies, ADLS, ADB, Synapse, and other Azure services. The role will be responsible for developing and support IT products and solutions using these technologies and deploy them for business users Responsibilities 5 to 10 Years of IT & Azure Data engineering technologies experience Prior experience in ETL, data pipelines, data flow techniques using Azure Data Services Working experience in Python, Py Spark, Azure Data Factory, Azure Data Lake Gen2, Databricks, Azure Synapse and file formats like JSON & Parquet. Experience in creating ADF Pipelines to source and process data sets. Experience in creating Databricks notebooks to cleanse, transform and enrich data sets. Development experience in orchestration of pipelines Good understanding about SQL, Databases, Datawarehouse systems preferably Teradata Experience in deployment and monitoring techniques. Working experience with Azure DevOps CI/CD pipelines to deploy Azure resources. Experience in handling operations/Integration with source repository Must have good knowledge on Datawarehouse concepts and Datawarehouse modelling. Working knowledge of SNOW including resolving incidents, handling Change requests /Service requests, reporting on metrics to provide insights. Collaborate with the project team to understand tasks to model tables using data warehouse best practices and develop data pipelines to ensure the efficient delivery of data. Strong expertise in performance tuning and optimization of data processing systems. Proficient in Azure Data Factory, Azure Databricks, Azure SQL Database, and other Azure data services. Develop and enforce best practices for data management, including data governance and security. Work closely with cross-functional teams to understand data requirements and deliver solutions that meet business needs. Proficient in implementing DataOps framework. Qualifications Azure data factory Azure Databricks Azure Synapse PySpark/SQL ADLS Azure DevOps with CI/CD implementation. Nice-to-Have Skill Sets Business Intelligence tools (preferredPower BI) DP-203 Certified.
Posted 1 week ago
6.0 - 11.0 years
25 - 27 Lacs
Hyderabad
Work from Office
Overview We are seeking a highly skilled and experienced Azure Data Engineer to join our dynamic team. In this critical role, you will be responsible for designing, developing, and maintaining robust and scalable data solutions on the Microsoft Azure platform. You will work closely with data scientists, analysts, and business stakeholders to translate business requirements into effective data pipelines and data models. Responsibilities Design, develop, and implement data pipelines and ETL/ELT processes using Azure Data Factory, Azure Databricks, and other relevant Azure services. Develop and maintain data lakes and data warehouses on Azure, including Azure Data Lake Storage Gen2 and Azure Synapse Analytics. Build and optimize data models for data warehousing, data marts, and data lakes. Develop and implement data quality checks and data governance processes. Troubleshoot and resolve data-related issues. Collaborate with data scientists and analysts to support data exploration and analysis. Stay current with the latest advancements in cloud computing and data engineering technologies. Participate in all phases of the software development lifecycle, from requirements gathering to deployment and maintenance Qualifications 6+ years of experience in data engineering, with at least 3 years of experience working with Azure data services. Strong proficiency in SQL, Python, and other relevant programming languages. Experience with data warehousing and data lake architectures. Experience with ETL/ELT tools and technologies, such as Azure Data Factory, Azure Databricks, and Apache Spark. Experience with data modeling and data warehousing concepts. Experience with data quality and data governance best practices. Strong analytical and problem-solving skills. Excellent communication and collaboration skills. Experience with Agile development methodologies. Bachelor's degree in Computer Science, Engineering, or a related field (Master's degree preferred). Relevant Azure certifications (e.g., Azure Data Engineer Associate) are a plus
Posted 1 week ago
5.0 - 10.0 years
19 - 25 Lacs
Hyderabad
Work from Office
Overview Seeking an Associate Manager, Data Operations, to support our growing data organization. In this role, you will assist in maintaining data pipelines and corresponding platforms (on-prem and cloud) while working closely with global teams on DataOps initiatives. Support the day-to-day operations of data pipelines, ensuring data governance, reliability, and performance optimization on Microsoft Azure. Hands-on experience with Azure Data Factory (ADF), Azure Synapse Analytics, Azure Databricks, and real-time streaming architectures is preferred. Assist in ensuring the availability, scalability, automation, and governance of enterprise data pipelines supporting analytics, AI/ML, and business intelligence. Contribute to DataOps programs, aligning with business objectives, data governance standards, and enterprise data strategy. Help implement real-time data observability, monitoring, and automation frameworks to improve data reliability, quality, and operational efficiency. Support the development of governance models and execution roadmaps to enhance efficiency across Azure, AWS, GCP, and on-prem environments. Work on CI/CD integration, data pipeline automation, and self-healing capabilities to improve enterprise-wide DataOps processes. Collaborate with cross-functional teams to support and maintain next-generation Data & Analytics platforms while promoting an agile and high-performing DataOps culture. Assist in the adoption of Data & Analytics technology transformations, ensuring automation for proactive issue identification and resolution. Partner with cross-functional teams to support process improvements, best practices, and operational efficiencies within DataOps. Responsibilities Assist in the implementation and optimization of enterprise-scale data pipelines using Azure Data Factory (ADF), Azure Synapse Analytics, Azure Databricks, and Azure Stream Analytics. Support data ingestion, transformation, orchestration, and storage workflows, ensuring data reliability, integrity, and availability. Help ensure seamless batch, real-time, and streaming data processing, focusing on high availability and fault tolerance. Contribute to DataOps automation efforts, including CI/CD for data pipelines, automated testing, and version control using Azure DevOps and Terraform. Collaborate with Data Engineering, Analytics, AI/ML, CloudOps, and Business Intelligence teams to support data-driven decision-making. Assist in aligning DataOps practices with regulatory and security requirements by working with IT, data stewards, and compliance teams. Support data operations and sustainment activities, including testing and monitoring processes for global products and projects. Participate in data capture, storage, integration, governance, and analytics efforts, working alongside cross-functional teams. Assist in managing day-to-day DataOps activities, ensuring adherence to service-level agreements (SLAs) and business requirements. Engage with SMEs and business stakeholders to ensure data platform capabilities align with business needs. Contribute to Agile work intake and execution processes, helping to maintain efficiency in data platform teams. Help troubleshoot and resolve issues related to cloud infrastructure and data services in collaboration with technical teams. Support the development and automation of operational policies and procedures, improving efficiency and resilience. Assist in incident response and root cause analysis, contributing to self-healing mechanisms and mitigation strategies. Foster a customer-centric approach, advocating for operational excellence and continuous improvement in service delivery. Help build a collaborative, high-performing team culture, promoting automation and efficiency within DataOps. Adapt to shifting priorities and support cross-functional teams in maintaining productivity and achieving business goals. Utilize technical expertise in cloud and data operations to support service reliability and scalability. Qualifications 5+ years of technology work experience in a large-scale global organization, with CPG industry experience preferred. 5+ years of experience in Data & Analytics roles, with hands-on expertise in data operations and governance. 2+ years of experience working within a cross-functional IT organization, collaborating with multiple teams. Experience in a lead or senior support role, with a focus on DataOps execution and delivery. Strong communication skills, with the ability to collaborate with stakeholders and articulate technical concepts to non-technical audiences. Analytical and problem-solving abilities, with a focus on prioritizing customer needs and operational improvements. Customer-focused mindset, ensuring high-quality service delivery and operational efficiency. Growth mindset, with a willingness to learn and adapt to new technologies and methodologies in a fast-paced environment. Experience supporting data operations in a Microsoft Azure environment, including data pipeline automation. Familiarity with Site Reliability Engineering (SRE) principles, such as monitoring, automated issue remediation, and scalability improvements. Understanding of operational excellence in complex, high-availability data environments. Ability to collaborate across teams, building strong relationships with business and IT stakeholders. Basic understanding of data management concepts, including master data management, data governance, and analytics. Knowledge of data acquisition, data catalogs, data standards, and data management tools. Strong execution and organizational skills, with the ability to follow through on operational plans and drive measurable results. Adaptability in a dynamic, fast-paced environment, with the ability to shift priorities while maintaining productivity.
Posted 1 week ago
2.0 - 4.0 years
8 - 12 Lacs
Hyderabad
Work from Office
Overview We are seeking a skilled and proactive business analyst with expertise in Azure Data Engineering to join our dynamic team. In this role, you will bridge the gap between business needs and technical solutions, leveraging your analytical skills and Azure platform knowledge to design and implement robust data solutions. You will collaborate closely with stakeholders to gather and translate requirements, develop data pipelines, and ensure data quality and governance. This position requires a strong understanding of Azure services, data modeling, and ETL processes, along with the ability to thrive in a fast-paced, evolving environment. Responsibilities Collaborate with stakeholders to understand business needs and translate them into technical requirements. Design, develop, and implement data solutions using Azure Data Engineering technologies. Analyze complex data sets to identify trends, patterns, and insights that drive business decisions. Create and maintain detailed documentation of business requirements, data models, and data flows. Work in an environment where requirements are not always clearly defined, demonstrating flexibility and adaptability. Conduct data quality assessments and implement data governance practices. Provide training and support to end-users on data tools and solutions. Continuously monitor and optimize data processes for efficiency and performance. Qualifications Minimum of 2-4 years of experience as a data analyst with hands-on experience in Azure Data Engineering. Proficiency in Azure Data Factory, Azure Databricks, Azure SQL Database, and other Azure data services. Strong analytical and problem-solving skills with the ability to work in a fast-paced, ambiguous environment. Excellent communication and interpersonal skills to effectively collaborate with cross-functional teams. Experience with data modeling, ETL processes, and data warehousing. Knowledge of data governance and data quality best practices. Ability to manage multiple projects and priorities simultaneously. Preferred Skills: Experience with other cloud platforms and data engineering tools. Certification in Azure Data Engineering or related fields.
Posted 1 week ago
10.0 - 15.0 years
4 - 8 Lacs
Noida
Work from Office
Highly skilled and experienced Data Modeler to join Enterprise Data Modelling team. The candidate will be responsible for creating and maintaining conceptual logical and physical data models ensuring alignment with industry best practices and standards. Working closely with business and functional teams the Data Modeler will play a pivotal role in standardizing data models at portfolio and domain levels driving efficiencies and maximizing the value of clients data assets. Preference will be given to candidates with prior experience within an Enterprise Data Modeling team. The ideal domain experience would be Insurance or Investment Banking. Roles and Responsibilities: Develop comprehensive conceptual logical and physical data models for multiple domains within the organization leveraging industry best practices and standards. Collaborate with business and functional teams to understand their data requirements and translate them into effective data models that support their strategic objectives. Serve as a subject matter expert in data modeling tools such as ERwin Data Modeler providing guidance and support to other team members and stakeholders. Establish and maintain standardized data models across portfolios and domains ensuring consistency governance and alignment with organizational objectives. Identify opportunities to optimize existing data models and enhance data utilization particularly in critical areas such as fraud banking AML. Provide consulting services to internal groups on data modeling tool usage administration and issue resolution promoting seamless data flow and application connections. Develop and deliver training content and support materials for data models ensuring that stakeholders have the necessary resources to understand and utilize them effectively. Collaborate with the enterprise data modeling group to develop and implement a robust governance framework and metrics for model standardization with a focus on longterm automated monitoring solutions. Qualifications: Bachelors or masters degree in computer science Information Systems or a related field. 10 years of experience working as a Data Modeler or in a similar role preferably within a large enterprise environment. Expertise in data modeling concepts and methodologies with demonstrated proficiency in creating conceptual logical and physical data models. Handson experience with data modeling tools such as Erwin Data Modeler as well as proficiency in database environments such as Snowflake and Netezza. Strong analytical and problemsolving skills with the ability to understand complex data requirements and translate them into effective data models. Excellent communication and collaboration skills with the ability to work effectively with crossfunctional teams and stakeholders. problem-solving skills,business intelligence platforms,erwin,data modeling,database management systems,data warehousing,etl processes,big data technologies,agile methodologies,data governance,sql,enterprise data modelling,data visualization tools,cloud data services,analytical skills,data modelling tool,,data architecture,communication skills
Posted 1 week ago
0.0 - 5.0 years
1 - 5 Lacs
Bengaluru
Work from Office
Job Title:Data Engineer - DBT (Data Build Tool)Experience0-5 YearsLocation:Bengaluru : Job Responsibilities Assist in the design and implementation of Snowflake-based analytics solution(data lake and data warehouse) on AWSRequirements definition, source data analysis and profiling, the logical and physical design of the data lake and datawarehouse as well as the design of data integration and publication pipelines Develop Snowflake deployment and usage best practices Help educate the rest of the team members on the capabilities and limitations of Snowflake Build and maintain data pipelines adhering to suggested enterprise architecture principles and guidelines Design, build, test, and maintain data management systemsWork in sync with internal and external team members like data architects, data scientists, data analysts to handle all sorts of technical issue Act as technical leader within the team Working in Agile/Lean model Deliver quality deliverables on time Translating complex functional requirements into technical solutions. EXPERTISE AND QUALIFICATIONSEssential Skills, Education and Experience Should have a B.E. / B.Tech. / MCA or equivalent degree along with 4-7 years of experience in Data Engineering Strong experience in DBT concepts like Model building and configurations, incremental load strategies, macro, DBT tests. Strong experience in SQL Strong Experience in AWS Creation and maintenance of optimum data pipeline architecture for ingestion, processing of data Creation of necessary infrastructure for ETL jobs from a wide range of data sources using Talend, DBT, S3, Snowflake. Experience in Data storage technologies like Amazon S3, SQL, NoSQL Data modeling technical awareness Experience in working with stakeholders working in different time zones Good to have AWS data services development experience. Working knowledge on using Bigdata technologies. Experience in collaborating data quality and data governance team. Exposure to reporting tools like Tableau Apache Airflow, Apache Kafka (nice to have) Payments domain knowledge CRM, Accounting, etc. in depth understanding Regulatory reporting exposureOther skills Good Communication skills Team Player Problem solver Willing to learn new technologies, share your ideas and assist other team members as neededStrong analytical and problem-solving skills; ability to define problems, collect data, establish facts, and drawconclusions.
Posted 1 week ago
3.0 - 6.0 years
13 - 18 Lacs
Bengaluru
Work from Office
We are looking to hire Data engineer for the Platform Engineering team. It is a collection of highly skilled individuals ranging from development to operations with a security first mindset who strive to push the boundaries of technology. We champion a DevSecOps culture and raise the bar on how and when we deploy applications to production. Our core principals are centered around automation, testing, quality, and immutability all via code. The role is responsible for building self-service capabilities that improve our security posture, productivity, and reduce time to market with automation at the core of these objectives. The individual collaborates with teams across the organization to ensure applications are designed for Continuous Delivery (CD) and are well-architected for their targeted platform which can be on-premise or the cloud. If you are passionate about developer productivity, cloud native applications, and container orchestration, this job is for you! Principal Accountabilities: The incumbent is mentored by senior individuals on the team to capture the flow and bottlenecks in the holistic IT delivery process and define future tool sets Skills and Software Requirements: Experience with a language such as Python, Go,SQL, Java, or Scala GCP data services (BigQuery; Dataflow; Dataproc; Cloud Composer; Pub/Sub; Google Cloud Storage; IAM) Experience with Jenkins, Maven, Git, Ansible, or CHEF Experience working with containers, orchestration tools (like Kubernetes, Mesos, Docker Swarm etc.) and container registries (GCE, Docker hub etc.) Experience with [SPI]aaS- Software-as-a-Service, Platform-as-a-Service, or Infrastructure-as- a-Service Acquire, cleanse, and ingest structured and unstructured data on the cloud Combine data from disparate sources in to a single, unified, authoritative view of data (e.g., Data Lake) Enable and support data movement from one system service to another system service Experience implementing or supporting automated solutions to technical problems Experience working in a team environment, proactively executing on tasks while meeting agreed delivery timelines Ability to contribute to effective and timely solutions Excellent oral and written communication skills
Posted 1 week ago
5.0 - 9.0 years
0 Lacs
maharashtra
On-site
As a Database Administrator at NTT DATA, you will be a seasoned subject matter expert responsible for ensuring the availability, integrity, and performance of critical data assets. You will work closely with cross-functional teams to support data-driven applications, troubleshoot issues, and implement robust backup and recovery strategies. Collaboration with Change Control, Release Management, Asset and Configuration Management, and Capacity and Availability Management will be essential to meet the needs of users and ensure database security and integrity. Key responsibilities include performing installation, configuration, and maintenance of database management systems, collaborating with software developers/architects to optimize database-related applications, designing backup and disaster recovery strategies, monitoring database performance, and providing technical support to end-users. You will also participate in database software upgrades, data validation activities, and work collaboratively with cross-functional teams to support database-related initiatives. To excel in this role, you should have seasoned proficiency in database administration tasks, a strong understanding of SQL, database security principles, and backup strategies. Effective communication, problem-solving, and analytical skills are crucial, along with the ability to manage multiple projects concurrently while maintaining attention to detail. Academic qualifications in computer science or related fields, along with relevant certifications like MCSE DBA or Oracle Certified Professional, are preferred. NTT DATA is a trusted global innovator of business and technology services, committed to helping clients innovate, optimize, and transform for long-term success. With a diverse workforce and a focus on R&D, NTT DATA is dedicated to moving organizations confidently into the digital future. As an Equal Opportunity Employer, NTT DATA offers a dynamic workplace where employees can thrive, grow, and make a difference.,
Posted 2 weeks ago
8.0 - 12.0 years
0 Lacs
karnataka
On-site
You should have a solid working knowledge of AWS database and data services as well as the Power BI stack. Your experience in gathering requirements, modeling data, designing, and supporting high-performance big data backend and data visualization systems will be crucial. You should be adept at utilizing methodologies and platform stacks such as Map Reduce, Spark, streaming solutions like Kafka and Kinesis, ETL systems like Glue and Firehose, storage solutions like S3, warehouse stacks like Redshift and DynamoDB, and equivalent open source stacks. Designing and implementing solutions using visualization technologies like Power BI and Quick Sight should be within your expertise. You will be responsible for maintaining and continuously grooming the product backlog, the release pipeline, and the product roadmap. It will be your responsibility to capture problem statements and opportunities raised by customers as demand items, epics, and stories. Leading database physical design sessions with the engineers in the team and ensuring quality assurance and load testing of the solution to maintain customer experience are also part of the role. Additionally, you will be supporting data governance and data quality (cleansing) efforts. Your primary skills should include proficiency in AWS database, data services, PowerBi stack, and big data.,
Posted 2 weeks ago
6.0 - 7.0 years
6 - 11 Lacs
Noida
Work from Office
Responsibilities Data Architecture: Develop and maintain the overall data architecture, ensuring scalability, performance, and data quality. AWS Data Services: Expertise in using AWS data services such as AWS Glue, S3, SNS, SES, Dynamo DB, Redshift, Cloud formation, Cloud watch, IAM, DMS, Event bridge scheduler etc. Data Warehousing: Design and implement data warehouses on AWS, leveraging AWS Redshift or other suitable options. Data Lakes: Build and manage data lakes on AWS using AWS S3 and other relevant services. Data Pipelines: Design and develop efficient data pipelines to extract, transform, and load data from various sources. Data Quality: Implement data quality frameworks and best practices to ensure data accuracy, completeness, and consistency. Cloud Optimization: Optimize data engineering solutions for performance, cost-efficiency, and scalability on the AWS cloud. Team Leadership: Mentor and guide data engineers, ensuring they adhere to best practices and meet project deadlines. Qualifications Bachelors degree in computer science, Engineering, or a related field. 6-7 years of experience in data engineering roles, with a focus on AWS cloud platforms. Strong understanding of data warehousing and data lake concepts. Proficiency in SQL and at least one programming language ( Python/Pyspark ). Good to have - Experience with any big data technologies like Hadoop, Spark, and Kafka. Knowledge of data modeling and data quality best practices. Excellent problem-solving, analytical, and communication skills. Ability to work independently and as part of a team. Preferred Qualifications Certifications in AWS Certified Data Analytics - Specialty or AWS Certified Solutions Architect - Data. Mandatory Competencies Big Data - Big Data - Pyspark Data on Cloud - Azure Data Lake (ADL) Beh - Communication and collaboration Cloud - AWS - AWS Lambda,AWS EventBridge, AWS Fargate Cloud - AWS - AWS S3, S3 glacier, AWS EBS Cloud - Azure - Azure Data Factory (ADF), Azure Databricks, Azure Data Lake Storage, Event Hubs, HDInsight Cloud - AWS - Tensorflow on AWS, AWS Glue, AWS EMR, Amazon Data Pipeline, AWS Redshift Database - Sql Server - SQL Packages Data Science and Machine Learning - Data Science and Machine Learning - Python
Posted 2 weeks ago
1.0 - 4.0 years
12 - 17 Lacs
Mumbai, Nagpur, Thane
Work from Office
Xpetize Technology Solutions Private Limited is looking for Field Sales Executive to join our dynamic team and embark on a rewarding career journey Promote and sell products or services to customers in the field Meet or exceed sales targets by developing and implementing sales strategies Build and maintain relationships with customers to increase product awareness and loyalty Conduct product presentations and provide product information to customers Provide excellent customer service, including addressing customer inquiries and resolving customer issues Monitor market trends and competitor activities to identify business opportunities and threats Maintain accurate sales records and report sales activity to management Strong knowledge of sales techniques, product knowledge, and customer service best practices Excellent communication and interpersonal skills
Posted 2 weeks ago
1.0 - 5.0 years
3 - 7 Lacs
Ahmedabad
Work from Office
Develop, record, and maintain cutting-edge web-based PHP applications for the company. Build innovative, state-of-the-art applications and collaborate with the User Experience (UX) team Ensure HTML, CSS, and shared JavaScript is valid and consistent across applications Prepare and maintain all applications utilizing standard development tools Utilize backend data services and contribute to increasing existing data services API Lead the entire web application development life cycle right from concept stage to delivery and post-launch support Convey effectively with all task progress, evaluations, suggestions, schedules along with technical and process issues Document the development process, architecture, and standard components Coordinate with co-developers and keeps project manager well informed of the status of development effort and serves as a liaison between development staff and project manager.
Posted 2 weeks ago
4.0 - 10.0 years
4 - 8 Lacs
Bengaluru
Work from Office
We are looking for a skilled SAP ABAP Developer with a strong programming background and hands-on experience in modern SAP technologies. The ideal candidate should have solid technical expertise across ABAP and related frameworks, along with a strong educational foundation. Design, develop, and implement SAP applications using ABAP/ABAP OO Develop SAP Core Data Services (CDS), OData services, and Fiori/UI5 applications Work with technologies such as BOPF, RAP, and HANA Integrate SOAP APIs and work with frontend scripting using JavaScript Collaborate with functional teams to translate business requirements into technical solutions Required Skills: 4 to 10 years of experience in software development Strong educational background: Bachelor s degree in Engineering or MCA from reputed institutes Expertise in: ABAP/ABAP OO CDS Views / OData Services Fiori / UI5 BOPF / HANA / RAP SOAP API / JavaScript Sap Abap, Migration, Implemenataion
Posted 2 weeks ago
7.0 - 12.0 years
7 - 11 Lacs
Kolkata
Work from Office
Your day in the role will include.. Assists clients in the selection, implementation, and support of Data Services for SAP. Lead multiple sized projects as team member or lead to implement new functionalities and improve existing functionalities including articulating, analyzing requirements and translating them into effective solutions Prepare and conduct Unit Testing and User Acceptance Testing You will come with.. Knowledge and experience in implementation planning, fit analysis, configuration, testing, rollout and post-implementation support. Experience in working with teams to prioritize work and drive system solutions by clearly articulating business needs Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise Overall, 7-12 years of relevant experience in SAP BODS/BOIS/SDI/SDQ and 3+ Years of SAP functional experience specializing in design and configuration of SAP BODS/HANA SDI modules. At least 3 years of hands-on experience in Syniti ADM, apart from other Data Migration exposure Syniti ADM Consultant, primarily have Data migration experience from Different Legacy Systems to SAP or Non-SAP systems. Data Migration experience from SAP ECC or Non SAP to SAP S4 HANA or other SAP Target Systems. Should be able to prepare mapping sheet combining his/her Functional and technical expertise Preferred technical and professional experience Having worked or strong Knowledge of SAP DATA HUB. Experience/Strong knowledge of HANA SDI (Smart data Integration) to use this as an ETL and should be able to develop flow graphs to Validate/Transform data. Consultant should Develop Workflows, Data flows based on the specifications using various stages in BODS
Posted 2 weeks ago
4.0 - 7.0 years
8 - 9 Lacs
Hyderabad
Work from Office
Req ID: 332502 NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now. We are currently seeking a Senior SAP FICO S4 Hana Consultant to join our team in Hyderabad, Telangana (IN-TG), India (IN). Senior SAP S/4 HANA FICO Consultant At NTT DATA Services, we know that with the right people on board, anything is possible. The quality, integrity, and commitment of our employees are key factors in our company s growth, market presence and our ability to help our clients stay a step ahead of the competition. By hiring the best people and helping them grow both professionally and personally, we ensure a bright future for NTT DATA Services and for the people who work here. NTT DATA Services currently seeks a senior SAP S/4 HANA FICO Consultant to join our team. SAP S/4 HANA FICO Consultant: An SAP S/4HANA FICO (Financial Accounting and Controlling) consultant is responsible for implementing, configuring, and supporting SAP S/4HANA Finance solutions . This involves understanding business requirements, designing solutions, performing configurations, testing, and providing user support. They also play a key role in integrating FICO with other SAP modules and external systems. Gathering and Analyzing Requirements: Working with business users to understand their needs and translating them into functional specifications for SAP S/4HANA FICO. Solution Design and Configuration: Designing and configuring the SAP S/4HANA FICO module to meet business requirements, including setting up company codes, chart of accounts, fiscal year variants, and other relevant configurations. Engage with customers in executive meetings, solution workshops, and design sessions and clearly articulate the business value of SAP solutions through tailored presentations and demos. Identify and describe solution options, evaluate pros/cons, and make recommendations regarding best solution strategies and estimated opportunity timelines. Create business requirements documents, process maps, high-level solution design diagrams, and scope of works for various business implementation scenarios and solution designs. Shape conversations with customers by guiding deep cross-functional Discovery based on industry best practices, customer research, and SAP customer experiences across SAP solutions. Stay updated on SAP product roadmaps and trends, leveraging this knowledge to influence presales strategies. Deliver customized demos, workshops, and presentations to illustrate the value of SAP s cloud portfolio. Collaborate with SAP s technical and implementation teams to ensure a seamless transition from presales to delivery. Monitor market and competitive trends in the SAP cloud landscape to keep the presales strategy relevant Training other members of the pre-sales team on the technical aspects of the service offerings and products. Positions General Duties and Tasks: 10+ years of experience in SAP S/4 HANA FICO implementing , solution architecture, SAP consulting and solution architecture experience Strong knowledge of SAP ERP, S/4HANA, and other SAP Finance . Deep understanding of SAP S/4HANA Finance concepts, including General Ledger (GL), Accounts Payable (AP), Accounts Receivable (AR), Asset Accounting (AA), Controlling (CO), and Profitability Analysis (PA). Excellent communication, presentation, and client relationship skills. Ability to translate technical details into business language and value-driven solutions. Experience in industries like manufacturing or Life Sciences. Bachelor s degree in Computer Science, Engineering, Business, or related field. Location: Hyderabad, Bangalore
Posted 2 weeks ago
3.0 - 8.0 years
5 - 9 Lacs
Hyderabad
Work from Office
Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : SAP ABAP Development for HANA Good to have skills : NAMinimum 3 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Developer, you will design, build, and configure applications to meet business process and application requirements. A typical day involves collaborating with team members to understand project needs, developing application features, and ensuring that the solutions align with business objectives. You will also engage in testing and troubleshooting to enhance application performance and user experience, while continuously seeking opportunities for improvement and innovation in application development processes. Roles & Responsibilities:- Expected to perform independently and become an SME.- Required active participation/contribution in team discussions.- Contribute in providing solutions to work related problems.- Assist in the documentation of application specifications and user guides.- Engage in code reviews to ensure quality and adherence to best practices. Professional & Technical Skills: - Must Have Skills: Proficiency in SAP ABAP Development for HANA.- Strong understanding of application design principles and methodologies.- Experience with database management and optimization techniques.- Familiarity with integration processes and tools within SAP environments.- Ability to troubleshoot and resolve application issues efficiently.- Aware of Clean Clore principles (In App-, on stack-development, side-by-side), Clean core extension concepts- ABAP cloud, ABAP RESTful Application Programming Model (RAP)- FIORI development/extensions, FIORI Elements, RAP UI - ABAP Development Tools (ADT):Development using Eclipse and ADT- Debugging (ABAP Development Tools - ADT) and Unit Testing knowledge (ABAP unit testing framework)- GIT, ABAP Git knowledge, - SAP BTP & ABAP Environment:BTP, development and deployment knowledge in BTP, Software Components, Packages, Transport Requests knowledge in BTP, BAS (Business Application Studio) knowledge, BPA (Build process Automation) knowledge - Core Data Services (CDS) knowledge- Behavior Definitions & Implementations:Definition and Implementation BO behavior (CRUD) and also use of these.- OData V4 & Service Exposure:Service Definition and Service Binding (create and expose service)- ADOBE Forms development knowledge (create change existing WF)- BPA (Build process Automation), Classical, flexible WF knowledge (create, change forms) Additional Information:- The candidate should have minimum 3 years of experience in SAP ABAP Development for HANA.- This position is based at our Hyderabad office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 weeks ago
15.0 - 25.0 years
13 - 18 Lacs
Kolkata
Work from Office
Project Role : Application Architect Project Role Description : Provide functional and/or technical expertise to plan, analyze, define and support the delivery of future functional and technical capabilities for an application or group of applications. Assist in facilitating impact assessment efforts and in producing and reviewing estimates for client work requests. Must have skills : Customer Data Platform & Integration, Google Cloud Data Services Good to have skills : NAMinimum 15 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Architect, you will provide functional and/or technical expertise to plan, analyze, define, and support the delivery of future functional and technical capabilities for an application or group of applications. You will also assist in facilitating impact assessment efforts and in producing and reviewing estimates for client work requests. Roles & Responsibilities:- Expected to be a SME with deep knowledge and experience.- Should have Influencing and Advisory skills.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Expected to provide solutions to problems that apply across multiple teams.- Lead the design and implementation of complex application solutions.- Collaborate with stakeholders to understand business requirements and translate them into technical solutions.- Provide guidance and expertise on application architecture and integration strategies. Professional & Technical Skills: - Must To Have Skills: Proficiency in Customer Data Platform & Integration, Google Cloud Data Services.- Strong understanding of data integration and data management principles.- Experience in designing and implementing scalable and secure data platforms.- Knowledge of cloud-based technologies and services.- Hands-on experience with data modeling and database design. Additional Information:- The candidate should have a minimum of 15 years of experience in Customer Data Platform & Integration.- This position is based at our Kolkata office.- A 15 years full-time education is required. Qualification 15 years full time education
Posted 2 weeks ago
15.0 - 20.0 years
10 - 14 Lacs
Bengaluru
Work from Office
Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : SAP BusinessObjects Data Services Good to have skills : NAMinimum 5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Lead, you will lead the effort to design, build, and configure applications, acting as the primary point of contact. Your typical day will involve collaborating with various teams to ensure project milestones are met, facilitating discussions to address challenges, and guiding your team through the development process while maintaining a focus on quality and efficiency. You will also engage in strategic planning sessions to align project goals with organizational objectives, ensuring that all stakeholders are informed and involved in the decision-making process. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Facilitate training and development opportunities for team members to enhance their skills.- Monitor project progress and implement necessary adjustments to ensure timely delivery. Professional & Technical Skills: - Must To Have Skills: Proficiency in SAP BusinessObjects Data Services.- Strong understanding of data integration and transformation processes.- Experience with ETL (Extract, Transform, Load) methodologies.- Familiarity with database management systems and SQL.- Ability to troubleshoot and resolve technical issues efficiently. Additional Information:- The candidate should have minimum 5 years of experience in SAP BusinessObjects Data Services.- This position is based at our Bengaluru office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 weeks ago
3.0 - 8.0 years
5 - 9 Lacs
Hyderabad
Work from Office
Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : SAP ABAP Development for HANA Good to have skills : NAMinimum 3 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Developer, you will engage in the design, construction, and configuration of applications tailored to fulfill specific business processes and application requirements. Your typical day will involve collaborating with team members to understand project needs, developing innovative solutions, and ensuring that applications are optimized for performance and usability. You will also participate in testing and debugging processes to ensure the highest quality of deliverables, while continuously seeking opportunities for improvement and efficiency in application development. Roles & Responsibilities:- Expected to perform independently and become an SME.- Required active participation/contribution in team discussions.- Contribute in providing solutions to work related problems.- Assist in the documentation of application specifications and user guides.- Engage in code reviews to ensure adherence to best practices and standards. Professional & Technical Skills: - Must To Have Skills: Proficiency in SAP ABAP Development for HANA.- Strong understanding of object-oriented programming principles.- Experience with database management and SQL.- Familiarity with SAP modules and integration techniques.- Ability to troubleshoot and resolve application issues efficiently.- Aware of Clean Clore principles (In App-, on stack-development, side-by-side), Clean core extension concepts- ABAP cloud, ABAP RESTful Application Programming Model (RAP)- FIORI development/extensions, FIORI Elements, RAP UI - ABAP Development Tools (ADT):Development using Eclipse and ADT- Debugging (ABAP Development Tools - ADT) and Unit Testing knowledge (ABAP unit testing framework)- GIT, ABAP Git knowledge, - SAP BTP & ABAP Environment:BTP, development and deployment knowledge in BTP, Software Components, Packages, Transport Requests knowledge in BTP, BAS (Business Application Studio) knowledge, BPA (Build process Automation) knowledge - Core Data Services (CDS) knowledge- Behavior Definitions & Implementations:Definition and Implementation BO behavior (CRUD) and also use of these.- OData V4 & Service Exposure:Service Definition and Service Binding (create and expose service)- ADOBE Forms development knowledge (create change existing WF)- BPA (Build process Automation), Classical, flexible WF knowledge (create, change forms) Additional Information:- The candidate should have minimum 3 years of experience in SAP ABAP Development for HANA.- This position is based at our Hyderabad office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 weeks ago
4.0 - 5.0 years
5 - 9 Lacs
Noida
Work from Office
Responsibilities Data Architecture: Develop and maintain the overall data architecture, ensuring scalability, performance, and data quality. AWS Data Services: Expertise in using AWS data services such as AWS Glue, S3, SNS, SES, Dynamo DB, Redshift, Cloud formation, Cloud watch, IAM, DMS, Event bridge scheduler etc. Data Warehousing: Design and implement data warehouses on AWS, leveraging AWS Redshift or other suitable options. Data Lakes: Build and manage data lakes on AWS using AWS S3 and other relevant services. Data Pipelines: Design and develop efficient data pipelines to extract, transform, and load data from various sources. Data Quality: Implement data quality frameworks and best practices to ensure data accuracy, completeness, and consistency. Cloud Optimization: Optimize data engineering solutions for performance, cost-efficiency, and scalability on the AWS cloud. Qualifications Bachelors degree in computer science, Engineering, or a related field. 4-5 years of experience in data engineering roles, with a focus on AWS cloud platforms. Strong understanding of data warehousing and data lake concepts. Proficiency in SQL and at least one programming language ( Python/Pyspark ). Good to have - Experience with any big data technologies like Hadoop, Spark, and Kafka. Knowledge of data modeling and data quality best practices. Excellent problem-solving, analytical, and communication skills. Ability to work independently and as part of a team. Preferred Qualifications Certifications in AWS Certified Data Analytics - Specialty or AWS Certified Solutions Architect - Data. Mandatory Competencies Big Data - Big Data - Pyspark Beh - Communication and collaboration Cloud - AWS - AWS Lambda,AWS EventBridge, AWS Fargate Database - Sql Server - SQL Packages Data Science and Machine Learning - Data Science and Machine Learning - Python
Posted 2 weeks ago
1.0 - 3.0 years
2 - 6 Lacs
Bengaluru
Work from Office
Skill required: Clinical Data Services - Clinical Reports Management Designation: Clinical Data Svs Associate Qualifications: BE/BTech Years of Experience: 1 to 3 years About Accenture Combining unmatched experience and specialized skills across more than 40 industries, we offer Strategy and Consulting, Technology and Operations services, and Accenture Song all powered by the worlds largest network of Advanced Technology and Intelligent Operations centers. Our 699,000 people deliver on the promise of technology and human ingenuity every day, serving clients in more than 120 countries. Visit us at www.accenture.com What would you do You will be aligned with our Life Sciences R&D vertical. Our services span across the entire life sciences enterprise, from research laboratories, clinical trials support, and regulatory services, to pharmacovigilance and patient services solutions. Employees under this span will be a part of one of the sub-offerings - Clinical, Pharmacovigilance & Regulatory, helping the worlds leading biopharma companies bring their vision to life enabling them to improve outcomes by converging around the patient, connecting scientific expertise with unique insights into the patient experience.The Clinical Database Management team is committed to developing and maintaining databases for clinical trials, which includes creating Case Report Forms (CRFs) and developing Edit and Validation Checks using programming skills like C# and PL/SQL. They also support data integration from various external sources such as labs and Interactive Response Technology (IRT). The team ensures compliance with regulatory standards and guidelines, maintaining high levels of data quality and security. Additionally, they continuously seek to improve database processes and tools by leveraging the latest technologies and best practices to enhance efficiency and accuracyDevelop clinical study databases by building electronic case report forms and program edit checks as per specifications. Support any updates or changes to the study database (e.g. Protocol amendments) through the change control process. What are we looking for Clinical Database ProgrammingClinical Data ManagementSAS ClinicalAdaptable and flexibleAbility to perform under pressureProblem-solving skillsAbility to establish strong client relationshipAgility for quick learningOracle Procedural Language Extensions to SQL (PLSQL)Structured Query Language (SQL)C, C++, C# Programming SkillsExperience in writing programs using Javascript Roles and Responsibilities: In this role you are required to solve routine problems, largely through precedent and referral to general guidelines Your expected interactions are within your own team and direct supervisor You will be provided detailed to moderate level of instruction on daily work tasks and detailed instruction on new assignments The decisions that you make would impact your own work You will be an individual contributor as a part of a team, with a predetermined, focused scope of work Please note that this role may require you to work in rotational shifts Qualification BE,BTech
Posted 2 weeks ago
15.0 - 20.0 years
18 - 22 Lacs
Hyderabad
Work from Office
Project Role : Data Platform Architect Project Role Description : Architects the data platform blueprint and implements the design, encompassing the relevant data platform components. Collaborates with the Integration Architects and Data Architects to ensure cohesive integration between systems and data models. Must have skills : Microsoft Azure Data Services Good to have skills : Microsoft Azure Databricks, Python (Programming Language), Microsoft SQL ServerMinimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As a Data Platform Architect, you will be responsible for architecting the data platform blueprint and implementing the design, which includes various data platform components. Your typical day will involve collaborating with Integration Architects and Data Architects to ensure seamless integration between systems and data models, while also addressing any challenges that arise during the implementation process. You will engage in discussions with stakeholders to gather requirements and provide insights that drive the overall architecture of the data platform, ensuring it meets the needs of the organization effectively. Roles & Responsibilities:- Expected to be an SME.- Collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Facilitate knowledge sharing sessions to enhance team capabilities.- Develop and maintain documentation related to data architecture and design. Professional & Technical Skills: - Must To Have Skills: Proficiency in Microsoft Azure Data Services.- Good To Have Skills: Experience with Microsoft Azure Databricks, Python (Programming Language), Microsoft SQL Server.- Strong understanding of data modeling techniques and best practices.- Experience with cloud-based data storage solutions and data processing frameworks.- Familiarity with data governance and compliance standards. Additional Information:- The candidate should have minimum 7.5 years of experience in Microsoft Azure Data Services.- This position is based at our Hyderabad office.- A 15 years full time education is required. Qualification 15 years full time education
Posted 2 weeks ago
4.0 - 9.0 years
7 - 11 Lacs
Thiruvananthapuram
Work from Office
Workday Consultant Technical Role - Xpetize Technology Solutions PVT LTD Experience : 3 4 years Notice Period Immediate Joiner Mode Of Work / Location: work from office (Trivandrum). Work Location: Bangalore (Work From Office) Skill Set Required Working knowledge of Workday Studio, EIB, Core Connectors, Workday Cloud Connect and Web Services. Working knowledge of SOAP, REST, XML, and XSLT. Strong Experience with calculated fields and reporting tools like BIRT, Matrix, Composite, discovery boards and future report specific functionality. Workday Security and Business Process Configuration. Workday Prism Analytics and Accounting Center experience Workday certification About Us Xpetize is a technology solutions company, supporting customers in IoT, application and engineering services, data services, cybersecurity, cloud and social services. We are headquartered in Trivandrum with offices in Bengaluru, Pune, USA and Japan. We work tirelessly to help our customers across the globe since 2011 and relentlessly trying to grow our expertise across geographies. Our flagship Industry 4.0 product XPETICS, is a fully managed IIoT platform that lets customers securely connect and process IoT data at scale. We have a flexible and open work culture, lots of fun, flexi work hours, up skilling programs, medical insurance for family and parents and ensuring work life balance.
Posted 2 weeks ago
3.0 - 5.0 years
30 - 35 Lacs
Ahmedabad
Work from Office
As a Data Engineer , you'll be at the heart of this mission-designing, building, and optimizing data solutions that drive real impact for our customers. This role sits in our Centre of Excellence team, which is an innovative hub dedicated to helping customers advance their Azure data roadmaps and strategies. Key Responsibilities: Build complex, scalable data solutions based on detailed architectural designs Test & troubleshoot data pipelines and integrations using your technical expertise Document your work clearly to support collaboration and future development Research & develop innovative approaches to data engineering challenges Define and promote best practices across the team and wider business Skills & Qualifications: Proven experience in data engineering and a range of Azure data services Experience with Fabric and Synapse Skilled in coding and building ETL pipelines Strong analytical and problem-solving mindset Comfortable working with a diverse range of customers Familiar with engineering practices and processes Committed to best practice adherence in everything you do What You Will Get? Competitive compensation and benefits Opportunity to work on enterprise-grade projects with leading clients A collaborative, learning-focused culture Flexible work environment and growth opportunities
Posted 2 weeks ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
32455 Jobs | Dublin
Wipro
16590 Jobs | Bengaluru
EY
11025 Jobs | London
Accenture in India
10991 Jobs | Dublin 2
Amazon
8878 Jobs | Seattle,WA
Uplers
8715 Jobs | Ahmedabad
IBM
8204 Jobs | Armonk
Oracle
7750 Jobs | Redwood City
Capgemini
6181 Jobs | Paris,France
Muthoot FinCorp (MFL)
6170 Jobs | New Delhi