Jobs
Interviews

16161 Spark Jobs - Page 24

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

0 years

0 Lacs

Chennai, Tamil Nadu, India

On-site

Duties for this role include but not limited to: supporting the design, build, test and maintain data pipelines at big data scale. Assists with updating data from multiple data sources. Work on batch processing of collected data and match its format to the stored data, make sure that the data is ready to be processed and analyzed. Assisting with keeping the ecosystem and the pipeline optimized and efficient, troubleshooting standard performance, data related problems and provide L3 support. Implementing parsers, validators, transformers and correlators to reformat, update and enhance the data. Data Engineers play a pivotal role within Dataworks, focused on creating and driving engineering innovation and facilitating the delivery of key business initiatives. Acting as a “universal translator” between IT, business, software engineers and data scientists, data engineers collaborate across multi-disciplinary teams to deliver value. Data Engineers will work on those aspects of the Dataworks platform that govern the ingestion, transformation, and pipelining of data assets, both to end users within FedEx and into data products and services that may be externally facing. Day-to-day, they will be deeply involved in code reviews and large-scale deployments. Essential Job Duties & Responsibilities Understanding in depth both the business and technical problems Dataworks aims to solve Building tools, platforms and pipelines to enable teams to clearly and cleanly analyze data, build models and drive decisions Scaling up from “laptop-scale” to “cluster scale” problems, in terms of both infrastructure and problem structure and technique Collaborating across teams to drive the generation of data driven operational insights that translate to high value optimized solutions. Delivering tangible value very rapidly, collaborating with diverse teams of varying backgrounds and disciplines Codifying best practices for future reuse in the form of accessible, reusable patterns, templates, and code bases Interacting with senior technologists from the broader enterprise and outside of FedEx (partner ecosystems and customers) to create synergies and ensure smooth deployments to downstream operational systems Skill/Knowledge Considered a Plus Technical background in computer science, software engineering, database systems, distributed systems Fluency with distributed and cloud environments and a deep understanding of optimizing computational considerations with theoretical properties Experience in building robust cloud-based data engineering and curation solutions to create data products useful for numerous applications Detailed knowledge of the Microsoft Azure tooling for large-scale data engineering efforts and deployments is highly preferred. Experience with any combination of the following azure tools: Azure Databricks, Azure Data Factory, Azure SQL D, Azure Synapse Analytics Developing and operationalizing capabilities and solutions including under near real-time high-volume streaming conditions. Hands-on development skills with the ability to work at the code level and help debug hard to resolve issues. A compelling track record of designing and deploying large scale technical solutions, which deliver tangible, ongoing value Direct experience having built and deployed robust, complex production systems that implement modern, data processing methods at scale Ability to context-switch, to provide support to dispersed teams which may need an “expert hacker” to unblock an especially challenging technical obstacle, and to work through problems as they are still being defined Demonstrated ability to deliver technical projects with a team, often working under tight time constraints to deliver value An ‘engineering’ mindset, willing to make rapid, pragmatic decisions to improve performance, accelerate progress or magnify impact Comfort with working with distributed teams on code-based deliverables, using version control systems and code reviews Ability to conduct data analysis, investigation, and lineage studies to document and enhance data quality and access Use of agile and devops practices for project and software management including continuous integration and continuous delivery Demonstrated expertise working with some of the following common languages and tools: Spark (Scala and PySpark), Kafka and other high-volume data tools SQL and NoSQL storage tools, such as MySQL, Postgres, MongoDB/CosmosDB Java, Python data tools Azure DevOps experience to track work, develop using git-integrated version control patterns, and build and utilize CI/CD pipelines Working knowledge and experience implementing data architecture patterns to support varying business needs Experience with different data types (json, xml, parquet, avro, unstructured) for both batch and streaming ingestions Use of Azure Kubernetes Services, Eventhubs, or other related technologies to implement streaming ingestions Experience developing and implementing alerting and monitoring frameworks Working knowledge of Infrastructure as Code (IaC) through Terraform to create and deploy resources Implementation experience across different data stores, messaging systems, and data processing engines Data integration through APIs and/or REST service PowerPlatform (PowerBI, PowerApp, PowerAutomate) development experience a plus Additional Job Description Analytical Skills, Accuracy & Attention to Detail, Planning & Organizing Skills, Influencing & Persuasion Skills, Presentation Skills FedEx was built on a philosophy that puts people first, one we take seriously. We are an equal opportunity/affirmative action employer and we are committed to a diverse, equitable, and inclusive workforce in which we enforce fair treatment, and provide growth opportunities for everyone. All qualified applicants will receive consideration for employment regardless of age, race, color, national origin, genetics, religion, gender, marital status, pregnancy (including childbirth or a related medical condition), physical or mental disability, or any other characteristic protected by applicable laws, regulations, and ordinances. Our Company FedEx is one of the world's largest express transportation companies and has consistently been selected as one of the top 10 World’s Most Admired Companies by "Fortune" magazine. Every day FedEx delivers for its customers with transportation and business solutions, serving more than 220 countries and territories around the globe. We can serve this global network due to our outstanding team of FedEx team members, who are tasked with making every FedEx experience outstanding. Our Philosophy The People-Service-Profit philosophy (P-S-P) describes the principles that govern every FedEx decision, policy, or activity. FedEx takes care of our people; they, in turn, deliver the impeccable service demanded by our customers, who reward us with the profitability necessary to secure our future. The essential element in making the People-Service-Profit philosophy such a positive force for the company is where we close the circle, and return these profits back into the business, and invest back in our people. Our success in the industry is attributed to our people. Through our P-S-P philosophy, we have a work environment that encourages team members to be innovative in delivering the highest possible quality of service to our customers. We care for their well-being, and value their contributions to the company. Our Culture Our culture is important for many reasons, and we intentionally bring it to life through our behaviors, actions, and activities in every part of the world. The FedEx culture and values have been a cornerstone of our success and growth since we began in the early 1970’s. While other companies can copy our systems, infrastructure, and processes, our culture makes us unique and is often a differentiating factor as we compete and grow in today’s global marketplace.

Posted 4 days ago

Apply

2.0 years

0 Lacs

Bengaluru, Karnataka, India

On-site

Job Title: Associate Career Level - C2 Introduction to role Are you ready to disrupt an industry and change lives? As an Associate, you will be at the forefront of GCC’s Sampling Allocation Service, developing and implementing analytical programs that optimize sampling distributions for AstraZeneca's branded products. This role is both managerial and hands-on, requiring proactive consultation with brand team collaborators and guiding the internal GCC Sampling team to ensure deliverables meet specifications. You'll work with brand teams to understand rules, requirements, and sampling strategies, applying your proficiency across multiple commercial datasets for proper implementation. Your efforts will feed into the sample ordering and distribution system, directly impacting our ability to develop life-changing medicines. Accountabilities In this dynamic role, you'll bring to bear your strong analytical skills and excellent communication abilities to forge effective business partnerships that drive tangible business impact. You'll continuously evaluate new quantitative analysis methods and technologies, manage sample allocation priorities across brands and therapeutic areas, allocate resources based on demand, liaise with AZ Sampling Stakeholders, and pull through analytics and coding standard methodologies. Your understanding of AZ core therapy areas and familiarity with core functions within AZ will be crucial as you lead a talented team. Essential Skills/Experience Quantitative Bachelor’s degree from an accredited college or university is required in one of the following or related fields: Engineering, Operations Research, Management Science, Economics, Statistics, Applied Math, Computer Science or Data Science. An advanced degree is preferred (Master’s, MBA or PhD). 2+ years of experience in Pharmaceutical / Biotech / Healthcare analytics or secondary data analysis. 3+ years of experience in application of advanced methods and statistical procedures on large and disparate datasets, specifically: Data Mining, Predictive Modelling algorithms. Optimisation & Simulation. 2+ years of recent experience and proficiency with Python, R, SQL and big data technology - Hadoop ecosystem (Cloudera distribution - Impala, Hive, Hbase, Spark, MapReduce etc.,). Understanding of the Veeva system and Veeva data, Alignment, Personal and Non-personal interactions and channels. Working knowledge of data visualisation – PowerBI, VBA or similar tools. Experience in MS Office products - PowerApps, Excel and PowerPoint skills required. Proficiency in manipulating and extracting insights from large longitudinal data sources, such as Claims, EMR and other patient-level data sets. Expertise in managing and analysing a range of large, secondary transactional databases is required. Statistical analysis and modelling background ML a plus Experience with IQVIA datasets as well as sales-related data sets such as targeting and alignment, HCP eligibility (blocking), and call data Experience with data visualisation methods and tools Ability to derive, summarise and communicate insights from analyses Organisation and time management skills Desirable Skills/Experience Strong leadership and interpersonal skills with demonstrated ability to work collaboratively with a significant number of business leaders and cross-functional business partners. Strong communication and influencing skills with demonstrated ability to develop and effectively present succinct, compelling reviews of independently developed analyses infused with insight and business implications/actions to be considered. Strategic and critical thinking with the ability to engage, build and maintain credibility with the Commercial Leadership Team. Strong organisational skills and time management; ability to manage a diverse range of simultaneous projects. Knowledge of the AZ brand and science is mandatory Experience using Big Data is a plus. Exposure to SPARK is desirable Should have Excellent Analytical problem-solving ability. Should be able to grasp new concepts quickly When we put unexpected teams in the same room, we unleash bold thinking with the power to inspire life-changing medicines. In-person working gives us the platform we need to connect, work at pace and challenge perceptions. That's why we work, on average, a minimum of three days per week from the office. But that doesn't mean we're not flexible. We balance the expectation of being in the office while respecting individual flexibility. Join us in our unique and ambitious world. At AstraZeneca, our work has a direct impact on patients by transforming our ability to develop life-changing medicines. We empower the business to perform at its peak by combining brand new science with leading digital technology platforms and data. Our dynamic environment offers countless opportunities to learn and grow through hackathons, exploring new technologies, and transforming roles forever. With a diversity of expertise unique to AstraZeneca, you'll dive deep into groundbreaking technology while broadening your understanding of our wider work. Ready to make a meaningful impact? Apply now to join our team! Date Posted 24-Jul-2025 Closing Date 30-Jul-2025 AstraZeneca embraces diversity and equality of opportunity. We are committed to building an inclusive and diverse team representing all backgrounds, with as wide a range of perspectives as possible, and harnessing industry-leading skills. We believe that the more inclusive we are, the better our work will be. We welcome and consider applications to join our team from all qualified candidates, regardless of their characteristics. We comply with all applicable laws and regulations on non-discrimination in employment (and recruitment), as well as work authorization and employment eligibility verification requirements.

Posted 4 days ago

Apply

2.0 - 4.0 years

0 Lacs

Mysore, Karnataka, India

On-site

Training-related experience Must have Teaching experience: conducting training sessions in classroom and dynamically responding to different capabilities of learners; experience in analyzing the feedback from sessions and identifying action areas for self-improvement Developing teaching material: Experience in developing teaching material, including exercises and assignments Good presentation skills, excellent oral / written communication skills Nice to have Teaching experience: Experience in delivering session over virtual classrooms Instructional Design: Developing engaging content Designing Assessments: Experience in designing assessments to evaluate the effectiveness of training and gauging the proficiency of the learner Participated in activities of the software development lifecycle like development, testing, configuration management Job Responsibilities Develop teaching materials including exercises & assignments Conduct classroom training / virtual training Design assessments Enhance course material & course delivery based on feedback to improve training effectiveness Location: Mysore, Mangalore, Bangalore, Chennai, Pune, Hyderabad, Chandigarh Description of the Profile We are looking for trainers with 2 to 4 years of teaching experience and technology know-how in one or more of the following areas: Java – Java programming, Spring, Angular / React, Bootstrap Microsoft – C# programming, SQL Server, ADO.NET, ASP.NET, MVC design pattern, Azure, MS Power platforms, MS Dynamics 365 CRM, MS Dynamics 365 ERP, SharePoint Testing – Selenium, Microfocus - UFT, Microfocus-ALM tools, SOA testing, SOAPUI, Rest assured, Appium Big Data – Python programming, Hadoop, Spark, Scala, Mongo DB, NoSQL SAP – SAP ABAP programming / SAP MM / SAP SD /SAP BI / SAP S4 HANA Oracle – Oracle E-Business Suite (EBS) / PeopleSoft / Siebel CRM / Oracle Cloud / OBIEE / Fusion Middleware API and integration – API, Microservices, TIBCO, APIGee, Mule Digital Commerce – SalesForce, Adobe Experience Manager Digital Process Automation - PEGA, Appian, Camunda, Unqork, UIPath MEAN / MERN stacks Business Intelligence – SQL Server, ETL using SQL Server, Analysis using SQL Server, Enterprise reporting using SQL, Visualization Data Science – Python for data science, Machine learning, Exploratory data analysis, Statistics & Probability Cloud & Infrastructure Management – Network administration / Database administration / Windows administration / Linux administration / Middleware administration / End User Computing / ServiceNow Cloud platforms like AWS / GCP/ Azure / Oracle Cloud, Virtualization Cybersecurity - Infra Security / Identity & Access Management / Application Security / Governance & Risk Compliance / Network Security Mainframe – COBOL, DB2, CICS, JCL Open source – Python, PHP, Unix / Linux, MySQL, Apache, HTML5, CSS3, JavaScript DBMS – Oracle / SQL Server / MySQL / DB2 / NoSQL Design patterns, Agile, DevOps

Posted 4 days ago

Apply

7.0 years

0 Lacs

Gurugram, Haryana, India

On-site

What’s up? We’re VaynerMedia ! We are a contemporary global creative and media agency with expertise in driving relevance for clients and delivering impactful business results. The independently-owned agency was founded in 2009 and has offices in New York, Los Angeles, London, Singapore, Australia, Thailand, Japan and Mexico City. VaynerMedia has been recognized for its work at Cannes Lions, the Clio Awards and The Webby Awards. It is part of the VaynerX family of companies. Culture is our key and Empathy is how we build it. VaynerMediaAPAC is strong with a world-class combination of diverse backgrounds collaborating to put forward innovative, creative and strategic solutions to the one advantage that stands the test of time --- PEOPLE. Day in and day out, we propel some of the biggest brands in the world to the intersection of attention and culture. In APAC, we are proudly +250, servicing a diverse client all around the region. The PITCH We’re lucky that everyone within our walls thinks creatively, shares a passion for building communities, and understands the world of both big campaigns and social micro-content. Our ACDs here serve a slightly different (and more senior) role than more traditional “digital creative types,” but thankfully we’ll never use the term “digital ninja”. Are you based in Delhi/ NCR? Geeking the creative world in the India market for 7+ years now? Oh well, we could be meant to be -- Here’s where you come in. I use my creative craft, understanding of platforms and grasp of culture to solve marketing problems for a variety of brands. I have the most challenging role in the creative department as my duties are split between both doing the work as well as helping manage some of the work. I spend roughly 80% of my time working on solving briefs and 20% of my time managing projects under the direction of my CD. I don’t have an assigned partner so I often work alone. Occasionally, I’ll work with another creative in the pod. And sometimes I’ll partner with the CD. While other members of the pod do not technically report to me, it’s expected that I help mentor them when possible. TL;DR - I have the most challenging role in the creative department. I fulfill the duties of an AD or CW but also help manage projects under my CD’s discretion. #thenavysealsofcreative Qualifications Include Lead with ideas, not just execution. You're a master at separating big, bold concepts from the noise of tactics, and knowing exactly when to dial up imagination or drill down into execution Brings 7+ years of hands-on experience in art direction, design, and copywriting, with a portfolio that showcases original thinking across platforms Thinks in stories, not just assets. You’re fluent in social-first creative, social-first campaigns, and know how to craft narratives that stop thumbs and spark conversations Is a natural leader. You inspire teams, guide junior talent, and know how to elevate an idea without overshadowing others Thrives on collaboration. You love being client-facing, can present work with clarity and confidence, and build trust through thoughtful engagement Navigates feedback like a pro. You take input with humility and give direction with clarity—turning every round into a better version of the last You are hands on with both design and copy aspects of the creative process and execution and are able to derive the best productive output Carries curiosity as a superpower. You’re always asking "what if?", constantly exploring cultural trends, platforms, and creative tools Is hungry but humble. You're ambitious but grounded, driven but kind. You lead with empathy and roll up your sleeves when needed Has a strategic mind with a creative soul. You don't just write or design—you think through the "why" behind every creative choice, aligning brand, business, and audience needs Natural leadership tendencies Exceptional video production experience Great with feedback and revisions What You Should Know About VaynerMedia Think “lab” and not “agency”. We get excited about solving business problems, not creating advertising for the sake of making advertising. Our entrepreneurship DNA runs deep. We’re willing to break rules, try new things, and test hypotheses if it means better understanding our craft. It’s fun! Our environment (and pace) feels much more like a start-up than most agencies you might be used to. Hopefully that, and the opportunity to collaborate with the talented and enthusiastic crew, work with big brands excites you, as it does us. We can’t wait to meet you.

Posted 4 days ago

Apply

3.0 years

0 Lacs

Chennai, Tamil Nadu, India

On-site

Job Description As a Data Engineer, you will leverage your technical expertise in data, analytics, cloud technologies, and analytic software tools to identify best designs, improve business processes, and generate measurable business outcomes. You will work with Data Engineering teams from within D&A, across the Pro Tech portfolio and additional Ford organizations such as GDI&A (Global Data Insight & Analytics), Enterprise Connectivity, Ford Customer Service Division, Ford Credit, etc. Develop EL/ELT/ETL pipelines to make data available in BigQuery analytical data store from disparate batch, streaming data sources for the Business Intelligence and Analytics teams. Work with on-prem data sources (Hadoop, SQL Server), understand the data model, business rules behind the data and build data pipelines (with GCP, Informatica) for one or more Ford Pro verticals. This data will be landed in GCP BigQuery. Build cloud-native services and APIs to support and expose data-driven solutions. Partner closely with our data scientists to ensure the right data is made available in a timely manner to deliver compelling and insightful solutions. Design, build and launch shared data services to be leveraged by the internal and external partner developer community. Building out scalable data pipelines and choosing the right tools for the right job. Manage, optimize and Monitor data pipelines. Provide extensive technical, strategic advice and guidance to key stakeholders around data transformation efforts. Understand how data is useful to the enterprise. Responsibilities Bachelors Degree 3+ years of experience with SQL and Python 2+ years of experience with GCP or AWS cloud services; Strong candidates with 5+ years in a traditional data warehouse environment (ETL pipelines with Informatica) will be considered 3+ years of experience building out data pipelines from scratch in a highly distributed and fault-tolerant manner. Comfortable with a broad array of relational and non-relational databases. Proven track record of building applications in a data-focused role (Cloud and Traditional Data Warehouse) Qualifications Experience with GCP cloud services including BigQuery, Cloud Composer, Dataflow, CloudSQL, GCS, Cloud Functions and Pub/Sub. Inquisitive, proactive, and interested in learning new tools and techniques. Familiarity with big data and machine learning tools and platforms. Comfortable with open source technologies including Apache Spark, Hadoop, Kafka. 1+ year experience with Hive, Spark, Scala, JavaScript. Strong oral, written and interpersonal communication skills Comfortable working in a dynamic environment where problems are not always well-defined. M.S. in a science-based program and/or quantitative discipline with a technical emphasis.

Posted 4 days ago

Apply

5.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Join Amgen’s Mission of Serving Patients At Amgen, if you feel like you’re part of something bigger, it’s because you are. Our shared mission—to serve patients living with serious illnesses—drives all that we do. Since 1980, we’ve helped pioneer the world of biotech in our fight against the world’s toughest diseases. With our focus on four therapeutic areas –Oncology, Inflammation, General Medicine, and Rare Disease– we reach millions of patients each year. As a member of the Amgen team, you’ll help make a lasting impact on the lives of patients as we research, manufacture, and deliver innovative medicines to help people live longer, fuller happier lives. Our award-winning culture is collaborative, innovative, and science based. If you have a passion for challenges and the opportunities that lay within them, you’ll thrive as part of the Amgen team. Join us and transform the lives of patients while transforming your career. What You Will Do Let’s do this. Let’s change the world. In this vital role you will be member of the technical team that develops and implements technology strategies, solutions, and services to support Amgen’s Global Medical business function. As the Software Engineer, you will be the primary accountable for developing integration of several capabilities and the point of contact for the business & technology product owner and other stakeholders on integration pipelines. This role will work closely with business and technology teams to efficiently build, develop, deploy, scale, and operate technology solutions and software products and be responsible for development and lifecycle management of integration pipelines. The Senior Associate will be a responsible member of a high-performing Agile Product Team consisting of business analysts, software engineers, solution architects, system configurators, etc. Roles & Responsibilities: Be an accountable member on the identified agile product team and portfolio of software applications Enable the technology ecosystem for Global Medical Data and Analytics platform team and ensure that the platforms meet their business needs and enterprise technology standards Implement a robust DevOps pipeline for the product using Amgen’s CI/CD frameworks and tools Development to deployment of data integration User Stories, Data pipelines and APIs using city planning technology using tools like databricks, MuleSoft, AWS, salesforce connectors, Python, GIT Support Lifecycle management and ownership of technology assets. Develop cross-functional collaborations with platform teams to maintain enterprise standards Drive a Product First attitude amongst the team and peers with a goal towards global expansion of the product’s capabilities Drive an attitude of excellence in running robust, reliable, and secure software products following Amgen’s software development, operating procedures, and change management requirements Unblock and drive engagement to improve results delivery for the team and the organization Drive external software product and service partners to deliver excellence Develop and mentor the team of engineers, architects, and data scientists for higher performance, and individual career growth. Act as a role model demonstrating leadership behaviors Partner with leadership in the Global Medical organization to define, develop, and implement technology solutions aligned with the organization’s goals and objectives, with a focus on increasing the effectiveness of the organization. Establish and maintain successful relationships with internal DTI platform and shared services product teams, strategic external partners, industry bodies, to learn, influence, and attract the best talent. What We Expect Of You We are all different, yet we all use our unique contributions to serve patients. The [vital attribute] professional we seek is a [type of person] with these qualifications. Basic Qualifications: Master's degree / Bachelor's degree and 5 to 9 years of Information Systems experience. Preferred Qualifications: Functional Skills 5+ years of IT Industry work experience in analysis, design, development, testing and maintenance of software applications in the Health & Life Sciences industry 2+ years' of experience supporting Medical Affairs business teams Experience working on major programs/projects that involve multiple partners and external vendors Experience in lifecycle management of technology applications and data pipelines Experience working in Scaled Agile Framework (SAFe), especially in regulated setting including ability to lead the transformation of teams from a service based to a product-based model Working experience on Enterprise Datalake(EDL) and technologies like Databricks, AWS, Mulesoft, SQL, Spark, Python, API, REACT framework, GITLab, MLOps etc. In-depth experience in all aspects of SDLC from requirements, design, testing, data analysis and Change Control process combined with experience developing project charters, statement of work and project financials Experience with data warehousing and EDL technologies like databricks to build integrations across multiple systems An ongoing commitment to learning and staying at the forefront of AI/ML advancements. Experience or working knowledge of SaaS solutions like Salesforce, Veeva, AWS, etc. Experience within a highly regulated pharmaceutical or technology organization, with the ability to ensure compliance with industry regulations and standard methodologies for data integrity, security, and audit readiness. Experience driving a collaborative culture that values technical depth, accountability, and customer service. Strong problem-solving and analytical skills. Demonstrated ability to work effectively in a fast-paced, dynamic environment. Experience handling vendor relationships and working with external partners or consultants to ensure optimal performance, support, and development of the ServiceNow platform Professional Certifications Preferred certification in SAFe for Teams Preferred certification in EDL tools like databricks Preferred AWS certifications Soft Skills: Excellent leadership skills with the ability to work in collaborative teams and cultivate talent. Demonstrated ability to influence decisions with business & technology collaborators & leadership Data-driven decision making Excellent verbal and written communication skills for technical and non-technical audiences of various levels in the organization. High degree of initiative and self-motivation Ability to handle multiple priorities successfully Team-oriented with a focus on achieving team goals Strong presentation and public speaking skills Excellent analytical and troubleshooting skills Ability to work effectively with global, virtual teams Equal opportunity statement Amgen is an Equal Opportunity employer and will consider you without regard to your race, color, religion, sex, sexual orientation, gender identity, national origin, protected veteran status, or disability status. We will ensure that individuals with disabilities are provided with reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.

Posted 4 days ago

Apply

0 years

0 Lacs

Andhra Pradesh, India

On-site

We are seeking a highly skilled and motivated Big Data Engineer to join our data engineering team. The ideal candidate will have hands-on experience with Hadoop ecosystem, Apache Spark, and programming expertise in Python (PySpark), Scala, and Java. You will be responsible for designing, developing, and optimizing scalable data pipelines and big data solutions to support analytics and business intelligence initiatives.

Posted 4 days ago

Apply

4.0 - 9.0 years

20 - 35 Lacs

Gurugram

Work from Office

Job Description - The candidate should have extensive production experience (2+ Years ) in GCP - Strong background in Data engineering 2-3 Years of exp in Big Data technologies including, Hadoop, NoSQL, Spark, Kafka etc. - Exposure to enterprise application development is a must. Roles & Responsibilities 4-10 years of IT experience range is preferred. Able to effectively use GCP managed services e.g. Dataproc, Dataflow, pub/sub, Cloud functions, Big Query, GCS - At least 4 of these Services. Good to have knowledge on Cloud Composer, Cloud SQL, Big Table, Cloud Function. Strong experience in Big Data technologies Hadoop, Sqoop, Hive and Spark including DevOPs. Good hands on expertise on either Python or Java programming. Good Understanding of GCP core services like Google cloud storage, Google compute engine, Cloud SQL, Cloud IAM. Good to have knowledge on GCP services like App engine, GKE, Cloud Run, Cloud Built, Anthos. Ability to drive the deployment of the customers workloads into GCP and provide guidance, cloud adoption model, service integrations, appropriate recommendations to overcome blockers and technical road-maps for GCP cloud implementations. Experience with technical solutions based on industry standards using GCP - IaaS, PaaS and SaaS capabilities. Extensive, real-world experience designing technology components for enterprise solutions and defining solution architectures and reference architectures with a focus on cloud technologies. Act as a subject-matter expert OR developer around GCP and become a trusted advisor to multiple teams. Technical ability to become certified in required GCP technical certifications.

Posted 4 days ago

Apply

5.0 - 8.0 years

15 - 20 Lacs

Noida, Pune, Bengaluru

Work from Office

Role: Business Data Analyst Location: All EXL Location Experience: 5+ Notice period- 15 days or less Key Responsibilities: Engage with stakeholders to gather and document data requirements Perform data mapping and define source-to-target transformations Create and maintain data dictionaries and technical documentation Collaborate on the design and layout of data pipelines Support data quality checks Required Skills: Strong hands-on experience with SQL and Spark (Python is a plus) Familiarity with metadata management and ETL concepts Good communication and documentation skills Strong hands-on experience with Excel Follow to keep yourself updated about future job openings linkedin.com/in/sonali-nayakawad-088b19199

Posted 4 days ago

Apply

0 years

0 Lacs

Pune, Maharashtra, India

On-site

Job Description Some careers have more impact than others. If you’re looking for a career where you can make a real impression, join HSBC and discover how valued you’ll be. HSBC is one of the largest banking and financial services organisations in the world, with operations in 64 countries and territories. We aim to be where the growth is, enabling businesses to thrive and economies to prosper, and, ultimately, helping people to fulfil their hopes and realise their ambitions. We are currently seeking an experienced professional to join our team in the role of a Senior Software Engineer In this role, you will Understand project requirements and develop design specifications as per business agreements Design, code, and maintain the Oracle systems based on established standards. Perform initial design reviews and recommend tactical as well as strategic improvements based on programme requirements. Write clear codes and prepare coding documentations as per programme requirements. Under DevOps initiative carry out in system integration and acceptance testing and perform bug fixes for production readiness. Analyse and troubleshoot production issues in a timely manner followed up with Root cause analysis documentation. Work independently with minimal supervision, closely collaborating with other solution architects, business analysts and project managers. Follow all relevant IT policies, processes and standard operating procedures so that work is carried out in a controlled and consistent manner Provide out of hours support to the production batch with a focus on performance tuning Delivering valuable working software to the business with a constant focus on technical process improvement Working in PODs to help deliver Stories / Tasks and also involved in providing the technical Support in Production tickets / Incidents on rota basis. Requirements To be successful in this role, you should meet the following requirements: Excellent communication skills (written or oral). Fluent written and spoken English is essential in order to communicate with the other teams / entities of the group (mainly teams based in Paris, New York, London and in Asia). Good exposure of core java, collections framework and OOPS Work experience on Spring boot framework and Junits. Data Analytics skills (as will be dealing with processing of large data sets) Good capabilities to work in a team split in different locations. Being responsive is essential, especially regarding the daily support of the application Being autonomous: knowing how to take responsibilities for actions to be undertaken and bring them to completion. Being “customer oriented”: knowing how to understand and interpret user needs. Being flexible. Additional Skills Knowledge of batch processing systems/tools (Choreographer, Airflow etc) Working knowledge of SQL and queries Exposure to Apache Spark or any big data framework Fair understanding of DevOps concepts You’ll achieve more when you join HSBC. www.hsbc.com/careers HSBC is committed to building a culture where all employees are valued, respected and opinions count. We take pride in providing a workplace that fosters continuous professional development, flexible working and opportunities to grow within an inclusive and diverse environment. Personal data held by the Bank relating to employment applications will be used in accordance with our Privacy Statement, which is available on our website. Issued by – HSBC Software Development India

Posted 4 days ago

Apply

5.0 - 10.0 years

15 - 25 Lacs

Hyderabad/Secunderabad, Bangalore/Bengaluru, Delhi / NCR

Hybrid

Genpact (NYSE: G) is a global professional services and solutions firm delivering outcomes that shape the future. Our 125,000+ people across 30+ countries are driven by our innate curiosity, entrepreneurial agility, and desire to create lasting value for clients. Powered by our purpose the relentless pursuit of a world that works better for people – we serve and transform leading enterprises, including the Fortune Global 500, with our deep business and industry knowledge, digital operations services, and expertise in data, technology, and AI. Inviting applications for the role of Lead Consultant-Data Engineer, AWS+Python, Spark, Kafka for ETL! Responsibilities Develop, deploy, and manage ETL pipelines using AWS services, Python, Spark, and Kafka. Integrate structured and unstructured data from various data sources into data lakes and data warehouses. Design and deploy scalable, highly available, and fault-tolerant AWS data processes using AWS data services (Glue, Lambda, Step, Redshift) Monitor and optimize the performance of cloud resources to ensure efficient utilization and cost-effectiveness. Implement and maintain security measures to protect data and systems within the AWS environment, including IAM policies, security groups, and encryption mechanisms. Migrate the application data from legacy databases to Cloud based solutions (Redshift, DynamoDB, etc) for high availability with low cost Develop application programs using Big Data technologies like Apache Hadoop, Apache Spark, etc with appropriate cloud-based services like Amazon AWS, etc. Build data pipelines by building ETL processes (Extract-Transform-Load) Implement backup, disaster recovery, and business continuity strategies for cloud-based applications and data. Responsible for analysing business and functional requirements which involves a review of existing system configurations and operating methodologies as well as understanding evolving business needs Analyse requirements/User stories at the business meetings and strategize the impact of requirements on different platforms/applications, convert the business requirements into technical requirements Participating in design reviews to provide input on functional requirements, product designs, schedules and/or potential problems Understand current application infrastructure and suggest Cloud based solutions which reduces operational cost, requires minimal maintenance but provides high availability with improved security Perform unit testing on the modified software to ensure that the new functionality is working as expected while existing functionalities continue to work in the same way Coordinate with release management, other supporting teams to deploy changes in production environment Qualifications we seek in you! Minimum Qualifications Experience in designing, implementing data pipelines, build data applications, data migration on AWS Strong experience of implementing data lake using AWS services like Glue, Lambda, Step, Redshift Experience of Databricks will be added advantage Strong experience in Python and SQL Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift. Advanced programming skills in Python for data processing and automation. Hands-on experience with Apache Spark for large-scale data processing. Experience with Apache Kafka for real-time data streaming and event processing. Proficiency in SQL for data querying and transformation. Strong understanding of security principles and best practices for cloud-based environments. Experience with monitoring tools and implementing proactive measures to ensure system availability and performance. Excellent problem-solving skills and ability to troubleshoot complex issues in a distributed, cloud-based environment. Strong communication and collaboration skills to work effectively with cross-functional teams. Preferred Qualifications/ Skills Master’s Degree-Computer Science, Electronics, Electrical. AWS Data Engineering & Cloud certifications, Databricks certifications Experience with multiple data integration technologies and cloud platforms Knowledge of Change & Incident Management process Genpact is an Equal Opportunity Employer and considers applicants for all positions without regard to race, color, religion or belief, sex, age, national origin, citizenship status, marital status, military/veteran status, genetic information, sexual orientation, gender identity, physical or mental disability or any other characteristic protected by applicable laws. Genpact is committed to creating a dynamic work environment that values diversity and inclusion, respect and integrity, customer focus, and innovation. Get to know us at genpact.com and on LinkedIn, X, YouTube, and Facebook. Furthermore, please do note that Genpact does not charge fees to process job applications and applicants are not required to pay to participate in our hiring process in any other way. Examples of such scams include purchasing a 'starter kit,' paying to apply, or purchasing equipment or training.

Posted 4 days ago

Apply

4.0 - 5.0 years

5 - 9 Lacs

Vadodara

Work from Office

Responsibilities - Conduct feature engineering, data analysis, and data exploration to extract valuable insights. - Develop and optimize Machine Learning models to achieve high accuracy and performance. - Design and implement Deep Learning models, including Artificial Neural Networks (ANN), Convolutional Neural Networks (CNN), and Reinforcement Learning techniques. - Handle real-time imbalanced datasets and apply appropriate techniques to improve model fairness and robustness. - Deploy models in production environments and ensure continuous monitoring, improvement, and updates based on feedback. - Collaborate with cross-functional teams to align ML solutions with business goals. - Utilize fundamental statistical knowledge and mathematical principles to ensure the reliability of models. - Bring in the latest advancements in ML and AI to drive innovation. Required Skills - 4-5 years of hands-on experience in Machine Learning and Deep Learning. - Strong expertise in feature engineering, data exploration, and data preprocessing. - Experience with imbalanced datasets and techniques to improve model generalization. - Proficiency in Python, TensorFlow, Scikit-learn, and other ML frameworks. - Strong mathematical and statistical knowledge with problem-solving skills. - Ability to optimize models for high accuracy and performance in real-world scenarios. Preferred Skills - Experience with Big Data technologies (Hadoop, Spark, etc.) - Familiarity with containerization and orchestration tools (Docker, Kubernetes). - Experience in automating ML pipelines with MLOps practices. - Experience in model deployment using cloud platforms (AWS, GCP, Azure) or MLOps tools.

Posted 4 days ago

Apply

4.0 - 5.0 years

5 - 9 Lacs

Patna

Work from Office

Responsibilities - Conduct feature engineering, data analysis, and data exploration to extract valuable insights. - Develop and optimize Machine Learning models to achieve high accuracy and performance. - Design and implement Deep Learning models, including Artificial Neural Networks (ANN), Convolutional Neural Networks (CNN), and Reinforcement Learning techniques. - Handle real-time imbalanced datasets and apply appropriate techniques to improve model fairness and robustness. - Deploy models in production environments and ensure continuous monitoring, improvement, and updates based on feedback. - Collaborate with cross-functional teams to align ML solutions with business goals. - Utilize fundamental statistical knowledge and mathematical principles to ensure the reliability of models. - Bring in the latest advancements in ML and AI to drive innovation. Required Skills - 4-5 years of hands-on experience in Machine Learning and Deep Learning. - Strong expertise in feature engineering, data exploration, and data preprocessing. - Experience with imbalanced datasets and techniques to improve model generalization. - Proficiency in Python, TensorFlow, Scikit-learn, and other ML frameworks. - Strong mathematical and statistical knowledge with problem-solving skills. - Ability to optimize models for high accuracy and performance in real-world scenarios. Preferred Skills - Experience with Big Data technologies (Hadoop, Spark, etc.) - Familiarity with containerization and orchestration tools (Docker, Kubernetes). - Experience in automating ML pipelines with MLOps practices. - Experience in model deployment using cloud platforms (AWS, GCP, Azure) or MLOps tools.

Posted 4 days ago

Apply

4.0 - 5.0 years

5 - 9 Lacs

Nagpur

Work from Office

Responsibilities - Conduct feature engineering, data analysis, and data exploration to extract valuable insights. - Develop and optimize Machine Learning models to achieve high accuracy and performance. - Design and implement Deep Learning models, including Artificial Neural Networks (ANN), Convolutional Neural Networks (CNN), and Reinforcement Learning techniques. - Handle real-time imbalanced datasets and apply appropriate techniques to improve model fairness and robustness. - Deploy models in production environments and ensure continuous monitoring, improvement, and updates based on feedback. - Collaborate with cross-functional teams to align ML solutions with business goals. - Utilize fundamental statistical knowledge and mathematical principles to ensure the reliability of models. - Bring in the latest advancements in ML and AI to drive innovation. Required Skills - 4-5 years of hands-on experience in Machine Learning and Deep Learning. - Strong expertise in feature engineering, data exploration, and data preprocessing. - Experience with imbalanced datasets and techniques to improve model generalization. - Proficiency in Python, TensorFlow, Scikit-learn, and other ML frameworks. - Strong mathematical and statistical knowledge with problem-solving skills. - Ability to optimize models for high accuracy and performance in real-world scenarios. Preferred Skills - Experience with Big Data technologies (Hadoop, Spark, etc.) - Familiarity with containerization and orchestration tools (Docker, Kubernetes). - Experience in automating ML pipelines with MLOps practices. - Experience in model deployment using cloud platforms (AWS, GCP, Azure) or MLOps tools.

Posted 4 days ago

Apply

4.0 - 5.0 years

5 - 9 Lacs

Pune

Work from Office

Responsibilities - Conduct feature engineering, data analysis, and data exploration to extract valuable insights. - Develop and optimize Machine Learning models to achieve high accuracy and performance. - Design and implement Deep Learning models, including Artificial Neural Networks (ANN), Convolutional Neural Networks (CNN), and Reinforcement Learning techniques. - Handle real-time imbalanced datasets and apply appropriate techniques to improve model fairness and robustness. - Deploy models in production environments and ensure continuous monitoring, improvement, and updates based on feedback. - Collaborate with cross-functional teams to align ML solutions with business goals. - Utilize fundamental statistical knowledge and mathematical principles to ensure the reliability of models. - Bring in the latest advancements in ML and AI to drive innovation. Required Skills - 4-5 years of hands-on experience in Machine Learning and Deep Learning. - Strong expertise in feature engineering, data exploration, and data preprocessing. - Experience with imbalanced datasets and techniques to improve model generalization. - Proficiency in Python, TensorFlow, Scikit-learn, and other ML frameworks. - Strong mathematical and statistical knowledge with problem-solving skills. - Ability to optimize models for high accuracy and performance in real-world scenarios. Preferred Skills - Experience with Big Data technologies (Hadoop, Spark, etc.) - Familiarity with containerization and orchestration tools (Docker, Kubernetes). - Experience in automating ML pipelines with MLOps practices. - Experience in model deployment using cloud platforms (AWS, GCP, Azure) or MLOps tools.

Posted 4 days ago

Apply

4.0 - 5.0 years

5 - 9 Lacs

Pimpri-Chinchwad

Work from Office

Responsibilities - Conduct feature engineering, data analysis, and data exploration to extract valuable insights. - Develop and optimize Machine Learning models to achieve high accuracy and performance. - Design and implement Deep Learning models, including Artificial Neural Networks (ANN), Convolutional Neural Networks (CNN), and Reinforcement Learning techniques. - Handle real-time imbalanced datasets and apply appropriate techniques to improve model fairness and robustness. - Deploy models in production environments and ensure continuous monitoring, improvement, and updates based on feedback. - Collaborate with cross-functional teams to align ML solutions with business goals. - Utilize fundamental statistical knowledge and mathematical principles to ensure the reliability of models. - Bring in the latest advancements in ML and AI to drive innovation. Required Skills - 4-5 years of hands-on experience in Machine Learning and Deep Learning. - Strong expertise in feature engineering, data exploration, and data preprocessing. - Experience with imbalanced datasets and techniques to improve model generalization. - Proficiency in Python, TensorFlow, Scikit-learn, and other ML frameworks. - Strong mathematical and statistical knowledge with problem-solving skills. - Ability to optimize models for high accuracy and performance in real-world scenarios. Preferred Skills - Experience with Big Data technologies (Hadoop, Spark, etc.) - Familiarity with containerization and orchestration tools (Docker, Kubernetes). - Experience in automating ML pipelines with MLOps practices. - Experience in model deployment using cloud platforms (AWS, GCP, Azure) or MLOps tools.

Posted 4 days ago

Apply

10.0 - 12.0 years

0 - 1 Lacs

Hyderabad, Chennai, Bengaluru

Hybrid

Experience Required: 10 to 12yrs (3-4 Years in Cloudera + Cloud Migration) Work location Hyderabad, Bangalore, Chennai, Noida, Pune Work Type - Hybrid model Work Time - Canada EST hours is a must Job Summary: We are seeking a skilled Cloudera Migration Specialist to lead the migration of our on-premises Cloudera cluster to Microsoft Azure. The ideal candidate will have 3–4 years of hands-on experience with Cloudera platform administration, optimization, and migration, along with a strong understanding of Azure cloud services and data engineering best practices. Key Responsibilities: Lead and execute the migration of Cloudera workloads (HDFS, Hive, Spark, Impala, HBase, etc.) from on-premise infrastructure to Azure. • Assess the existing Cloudera cluster, identify dependencies, and prepare a detailed migration roadmap. • Develop and implement data migration scripts, workflows, and cloud-native configurations. • Design and deploy equivalent services on Azure using Azure HDInsight, Azure Data Lake, Azure Synapse, or other relevant services. • Ensure data integrity, performance tuning, and post-migration validation. • Collaborate with infrastructure, security, and DevOps teams to ensure compliance and automation. • Prepare and maintain documentation of the migration plan, architecture, and troubleshooting playbooks. • Provide knowledge transfer and training to internal teams post-migration. Required Skills & Experience: 3–4 years of hands-on experience with Cloudera (CDH/CDP) ecosystem components (e.g., HDFS, YARN, Hive, Spark, Impala, HBase). • Proven experience in Cloudera cluster migrations, preferably to cloud platforms like Azure. • Solid understanding of cloud-native equivalents and data architectures on Azure. • Experience with Azure services such as HDInsight, Data Lake Storage, Synapse Analytics, Blob Storage. • Proficiency in Linux system administration, shell scripting, and automation tools. • Strong problem-solving and troubleshooting abilities in distributed data environments. • Familiarity with security controls, Kerberos, Ranger, LDAP integration, and data governance Preferred Qualifications: Cloudera Certified Administrator / Developer. • Experience with Azure DevOps, Terraform, or Ansible for infrastructure provisioning. • Knowledge of disaster recovery planning and HA architectures on Azure. • Familiarity with performance tuning in cloud vs. on-prem Hadoop environments

Posted 4 days ago

Apply

4.0 - 5.0 years

5 - 9 Lacs

Thane

Work from Office

Responsibilities - Conduct feature engineering, data analysis, and data exploration to extract valuable insights. - Develop and optimize Machine Learning models to achieve high accuracy and performance. - Design and implement Deep Learning models, including Artificial Neural Networks (ANN), Convolutional Neural Networks (CNN), and Reinforcement Learning techniques. - Handle real-time imbalanced datasets and apply appropriate techniques to improve model fairness and robustness. - Deploy models in production environments and ensure continuous monitoring, improvement, and updates based on feedback. - Collaborate with cross-functional teams to align ML solutions with business goals. - Utilize fundamental statistical knowledge and mathematical principles to ensure the reliability of models. - Bring in the latest advancements in ML and AI to drive innovation. Required Skills - 4-5 years of hands-on experience in Machine Learning and Deep Learning. - Strong expertise in feature engineering, data exploration, and data preprocessing. - Experience with imbalanced datasets and techniques to improve model generalization. - Proficiency in Python, TensorFlow, Scikit-learn, and other ML frameworks. - Strong mathematical and statistical knowledge with problem-solving skills. - Ability to optimize models for high accuracy and performance in real-world scenarios. Preferred Skills - Experience with Big Data technologies (Hadoop, Spark, etc.) - Familiarity with containerization and orchestration tools (Docker, Kubernetes). - Experience in automating ML pipelines with MLOps practices. - Experience in model deployment using cloud platforms (AWS, GCP, Azure) or MLOps tools.

Posted 4 days ago

Apply

9.0 - 14.0 years

22 - 32 Lacs

Bengaluru

Hybrid

NOTE: We are only looking for candidates who can join Immediately to available to join in 15 days Experience level- 6+ years Location: Bangalore, Hyderabad, Chennai (Candidates who are currently in one of these 3 locations can apply) Job Summary: We are seeking a highly skilled and experienced Senior Big Data + Python Developer to join our team in Bangalore . The ideal candidate will have a strong background in distributed computing, data engineering, and systems design. You will work on building scalable data pipelines, integrating with modern data platforms, and contributing to architectural decisions. Prior experience with AIML platforms is a plus but not mandatory. Key Responsibilities: Design, develop, and maintain large-scale data processing pipelines using Python , Spark , and Ray . Manage and optimize workflows with Airflow . Work with Apache Hive , Iceberg , and Druid for efficient data storage, querying, and real-time analytics. Deploy, monitor, and manage containerized applications using Kubernetes . Develop dashboards and visualizations using Apache Superset or similar tools. Collaborate with architects and data scientists to design high-performance systems. Ensure data quality, scalability, and performance in all solutions delivered. Participate in code reviews and provide technical mentorship to junior team members. Required Skill Set: Strong programming experience with Python . Hands-on experience with Ray , Apache Spark , and Hive . Solid understanding of Apache Iceberg for data lake management. Experience working with Kubernetes for container orchestration. Workflow orchestration with Apache Airflow . Familiarity with Apache Druid for real-time analytics. Expertise in data visualization tools such as Superset . Strong knowledge of data architecture and design patterns. Proven track record of delivering scalable and production-grade systems. Preferred Qualifications: Exposure to AI/ML workflows or platforms is a plus. Experience in system and data architecture design. Contributions to open-source or published technical blogs/articles. Excellent problem-solving and analytical skills.

Posted 4 days ago

Apply

5.0 - 8.0 years

0 Lacs

Chennai, Tamil Nadu, India

On-site

We are currently seeking a highly accomplished Senior Data Software Engineer to enhance our team and engage in significant projects that will advance both your skillset and your career trajectory. As a Senior Engineer, your efforts will be essential in architecting and deploying Data solutions across diverse projects. The candidate that is best suited for this role brings extensive proficiency in Big Data and related technologies, emphasizing Apache Spark, Python, Azure, and AWS. Responsibilities Develop and implement end-to-end Data solutions to satisfy complex business demands Collaborate with interdisciplinary teams to understand project requirements and deliver top-tier software solutions Employ your deep knowledge of Apache Spark, Python, Azure, and AWS to construct scalable and effective data processing systems Maintain and augment the performance, security, and scalability of Data applications Stay updated on industry trends and technological breakthroughs to enhance our development methodologies continuously Requirements 5-8 years of hands-on experience in Data and related technologies Advanced understanding of Apache Spark High-level skills in Hadoop and Hive Proficiency in Python Experience with native Cloud data services on AWS and Azure

Posted 4 days ago

Apply

1.0 years

0 Lacs

Noida, Uttar Pradesh, India

Remote

🌟 We're Hiring Creators at BettleByte! 🌟 Love crafting eye-catching designs or turning raw footage into scroll-stopping content? This one’s for you. BettleByte is on the hunt for a talented: 🎨 Graphic Designer 🎬 Video Editor Whether you're a pro with 1+ year of experience or an intern eager to prove your spark, we woud love to hear from you! 🛠️ What You Need to Know: ✔️ Proficiency in: – Adobe Photoshop – Adobe Illustrator – Adobe Premiere Pro 💼 What You’ll Work On: – Brand visuals, social media creatives, reels & shorts – Ad campaigns, product videos, motion graphics – Projects for startups, influencers, businesses & more 📍 Work Style: 3️⃣ Days Work From Office 3️⃣ Days Work From Home (Flexible, productive, and fun!) If you're passionate about visuals, vibes, and value-driven design, let’s build something brilliant at BettleByte 💥

Posted 4 days ago

Apply

5.0 - 8.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

We are seeking a highly experienced Senior Data Software Engineer to join our dynamic team and tackle challenging projects that will enhance your skills and career. As a Senior Engineer, your contributions will be critical in designing and implementing Data solutions across a variety of projects. The ideal candidate will possess deep experience in Big Data and associated technologies, with a strong emphasis on Apache Spark, Python, Azure and AWS. Responsibilities Develop and execute end-to-end Data solutions to meet complex business needs Work collaboratively with interdisciplinary teams to comprehend project needs and deliver superior software solutions Apply your expertise in Apache Spark, Python, Azure and AWS to create scalable and efficient data processing systems Maintain and enhance the performance, security, and scalability of Data applications Keep abreast of industry trends and technological advancements to foster continuous improvement in our development practices Requirements 5-8 years of direct experience in Data and related technologies Advanced knowledge and hands-on experience with Apache Spark High-level proficiency with Hadoop and Hive Proficiency in Python Prior experience with AWS and Azure native Cloud data services Technologies Hadoop Hive

Posted 4 days ago

Apply

9.0 - 15.0 years

7 - 11 Lacs

Pune, Chennai, Bengaluru

Work from Office

Skill: Scala Grade -C2/D1 Location: Pune/Chennai & Bangalore NP: Immediate to 15 Days Joiners Only Strong analytical skills with experience of previously working within an area of Risk, Finance and Treasury. Good experience on SCALA and Apache Spark open-source data analytics cluster computing framework. Experience in working with different file formats like JSON, Parquet, AVRO, ORC and XML. Excellent inter-personal skills with experience of briefing, de-briefing and presenting to senior executives and having effective listening skills. Able to communicate effectively, both orally and in writing, with clients, colleagues and external vendors. Excellent time management and planning skills with experience of working under pressure. Ability to remain organized and able to prioritize multiple incident priorities. Highest standards of personal integrity, professional conduct and ethics. Incident, problem and change management skills Minimum qualification should be BE/BTech or Equivalent Scala, Spark

Posted 4 days ago

Apply

9.0 - 15.0 years

6 - 10 Lacs

Gurugram

Work from Office

Skill: Scala with AWS Grade -C2 Location: Pan India NP: Immediate to 15 Days Joiners Only We are seeking a skilled Big Data Engineer with strong hands-on experience in Apache Spark, Scala, and AWS cloud services. The ideal candidate will be responsible for developing scalable data pipelines, processing large datasets, and deploying efficient Spark-based solutions on the AWS ecosystem. Design and develop distributed data processing pipelines using Apache Spark and Scala. Optimize Spark jobs for performance and scalability on AWS infrastructure. Integrate Spark applications with AWS services such as S3, EMR, Lambda, Glue, RDS, Athena, and Redshift. Write clean, reusable, and production-grade Scala code. Work with large-scale structured and unstructured data in real-time and batch modes. Ensure data quality, reliability, and consistency through monitoring and validations. Collaborate with data scientists, analysts, and business teams to understand requirements and deliver insights. Implement best practices for data engineering in a cloud-native environment. Required Skills: Strong programming skills in Scala 3+ years of hands-on experience with Apache Spark (RDD/DataFrame/SQL APIs) Experience working on AWS services like EMR, S3, Glue, Lambda, EC2, etc. Proficient in writing and optimizing complex Spark transformations and actions. Experience in working with large-scale data processing and distributed systems. Preferred Skills: Experience with Kafka, Airflow, or similar orchestration tools. Working knowledge of Python Experience with containerized deployments (Docker, Kubernetes). AWS certification (e.g., AWS Certified Data Analytics Specialty) is a plus. Aws, Scala, Spark

Posted 4 days ago

Apply

3.0 - 8.0 years

9 - 13 Lacs

Bengaluru

Work from Office

A Bachelors degree in Computer Science, Data Science, or a related field; a Master s is a plus 6 8 years of experience in data science with a strong track record of building and deploying ML solutions at scale Deep understanding of core ML techniques supervised, unsupervised, and semi-supervised learning along with strong foundations in statistics and linear algebra Exposure to deep learning concepts and architectures (e.g., CNNs, RNNs, Transformers) and their practical applications Proficiency in Python and SQL, with experience in building data pipelines and analytical workflows Hands-on experience with large-scale data processing using Apache Spark, Hadoop/Hive, or cloud platforms such as GCP Strong programming fundamentals and experience writing clean, maintainable, and production-ready code Excellent analytical and problem-solving skills the ability to extract actionable insights from messy and high-volume data Solid grasp of statistical testing, hypothesis validation, and common pitfalls in experimental design Experience designing and interpreting A/B tests, including uplift measurement and segmentation Ability to work closely with product and engineering teams to translate business goals into scalable ML or data solutions Bonus points for: Experience with reinforcement learning or sequence modeling techniques Contributions to ML libraries, internal tools, or research publications B2c, Cnn, Deploying, Fraud, Machine Learing, Network+, Search, Transformer

Posted 4 days ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies