Home
Jobs

45 R Program Jobs

Filter Interviews
Min: 0 years
Max: 25 years
Min: ₹0
Max: ₹10000000
Setup a job Alert
Filter
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

5.0 - 8.0 years

15 - 25 Lacs

Kolkata, Chennai, Bengaluru

Hybrid

Naukri logo

Global Gen AI Developer Enabling a software-defined, electrified future. Visteon is a technology company that develops and builds innovative digital cockpit and electrification products at the leading-edge of the mobility revolution. Founded in 2000, Visteon brings decades of automotive intelligence combined with Silicon Valley speed to apply global insights that help transform the software-defined vehicle of the future for many of the worlds largest OEMs. The company employs 10,000 employees in 18 countries around the globe. To know more about us click here. Mission of the Role: Facilitate Enterprise machine learning and artificial intelligence solutions using the latest technologies Visteon is adopting globally. Key Objectives of this Role: The primary goal of the Global ML/AI Developer is to leverage advanced machine learning and artificial intelligence techniques to develop innovative solutions that drive Visteons strategic initiatives. By collaborating with cross-functional teams and stakeholders, this role identifies opportunities for AI-driven improvements, designs and implements scalable ML models, and integrates these models into existing systems to enhance operational efficiency. Following development best practices, fostering a culture of continuous learning, and staying abreast of AI advancements, the Global ML/AI Developer ensures that all AI solutions align with organizational goals, support data-driven decision-making, and continuously improve Visteons technological capabilities. Qualification, Experience and Skills: 6-8 Yrs Technical Skills: Expertise in machine learning frameworks (e.g., TensorFlow, PyTorch), programming languages (e.g., Python, R, SQL), and data processing tools (e.g., Apache Spark, Hadoop). Proficiency in developing, training, and deploying ML models, including supervised and unsupervised learning, deep learning, and reinforcement learning. Strong understanding of data engineering concepts, including data preprocessing, feature engineering, and data pipeline development. Experience with cloud platforms (preferably Microsoft Azure) for deploying and scaling ML solutions. Business Acumen : Strong business analysis and ability to translate complex technical concepts into actionable business insights and recommendations. Key Behaviors: Innovation: Continuously seeks out new ideas, technologies, and methodologies to improve AI/ML solutions and drive the organization forward. Attention to Detail: Pays close attention to all aspects of the work, ensuring accuracy and thoroughness in data analysis, model development, and documentation. Effective Communication: Clearly and effectively communicates complex technical concepts to non-technical stakeholders, ensuring understanding and alignment across the organization.

Posted 11 hours ago

Apply

3.0 - 8.0 years

10 - 20 Lacs

Bengaluru

Work from Office

Naukri logo

Hiring for a FAANG company. Note: This position is part of a program designed to support women professionals returning to the workforce after a career break (9+ months career gap) About the Role: A global analytics team is seeking a Business Intelligence Engineer to drive data-backed insights, design robust reporting frameworks, and influence key business strategies across international markets. This role is ideal for professionals returning to the workforce and looking to re-engage in high-impact analytical work. You will collaborate closely with business stakeholders across geographies (Europe, US, Japan, Asia), working on payments and lending analytics. This is a high-ownership, high-impact role requiring a passion for data, a knack for storytelling through dashboards, and the ability to work independently in a fast-paced environment. Key Responsibilities: Design and maintain dashboards, reports, and metrics to support executive-level business decision-making Ensure data accuracy and integrity across tools, dashboards, and reporting pipelines Use SQL, Excel, and scripting languages (e.g. Python, R, Java) for deep-dive analysis Develop BI tools and data visualizations using platforms like Tableau, AWS QuickSight, Looker, etc. Analyze business trends and apply statistical rigor (t-tests, chi-squared tests, regressions, forecasting) to derive insights Lead alignment and standardization of key metrics across global BI teams Model data and metadata to support robust analytics infrastructure Automate manual reporting efforts to enhance operational efficiency Work with cross-functional teams to recommend data-driven growth strategies Present insights and narratives to stakeholders including global leaders and executives A Day in the Life: Define and refine performance metrics, reports, and insights for international payment systems Drive analytical alignment with global BI leaders and executive stakeholders Lead deep dives into metrics influencing revenue, signups, and operational performance Own VP- and Director-level reporting initiatives and decision-support analysis Collaborate across regions to deliver unified and actionable analytics strategies Basic Qualifications: 2+ years of experience in data analytics using Redshift, Oracle, NoSQL, or similar data sources Strong SQL skills for data retrieval and analysis Proficiency in data visualization using Tableau , QuickSight , Power BI , or similar tools Comfort with scripting languages like Python , Java , or R Experience applying statistical techniques to real-world data problems Preferred Qualifications: Masters degree or other advanced technical degree Experience with data modeling and data pipeline architecture Strong grasp of statistical analysis techniques, including correlation analysis and hypothesis testing Top 10 Must-Have Skills: Advanced SQL Data Visualization (Tableau, QuickSight, Power BI, Looker) Statistical Analysis (t-test, Chi-squared, Regression) Scripting (Python / R / Java) Redshift / Oracle / NoSQL Databases Dashboard & Report Development Data Modeling & Pipeline Design Cross-functional Global Collaboration Business Metrics & KPI Definition Executive-Level Reporting

Posted 4 days ago

Apply

8.0 - 13.0 years

14 - 24 Lacs

Bengaluru

Hybrid

Naukri logo

Job Description Are you interested in joining a dynamic global team that has a great impact on the Collins' organization? If so, then this is the job for you! The Global Technical Solutions (GTS) organization brings employees together across Collins that work on Electronic Computer-Aided Design (ECAD), Mechanical Computer-Aided Design (MCAD), Technical Documentation and Technical Excellence. As a Data Scientist specializing in process improvements for Mechanical Computer-Aided Design (MCAD), Electrical Computer-Aided Design (ECAD), and Technical Publications, you will be integral to enhancing the efficiency, accuracy, and consistency of design and documentation processes within GTS. Leveraging data analytics and automation, youll develop data-driven solutions to streamline workflows, reduce errors, and improve quality across mechanical and electrical design outputs as well as technical publications. Collaborating closely with MCAD and ECAD design teams, technical authors, and cross-functional stakeholders, youll contribute to advanced process optimization and best practices for aerospace engineering documentation. What YOU will do: Develop strategy on data in collaboration with senior leadership team, technology team and Tools leader by implementing innovative data technology to solve difficult business problems. Build and maintain data pipelines to collect, process, and store large volumes of data. Develop ETL/ELT pipelines to move and transform data. Ensure data systems are scalable, fast, and reliable. Work with big data tools like Data Lakes, Apache Spark, Kafka, and cloud platforms. Designing, creating and maintaining dashboards, reports, and other data visualizations for various departments within an organization by developing and implementing data visualization standards and guidelines, and providing technical support and training to other team members on data visualization tools and techniques by staying on top of latest data visualization trends and techniques. Data Analysis & Process Insights for Portfolio of programs: Collect, clean, and analyze data from MCAD and ECAD systems, technical documentations, finance, HR and related tools. Identify inefficiencies, error trends, and bottlenecks across the design-to-documentation lifecycle to drive improvements in process quality and speed. Workflow Optimization for Cross-Disciplinary Design & Documentation: Identify improvement opportunities within product areas of MCAD, ECAD, Technical Documentation and Technical Excellence processes. Develop and implement data-driven solutions to reduce manual steps, improve cross-functional integration, and enhance document consistency and accuracy. Dashboarding & Visualization for Process Monitoring: Build and maintain dashboards to provide real-time monitoring of key metrics for MCAD, ECAD, Technical Documentation and Technical Excellence processes. Enable data-driven insights into process performance, workload management, and error trends for continuous improvement Cross-Functional Collaboration: Work closely with MCAD, ECAD, and technical documentation teams to understand their challenges and requirements. Serve as a data-driven consultant to technical authors, design engineers, and quality teams, ensuring the integration of analytics solutions aligns with the needs of each function. Documentation & Reporting: Document analytical methods, model specifications, and process improvements. Provide regular impact reports and key insights for continuous process optimization across MCAD, ECAD, and technical publications. Process Improvement: Identify and implement improvements using CORE events and engagements. Travel up to 10% of the time per business needs. Qualifications You Must Have: BE/B.Tech or MS/M.Tech or Phd with 8 to 13 Years of experience. 8+ years of experience in data engineering, data science, process improvement, or analytics, preferably within aerospace or a similar industry with a focus on design and documentation 8+ years experience in programming languages and relevant experience developing front end and backend tools (eg, Python, R, SQL, Cloud Platforms, PowerApps, PowerBI, Tableau, etc.) for data engineering, analysis and visualization. 8+ years’ experience creating data visualizations to articulate messaging to executive leadership using tools such as Power BI, Tableau, Alteryx etc. Qualifications We Prefer: Bachelor’s or master’s degree in data science, Engineering, Computer Science, Business or a related field, with a focus on design processes, technical documentation, or process optimization Employee in Global Technical Solutions, effective January 2025 Awareness of MCAD and ECAD tools (e.g., CATIA, SolidWorks, Altium, or Siemens EDA) Ability to explain data-driven insights to both technical and non-technical audiences Ability to collaborate effectively in cross-functional teams and drive alignment across disciplines What We Offer: Some of our competitive benefits package includes: Transportation facility. Group Term Life Insurance. Group Health Insurance. Group Personal Accident Insurance. Entitled for 18 days of vacation and 12 days of sick leave annually. Employee scholar program Work life balance Car lease program National Pension Scheme LTA Fuel & Maintenance /Driver wages And more! Nothing matters more to Collins Aerospace than our strong ethical and safety commitments. As such, all India positions require a background check, which may include a drug screen. At Collins, the paths we pave together lead to limitless possibility. And the bonds we form – with our customers and with each other -- propel us all higher, again and again. Apply now and be part of the team that’s redefining aerospace, every day

Posted 4 days ago

Apply

3.0 - 8.0 years

18 - 30 Lacs

Hyderabad / Secunderabad, Telangana, Telangana, India

On-site

Foundit logo

Create data cleaning reporting solutions, study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. Contribute to process improvements and development of programming standards using R. Minimum of 3 years of experience in R programming including familiarity with R packages (e.g., Tidyverse, ggplot2, survival) Familiarity with data visualization techniques and tools in R. Write clean, efficient, and well-documented R code, conduct R code reviews, and R programming validation. Strong proficiency in R programming, including data manipulation, statistical analysis, and data visualization. Experience with POSIT Workbench, Connect, and/or Package Manager. Good to have python & SQL knowledge Develop library of reusable checks working closely with data engineers and config specialists. Collaborate with cross functional teams in creating business rules and automated check library across TAU. Serves as a technical resource for creating data analytics to help with data validation and cleaning. Provide programming expertise for data cleaning to efficiently ensure high quality data. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable.

Posted 4 days ago

Apply

7.0 - 10.0 years

15 - 20 Lacs

Hyderabad / Secunderabad, Telangana, Telangana, India

On-site

Foundit logo

Create data cleaning reporting solutions, study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. Contribute to process improvements and development of programming standards using R. Minimum of 3 years of experience in R programming including familiarity with R packages (e.g., Tidyverse, ggplot2, survival) Familiarity with data visualization techniques and tools in R. Write clean, efficient, and well-documented R code, conduct R code reviews, and R programming validation. Strong proficiency in R programming, including data manipulation, statistical analysis, and data visualization. Experience with POSIT Workbench, Connect, and/or Package Manager. Good to have python & SQL knowledge Develop library of reusable checks working closely with data engineers and config specialists. Collaborate with cross functional teams in creating business rules and automated check library across TAU. Serves as a technical resource for creating data analytics to help with data validation and cleaning. Provide programming expertise for data cleaning to efficiently ensure high quality data. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable.

Posted 4 days ago

Apply

7.0 - 12.0 years

15 - 20 Lacs

Mumbai, Maharashtra, India

On-site

Foundit logo

Minimum 2 years of experience for Senior Statistical programming is a mandate. Having strong understanding of SAS Base/Macros/GTL and experience of SDTM/ADAM/TLFs creation as per CDISC standards. Associate should be able to manage communication with customers, track and ensure quality on-time deliverables Implementation of process improvements for enhanced quality and efficiency gains, institute robust processes to ensure consistent outcomes with quicker turn-around time. Experience in SDTM, ADAM & TLF Experience in EDC & VEEVA Expertise in Base and Advance SAS programming. Expertise in developing macros and developing generalized programs. Development of the complex SAS macros that can be utilized across TAs in multiple studies, review of the developed SAS programs. Expertise in SQL, SAS Macros, Functions, SAS GTL, SAS, and other advanced SAS components. Mentor and supervise developers to build generic standard macros, generic listings or other ad-hoc requests in accordance with business requirements Key Skills: a) Minimum 2 years of Experience in SDTM, ADAM & TLF is must b) Any Graduate

Posted 4 days ago

Apply

7.0 - 12.0 years

15 - 20 Lacs

Nagpur, Maharashtra, India

On-site

Foundit logo

Minimum 2 years of experience for Senior Statistical programming is a mandate. Having strong understanding of SAS Base/Macros/GTL and experience of SDTM/ADAM/TLFs creation as per CDISC standards. Associate should be able to manage communication with customers, track and ensure quality on-time deliverables Implementation of process improvements for enhanced quality and efficiency gains, institute robust processes to ensure consistent outcomes with quicker turn-around time. Experience in SDTM, ADAM & TLF Experience in EDC & VEEVA Expertise in Base and Advance SAS programming. Expertise in developing macros and developing generalized programs. Development of the complex SAS macros that can be utilized across TAs in multiple studies, review of the developed SAS programs. Expertise in SQL, SAS Macros, Functions, SAS GTL, SAS, and other advanced SAS components. Mentor and supervise developers to build generic standard macros, generic listings or other ad-hoc requests in accordance with business requirements Key Skills: a) Minimum 2 years of Experience in SDTM, ADAM & TLF is must b) Any Graduate

Posted 4 days ago

Apply

5.0 - 10.0 years

15 - 20 Lacs

Mumbai, Maharashtra, India

On-site

Foundit logo

Create data cleaning reporting solutions, study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. Contribute to process improvements and development of programming standards using R. Minimum of 3 years of experience in R programming including familiarity with R packages (e.g., Tidyverse, ggplot2, survival) Familiarity with data visualization techniques and tools in R. Write clean, efficient, and well-documented R code, conduct R code reviews, and R programming validation. Strong proficiency in R programming, including data manipulation, statistical analysis, and data visualization. Experience with POSIT Workbench, Connect, and/or Package Manager. Good to have python & SQL knowledge Develop library of reusable checks working closely with data engineers and config specialists. Collaborate with cross functional teams in creating business rules and automated check library across TAU. Serves as a technical resource for creating data analytics to help with data validation and cleaning. Provide programming expertise for data cleaning to efficiently ensure high quality data. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable.

Posted 4 days ago

Apply

5.0 - 10.0 years

15 - 20 Lacs

Nagpur, Maharashtra, India

On-site

Foundit logo

Role & responsibilities: 1.Create data cleaning reporting solutions study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. 2. Develop library of reusable checks working closely with data engineers and config specialists. 3. Collaborate with cross functional teams in creating business rules and automated check library across TAU. 4. Serves as a technical resource for creating data analytics to help with data validation and cleaning. 5. Provide programming expertise for data cleaning to efficiently ensure high quality data. 6. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. 7. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. 8. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. 9. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) 10. Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable 11. Technical skills and experience using relational databases (e.g. MS SQL Server, Redshift, or Oracle), 12. EDC build and/or clinical data management experience preferred 13. Experience in AWS preferred 14. Experience setting up automated validation is preferred 15. Knowledge of Data lake and creating interactive Dashboards are preferred 16. Proficiency in the use of Microsoft Office Suite of tools (Outlook, Word, Excel, etc.)

Posted 4 days ago

Apply

5.0 - 10.0 years

15 - 20 Lacs

Nagpur, Maharashtra, India

On-site

Foundit logo

Programming, specifically J Review, SQL, R and GitHub. Good to have Python. Create data cleaning reporting solutions study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. Develop library of reusable checks working closely with data engineers and config specialists. Collaborate with cross functional teams in creating business rules and automated check library across TAU. Serves as a technical resource for creating data analytics to help with data validation and cleaning. Provide programming expertise for data cleaning to efficiently ensure high quality data. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable Technical skills and experience using relational databases (e.g. MS SQL Server, Redshift, or Oracle), EDC build and/or clinical data management experience preferred Experience in AWS preferred Experience setting up automated validation is preferred Knowledge of Data base and creating interactive Dashboards are preferred Proficiency in the use of Microsoft Office Suite of tools (Outlook, Word, Excel, etc.)

Posted 4 days ago

Apply

5.0 - 10.0 years

15 - 20 Lacs

Mumbai, Maharashtra, India

On-site

Foundit logo

1.Create data cleaning reporting solutions study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. 2. Develop library of reusable checks working closely with data engineers and config specialists. 3. Collaborate with cross functional teams in creating business rules and automated check library across TAU. 4. Serves as a technical resource for creating data analytics to help with data validation and cleaning. 5. Provide programming expertise for data cleaning to efficiently ensure high quality data. 6. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. 7. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. 8. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. 9. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) 10. Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable 11. Technical skills and experience using relational databases (e.g. MS SQL Server, Redshift, or Oracle), 12. EDC build and/or clinical data management experience preferred 13. Experience in AWS preferred 14. Experience setting up automated validation is preferred 15. Knowledge of Data lake and creating interactive Dashboards are preferred 16. Proficiency in the use of Microsoft Office Suite of tools (Outlook, Word, Excel, etc.)

Posted 4 days ago

Apply

5.0 - 10.0 years

15 - 20 Lacs

Mumbai, Maharashtra, India

On-site

Foundit logo

Role & responsibilities: 1.Create data cleaning reporting solutions study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. 2. Develop library of reusable checks working closely with data engineers and config specialists. 3. Collaborate with cross functional teams in creating business rules and automated check library across TAU. 4. Serves as a technical resource for creating data analytics to help with data validation and cleaning. 5. Provide programming expertise for data cleaning to efficiently ensure high quality data. 6. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. 7. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. 8. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. 9. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) 10. Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable 11. Technical skills and experience using relational databases (e.g. MS SQL Server, Redshift, or Oracle), 12. EDC build and/or clinical data management experience preferred 13. Experience in AWS preferred 14. Experience setting up automated validation is preferred 15. Knowledge of Data lake and creating interactive Dashboards are preferred 16. Proficiency in the use of Microsoft Office Suite of tools (Outlook, Word, Excel, etc.)

Posted 4 days ago

Apply

5.0 - 7.0 years

15 - 25 Lacs

Bengaluru

Work from Office

Naukri logo

i) Write and optimize SQL queries, primarily using MS SQL Server and AWS Athena, to extract and analyze data. ii) Investigate and research data discrepancies, providing clear documentation of findings and resolutions. iii) Collaborate with business and technical teams to translate data insights into actionable recommendations. iv) Develop reports and dashboards using Tableau or similar visualization tools (Nice to have). v) Communicate insights and findings effectively in English (verbal and written).

Posted 6 days ago

Apply

3.0 - 7.0 years

10 - 15 Lacs

Ahmedabad

Remote

Naukri logo

To use advanced data analysis for providing actionable insights that support data-driven decision-making within teams by utilizing R programming language, developing automation tools for oversight and standardized reporting templates to enhance the efficiency and consistency of information across the organization. Role & responsibilities Data Analytics and Reporting Design and develop detailed reports leveraging various databases, using R language to generate meaningful insights. Proactively develop automation tools, standardized reporting tools & templates (including slides) for use by stakeholders to enable consistent presentation of information across the teams. Create visualizations and trend analysis to identify patterns in study performance or other relevant areas. Monitor and document any bugs or issues related to the reporting tools and work with the concerned teams for timely resolution. Stakeholder Collaboration Collaborate with cross-functional teams proactively to gather and understand reporting requirements, translating them into programming Specifications. Assist leadership teams in data-driven decision making by providing actionable insights such as budget projections, resource allocation, Full Time Equivalent (FTE) predictions etc. through various reports. Act as the bridge between data teams and other departments, gathering requirements and output Specifications for new analytical tools or reports. Utilize data to perform analytics, thereby predicting future requirements, provide real-time analysis of study quality metrics, site performance metrics and generate reports for Source Data Verification (SDV). Analytics Project Management Maintain clear documentation of methodologies, data processes, and report-generation protocols to ensure transparency and reproducibility of analyses. Assist in the development of performance metrics based on communicated requirements, contributing to the overall efficiency and effectiveness of clinical operations. Develop clear and concise communication around project progress, key findings, and recommendations to various stakeholders. Keep track of the number of projects progress and provide regular reports to Associate Director, Biostatistics and Analytics. Policies, Processes & Procedures Conduct day to day activities & follow all relevant policies, processes, standard operating procedures and instructions so that work is carried out in a controlled and consistent manner. Assist in the updating departmental and role Specific programming standards, Standard Operating Procedures (SOPs), Working Procedural Documents (WPDs), and templates to reflect current practices and regulatory requirements. Implement new initiatives/ projects as per established policies to meet future requirements of the function. Comply with the company's Quality and Information Security Management Systems and applicable national and international legislation, including legislation for data protection Key interactions Clinical operations Biostatisticians Clinical Informatics Clinical Data Management Project Management Functional / Behavioral Skills required to execute the role R Programming Automation and Tool Development Data Analysis and Interpretation Documentation and Reporting Project Management Communication Detail orientation Problem solving Time management Hands on experience in data analysis, preferably within a clinical research or healthcare setting

Posted 1 week ago

Apply

5.0 - 9.0 years

17 - 25 Lacs

Pune, Gurugram, Bengaluru

Hybrid

Naukri logo

Desired Skills and experience Candidates should have a B.E./B.Tech/MCA/MBA in Finance, Information Systems, Computer Science or a related field Strong experience in R programming and package development Proficiency with GitHub and unit testing frameworks. Strong documentation and communication skills. A background or work experience in biostatistics or a similar discipline (Preferred). Expert knowledge in Survival Analysis (Preferred) Statistical model deployment, and end-to-end MLOps is nice to have. Having worked extensively on cloud infrastructure, preferably Databricks and Azure. Shinydevelopment is nice to have. Can work with customer stakeholders to understand business processes and workflows and can design solutions to optimize processes via streamlining and automation. DevOps experience and familiarity with software release process. Familiar with agile delivery methods. Excellent communication skills, both written and verbal Extremely strong organizational and analytical skills with strong attention to detail Strong track record of excellent results delivered to internal and external clients Able to work independently without the needs for close supervision and also collaboratively as part of cross-team efforts Experience with delivering projects within an agile environment

Posted 2 weeks ago

Apply

4.0 - 9.0 years

8 - 12 Lacs

Pune, Bengaluru, Mumbai (All Areas)

Work from Office

Naukri logo

We are seeking a Clinical Programmer with expertise in JReview listings, dashboards, and visualizations to support clinical data review, safety monitoring, data validation and clinical data visualization, analysis, and reporting for clinical trials. Required Candidate profile 4+ Years of experience. MS SQL Server and R, R packages e.g., Tidyverse, ggplot2, dplyr, Pipeline creation, or Oracle, SAS, Python, Elluminate, BOXI, Spotfire, Tableau, Macros, RShiny, JReview, Plotly

Posted 2 weeks ago

Apply

5.0 - 10.0 years

12 - 18 Lacs

Pune, Chennai, Bengaluru

Work from Office

Naukri logo

Hello, We are looking for Data Scientist role to work with one of the American Multinational based at Bangalore/Chennai/Pune/Gurgaon/Hyderabad. Key Responsibilities: • R Programming Transition: Analyze and understand the existing application to effectively transition it to R. Re-write or optimize existing code to ensure efficiency and scalability in R. Transform R models to Python as needed. • R Shiny Dashboard Development: Design, develop, and implement interactive and visually appealing R Shiny Dashboards. Integrate data sources and ensure real-time data updates in the dashboards. Collaborate with stakeholders to gather requirements and provide dashboard solutions that meet business needs. • Cloud Server Hosting: Set up and configure AWS for hosting R Shiny Dashboards. Ensure secure, scalable, and reliable AWS hosting environments. • Project Leadership and Collaboration: Collaborate with cross-functional teams including data scientists, analysts, and IT staff. Required Skills and Qualifications: • Bachelors or Masters degree in Computer Science, Data Science, Statistics, or related field. • Proven experience in R programming and developing R Shiny applications and strong python skills. • Strong understanding of data manipulation and visualization in R. • Experience with cloud platforms such as AWS, Google Cloud, or Azure, specifically in deploying R Shiny Dashboards. • Familiarity with DevOps practices, including CI/CD pipelines and version control (e.g., Git). • Excellent problem-solving skills and ability to work independently and as part of a team. • Strong communication skills to effectively collaborate with technical and non-technical stakeholders. Notice - Immediate/15 days/1 Month Interested candidates may apply on saritha.nair@reynasolutions.com Regards, Sarita Nair

Posted 2 weeks ago

Apply

8.0 - 13.0 years

3 - 6 Lacs

Bengaluru

Work from Office

Naukri logo

Location Bengaluru : We are seeking a highly skilled and motivated Data Engineer to join our dynamic team. The ideal candidate will have extensive experience in data engineering, with a strong focus on Databricks, Python, and SQL. As a Data Engineer, you will play a crucial role in designing, developing, and maintaining our data infrastructure to support various business needs. Key Responsibilities Develop and implement efficient data pipelines and ETL processes to migrate and manage client, investment, and accounting data in Databricks Work closely with the investment management team to understand data structures and business requirements, ensuring data accuracy and quality. Monitor and troubleshoot data pipelines, ensuring high availability and reliability of data systems. Optimize database performance by designing scalable and cost-effective solutions. What s on offer Competitive salary and benefits package. Opportunities for professional growth and development. A collaborative and inclusive work environment. The chance to work on impactful projects with a talented team. Candidate Profile Experience: 8+ years of experience in data engineering or a similar role. Proficiency in Apache Spark. Databricks Data Cloud, including schema design, data partitioning, and query optimization Exposure to Azure. Exposure to Streaming technologies. (e.g Autoloader, DLT Streaming) Advanced SQL, data modeling skills and data warehousing concepts tailored to investment management data (e.g., transaction, accounting, portfolio data, reference data etc). Experience with ETL/ELT tools like snap logic and programming languages (e.g., Python, Scala, R programing). Familiarity workload automation and job scheduling tool such as Control M. Familiar with data governance frameworks and security protocols. Excellent problem-solving skills and attention to detail. Strong communication and collaboration skills. Education Bachelor s degree in computer science, IT, or a related discipline. Not Ready to Apply Join our talent pool and we'll reach out when a job fits your skills.

Posted 2 weeks ago

Apply

10.0 - 15.0 years

35 - 55 Lacs

Gurugram, Chennai, Bengaluru

Hybrid

Naukri logo

Role: Data Science Leader Experience: 10-15 years of overall experience, with at least 5+ years in Data science roles along with 24 years in a Leadership or Managerial capacity. Technical Skills: Data Science & Machine Learning Deep understanding of statistical modeling, predictive analytics, clustering, NLP, and time series forecasting. Strong grasp of model evaluation, fairness, explainability, and business alignment. Programming Proficient in Python (or R) with experience in data science libraries like pandas, NumPy, scikit-learn, TensorFlow, or PyTorch. Ability to read and review code, debug, and advise on best practices. Cloud Computing Working knowledge of at least one major cloud platform (AWS, GCP, Azure). Experience with cloud-native tools for data storage (S3, BigQuery), compute (EC2, GKE, Lambda), and ML services (SageMaker, Vertex AI, Azure ML). Data Engineering Fundamentals Understanding of data pipelines, ETL/ELT processes. Familiarity with tools like Airflow, dbt, Spark, SQL. Data Visualization Experience with dashboards and reporting tools (Tableau, Power BI, Looker, or custom visualizations using Plotly/Altair). Skilled in transforming complex outputs into clear, compelling narratives for non-technical stakeholders. MLOps / Deployment Familiarity with modern ML development and deployment practices. Basic understanding of deploying models to production, CI/CD pipelines, and monitoring. Generative AI Familiarity with GenAI concepts, including large language models (LLMs), embeddings, prompt engineering, and RAG pipelines. Familiarity with tools and APIs like OpenAI, Hugging Face, LangChain. Leadership & People Management Team Management Experience leading and mentoring teams of 510 individuals across varying levels. Proven track record of building and scaling Data science teams, delivering impactful projects Conduct performance reviews, manage career growth, and foster a healthy team culture. Cross-Functional Collaboration Proven ability to work closely with product, engineering, marketing, and business teams. Hiring & Talent Development Skilled in identifying top talent, conducting interviews, onboarding, and team capability building. Project & Stakeholder Management Project Management Experienced in managing multiple projects simultaneously. Comfortable with Agile methodologies, sprint planning, and delivery tracking. Stakeholder Communication Translating technical insights into business terms. Communicates technical insights clearly to senior executives. Problem Solving & Scope Management Ability to break down ambiguous business problems into solvable components. Define scope and ensure projects align with business impact. Strategic Thinking Aligns team objectives with company vision and business goals. Drives roadmap planning, and long-term capability building. Business Acumen Domain Knowledge Deep understanding of any one business vertical among the following e.g., telecom, BFSI, fintech, e-commerce, healthcare etc. Impact Orientation Focus on delivering measurable business value from data science efforts. Strong grasp of metrics, KPIs, and ROI-driven thinking. Strategic Thinking Ability to align data science efforts with long-term business goals. Soft Skills Exceptional communication, both verbal and written. Decision-Making: Balanced between data, intuition, team input and strategic vision. Empathy & Emotional Intelligence: Especially for managing team dynamics and motivation. Adaptability: In the face of changing priorities or business goals and organizational change.

Posted 2 weeks ago

Apply

10.0 - 20.0 years

15 - 30 Lacs

Bengaluru

Hybrid

Naukri logo

Role : Name: Open Source Transition Lead Office : Client Deployed Mode : Hybrid Location : Banglore. Note: This role is an opportunity to Client Deployed model under EY payrolls. This is a initial duration of 6 months role, the tenure is subject to extension basis performance. Job Responsibility: • Transition from Sas to open source • Adopting R as open-source language Python • Scope is R&D wide ( 1000+ users) • Upskilling, right tools and program themselves and standard tools • Systems have Sas code and need to be decommissioned Big tech component • Business and Tech change • User engagement, comms, and training and getting voice of Customer (making training program. Please share ur updated resume to sai.m@in.ey.com

Posted 2 weeks ago

Apply

5.0 - 10.0 years

15 - 20 Lacs

Pune, Mumbai (All Areas)

Work from Office

Naukri logo

Hi , We are hiring for Leading ITES Company for SAS Programmer & CDB Programmer ( J Reviewer/Python/SAS) Role & responsibilities: 1.Create data cleaning reporting solutions study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. 2. Develop library of reusable checks working closely with data engineers and config specialists. 3. Collaborate with cross functional teams in creating business rules and automated check library across TAU. 4. Serves as a technical resource for creating data analytics to help with data validation and cleaning. 5. Provide programming expertise for data cleaning to efficiently ensure high quality data. 6. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. 7. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. 8. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. 9. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) 10. Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable 11. Technical skills and experience using relational databases (e.g. MS SQL Server, Redshift, or Oracle), 12. EDC build and/or clinical data management experience preferred 13. Experience in AWS preferred 14. Experience setting up automated validation is preferred 15. Knowledge of Data lake and creating interactive Dashboards are preferred 16. Proficiency in the use of Microsoft Office Suite of tools (Outlook, Word, Excel, etc.) To Apply, WhatsApp 'Hi' @ 9151555419 Follow the Steps Below: >Click on Start option to Apply and fill the details >Select the location as Other ( to get multiple location option ) a) To Apply for above Job Role ( Mumbai ) Type : Job Code # 83 b) To Apply for above Job Role ( Pune ) Type : Job Code # 84

Posted 3 weeks ago

Apply

3.0 - 8.0 years

15 - 20 Lacs

Nagpur, Pune, Bengaluru

Work from Office

Naukri logo

Hi, We are hiring for ITES Company for SAS + CDB Programmer Role. Job description: Programming, specifically J Review, SQL, R and GitHub. Good to have Python. • Create data cleaning reporting solutions study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. • Develop library of reusable checks working closely with data engineers and config specialists. • Collaborate with cross functional teams in creating business rules and automated check library across TAU. • Serves as a technical resource for creating data analytics to help with data validation and cleaning. • Provide programming expertise for data cleaning to efficiently ensure high quality data. • Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. • Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. • Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. • Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) • Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable • Technical skills and experience using relational databases (e.g. MS SQL Server, Redshift, or Oracle), • EDC build and/or clinical data management experience preferred • Experience in AWS preferred • Experience setting up automated validation is preferred • Knowledge of Data base and creating interactive Dashboards are preferred • Proficiency in the use of Microsoft Office Suite of tools (Outlook, Word, Excel, etc.) Key Skills: a) Minimum 3 years of Experience in SAS CDB Programmers with R programming. b) Good to have Python knowledge c) Any Graduate To Apply, WhatsApp 'Hi' @ 9151555419 Follow the Steps Below: >Click on Start option to Apply and fill the details >Select the location as Other ( to get multiple location option ) a) To Apply for above Job Role ( Pune ) Type : Job Code # 84 b) To Apply for above Job Role ( Nagpur ) Type : Job Code # 85 c) To Apply for above Job Role ( Bangalore ) Type : Job Code # 86

Posted 3 weeks ago

Apply

4.0 - 9.0 years

0 - 0 Lacs

Bengaluru

Work from Office

Naukri logo

Python developer+ R Programming 4 to 7 years Bangalore (Whitefiled) WFO CTC:-19 LPA NP:-Immediate (5 days WFO) 2 yrs relevant exp should be have in R Programing

Posted 3 weeks ago

Apply

3.0 - 8.0 years

18 - 30 Lacs

Pune, Mumbai (All Areas)

Work from Office

Naukri logo

Hi We are looking to hire for SAS+R Programming . Please go through the JD and Apply Roles & Responsibilities: Create data cleaning reporting solutions, study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. Contribute to process improvements and development of programming standards using R. Minimum of 3 years of experience in R programming including familiarity with R packages (e.g., Tidyverse, ggplot2, survival) Familiarity with data visualization techniques and tools in R. Write clean, efficient, and well-documented R code, conduct R code reviews, and R programming validation. Strong proficiency in R programming, including data manipulation, statistical analysis, and data visualization. Experience with POSIT Workbench, Connect, and/or Package Manager. Good to have python & SQL knowledge Develop library of reusable checks working closely with data engineers and config specialists. Collaborate with cross functional teams in creating business rules and automated check library across TAU. Serves as a technical resource for creating data analytics to help with data validation and cleaning. Provide programming expertise for data cleaning to efficiently ensure high quality data. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable. To Apply WhatsApp 'Hi' @ 9151555419 and Follow the Steps Below: a) For Position in Mumbai Search : Job Code # 373 b) For Position in Pune Search : Job Code # 374

Posted 3 weeks ago

Apply

7.0 - 10.0 years

10 - 20 Lacs

Hyderabad, Bengaluru

Work from Office

Naukri logo

Hi We are hiring for Leading ITES Company for SAS+R Programmer Profile. Role & responsibilities: a) Any Graduate b) Minimum of 3 years of experience in SAS + R programming c) Familiarity with data visualization techniques and tools in R d) CDISC knowledge & experience Roles & Responsibilities: Create data cleaning reporting solutions, study status metric, and monitoring reports and visualizations required per standard and study specific data review plans. Contribute to process improvements and development of programming standards using R. Minimum of 3 years of experience in R programming including familiarity with R packages (e.g., Tidyverse, ggplot2, survival) Familiarity with data visualization techniques and tools in R. Write clean, efficient, and well-documented R code, conduct R code reviews, and R programming validation. Strong proficiency in R programming, including data manipulation, statistical analysis, and data visualization. Experience with POSIT Workbench, Connect, and/or Package Manager. Good to have python & SQL knowledge Develop library of reusable checks working closely with data engineers and config specialists. Collaborate with cross functional teams in creating business rules and automated check library across TAU. Serves as a technical resource for creating data analytics to help with data validation and cleaning. Provide programming expertise for data cleaning to efficiently ensure high quality data. Ensure accurate delivery by creating macros for Standard listings for the department based on specifications. Develop programs that generate CRF and Query metrics, including ad-hoc Sponsor and internal programming requests. Ability to set up and configure analytics for coding reports, safety review and any risk-based monitoring activities. Operational experience in pharmaceutical drug development with significant direct exposure to clinical development and data operations required for the reporting of clinical trial data (e.g. data review, regulatory submissions, safety updates, etc.) Understanding of regulatory requirements and relevant data standards; CDISC knowledge and experience are preferable. To Apply, WhatsApp 'Hi' @ 9151555419 Follow the Steps Below: >Click on Start option to Apply and fill the details >Select the location as Other ( to get multiple location option ) a) To Apply for above Job Role ( Bangalore ) Type : Job Code # 375 b) To Apply for above Job Role ( Hyderabad ) Type : Job Code # 376

Posted 3 weeks ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies