Senior Data Engineer - Remote

Merck & Co.
Apply Now

Job Description

New hires in office-based roles in the US & Puerto Rico will be required, subject to applicable law, to demonstrate that they have been fully vaccinated for COVID-19 or qualify for a medical or religious exemption to this vaccination requirement that can be accommodated without an undue burden to the operation. However, subject to applicable law, employees working in roles that the Company determines require routine collaboration with external stakeholders, such as employees in health services, customer facing commercial, or research based roles, will be required to be fully vaccinated as a condition of employment.

Attention NJ Applicants: In April 2020, our Company announced plans to consolidate our New Jersey campuses into a single New Jersey headquarters location in Rahway, NJ by 2023. Therefore, the NJ work location for this role will be temporarily based and the exact timing of the departmental move to Rahway, NJ will be communicated at a future date.

We are leveraging analytics and technology, as we invent for life on behalf of patients around the world. We are seeking those who have a passion for using data, analytics, and insights to drive decision making, that will allow us to tackle some of the world’s greatest health threats.

Within our commercial Insights, Analytics, and Data organization we are transforming to better power decision-making across our end-to-end commercialization process, from business development to late lifecycle management. As we endeavor, we are seeking a dynamic talent for the role of Data Engineer

For the Data Engineer role, we are looking for professional with experience in designing, developing, and maintaining data pipelines. We intend to make data reliable, governed, secure and available for analytics within the organization. As part of a team this role will be responsible for data management with a broad range of activities like data ingestion to cloud data lakes and warehouses, quality control, metadata management and orchestration of machine learning models. We are also forward looking and plan to bring innovations like data mesh and data fabric into our ecosystem of tools and processes

Primary Responsibilities:

  • Play a key role in the success and growth of the Data Engineering team by mentoring and playing a leadership role within the team
  • Drive innovation within Data Engineering by playing a lead role in technology decisions for the future of our data science, analysis, and reporting needs
  • Work with business partners and software engineers to gather, understand, and bridge definitions and requirements
  • Lead the design and development for highly complex and critical data projects with strict timelines
  • Improvements to team efficiency and effectiveness through implementation of data tools (self-service, data quality, etc.)
  • Design, develop and maintain data pipelines to extract data from a variety of sources and populate data lake and data warehouse
  • Develop the various data transformation rules and data modeling capabilities
  • Collaborate with Data Analyst, Data Scientists, Machine Learning Engineers to identify and transform data for ingestion, exploration, and modeling
  • Work with data governance team and implement data quality checks and maintain data catalogs
  • Use Orchestration, logging, and monitoring tools to build resilient pipelines
  • Use test driven development methodology when building ELT/ETL pipelines
  • Understand and apply concepts like data lake, data warehouse, lake-house, data mesh and data-fabric where relevant
  • Develop data models for cloud data warehouses like Redshift and Snowflake
  • Develop pipelines to ingest data into cloud data warehouses
  • Understand and be able to use different databases like Relational, Document, Graph and Key/Value
  • Analyze data using SQL
  • Use serverless AWS services like Glue, Lambda, StepFunctions
  • Use Terraform Code to deploy on AWS
  • Containerize Python code using Docker
  • Use Git for version control and understand various branching strategies
  • Build pipelines to work with large datasets using PySpark
  • Develop proof of concepts using Jupyter Notebooks
  • Work as part of an agile team
  • Create technical documentation as needed

Education:

  • Bachelor’s degree or equivalent experience in a relevant field such as Mathematics, Computer Science, Engineering, Artificial Intelligence, etc.

Knowledge and Skills:

  • 3+ years of relevant knowledge
  • Good experience with AWS services like S3, ECS, Fargate, Glue, StepFunctions, CloudWatch, Lambda, EMR
  • SQL
  • Proficient in Python, PySpark
  • Good with Git, Docker, Terraform
  • Ability to work in cross functional teams

Preferred Knowledge and Skills:

  • Any AWS developer or architect certification
  • Agile development methodology

Our Human Health Division maintains a “patient first, profits later” ideology. The organization is comprised of sales, marketing, market access, digital analytics and commercial professionals who are passionate about their role in bringing our medicines to our customers worldwide.

Company Info.

Merck & Co.

Merck & Co., Inc. is a multinational pharmaceutical company headquartered in Kenilworth, New Jersey. It is named after the Merck family, which set up Merck Group in Germany in 1668. The company does business as MSD outside the United States and Canada.

  • Industry
    Pharmaceuticals
  • No. of Employees
    66,400
  • Location
    Kenilworth, NJ, USA
  • Website
  • Jobs Posted

Get Similar Jobs In Your Inbox

Merck & Co. is currently hiring Senior Data Scientist Jobs in Kenilworth, NJ, USA with average base salary of $160,000 - $240,000 / Year.

Similar Jobs View More