Principal Hadoop Data Engineer

HCA Healthcare
Apply Now

Job Description

Job Summary:

The role requires working closely with others, frequently in a matrixed environment, and with little supervision. As a Principal Data Engineer/Architect level, the role requires 'self-starters' who are proficient in problem solving and capable of bringing clarity to complex situations. It requires contributing to strategic technical direction and system architecture approaches for individual projects and platform migrations. The culture of the organization places an emphasis on teamwork, so social and interpersonal skills are equally important as technical capability. Due to the emerging and fast-evolving nature of GCP/Big Data technology and practice, the position requires that one stay well-informed of technological advancements and be proficient at putting new innovations into effective practice.

Responsible for leading GCP development efforts, driving adoption and appropriate use of technology and consulting on internal and external development efforts to ensure code quality and sound architecture. This position that assumes the responsibility for project success and the upward development of team members technical skills. They are the development team's point of contact that must interface with business partners of varying roles ranging from technical staff to executive leadership. In addition, this candidate will have a history of increasing responsibility in a multi-role team. This position requires a candidate who can analyze business requirements, perform design tasks, construct, test, and implement cutting-edge technical data solutions with minimal supervision.

As a Principal Data Engineer/Architect, you will work closely with all team members to create a modular, scalable solution that addresses current needs, but will also serve as a foundation for future success. The position will be critical in building the team’s engineering practices in test driven development, continuous integration, and automated deployment and is a hands-on team member who actively coaches the team to solve complex problems. She / he will be responsible for the design, development, performance and support of the Cloud Platform components.

This candidate will have a record of accomplishment of participation in successful projects in a fast-paced, mixed team (consultant and employee) environment. In addition, the applicant must be willing to train and mentor other developers to prepare them for assuming the responsibilities.

General Responsibilities:

  • Responsible for building and supporting a GCP/Hadoop-based ecosystem designed for enterprise-wide analysis of structured, semi-structured, and unstructured data.
  • Bring new data sources into GCP/HDFS, transform and load to databases.
  • Lead projects in delivering the data and projects on-time
  • Closely collaborates with team members to successfully execute development initiatives using Agile practices and principles
  • Leads efforts to design, development, deploy, and support software systems
  • Experience with HL7, FHIR, and Whistle mapping.
  • Collaborates with business analysts, project lead, management and customers on requirements
  • Participates in large-scale development projects involving multiple areas outside of core team
  • Designs fit-for-purpose products to ensure products align to the customer's strategic plans and technology road maps
  • Demonstrates deep understanding and coaches’ value-based decision making and Agile principles across teams
  • Coaches team on clinical data, existing system structure, constraints and deficiencies with product
  • Shares knowledge and experience to contribute to growth of overall team capabilities
  • Participates in the deployment, change, configuration, management, administration and maintenance of deployment process and systems
  • Work closely with management, architects and other teams to develop and implement the projects.
  • Actively participate in technical group discussions and adopt any new technologies to improve the development and operations.
  • Focuses on customer satisfaction
  • Rapidly prototypes and delivers just-in-time solutions
  • Gather requirements, designs, constructs and delivers solutions with minimal team interaction
  • Works in an environment with rapidly changing business requirements and priorities
  • Demonstrates deep understanding and acts as a leader in the team’s continuous integration and continuous delivery automation pipeline
  • Work collaboratively with Data Scientists, business, and IT leaders throughout the company to understand Cloud/Big Data needs and use cases.

Education, Experience and Certifications:

  • Bachelor's Degree in computer science or related field – Required
  • Master's Degree in computer science or related field – Preferred
  • 3+ years of experience in Data Engineer – Required
  • 1+ year(s) of experience in Healthcare – Preferred
  • 10+ years of experience in Information Technology – Required
  • GCP Cloud Professional Data Architect certification – Preferred
  • GCP Cloud Professional Data Engineer certification – Preferred 

Other Required Qualifications:

A successful candidate will have:

  • Strong understanding of best practices and standards for GCP application design and implementation.
  • Two Year of hands-on experience with GCP platform and experience with many of the following components:
    • GCS, Cloud Run, Cloud Functions
    • Bigtable, Cloud SQL
    • Kafka, Pub/Sub
    • Python, Golang, Spark, Scala or Java
    • BigQuery, Dataflow, Data Fusion
    • CICD process and Logging & Monitoring
    • OpenShift, Docker
  • Experience with Unstructured Data, Real-Time Streaming with GCP
  • Ability to multitask and to balance competing priorities.
  • Requires strong practical experience in agile application development, file systems management, and DevOps discipline and practice using short-cycle iterations to deliver continuous business value.
  • Knowledge of all facets of GCP Cloud ecosystem development including ideation, design, implementation, tuning, and operational support.
  • Ability to define and utilize best practice techniques and to impose order in a fast-changing environment. Must have strong problem-solving skills.
  • Strong verbal, written, and interpersonal skills, including a desire to work within a highly-matrixed, team-oriented environment.

A successful candidate may have:

  • Experience in Healthcare Domain
  • Experience in Patient Data
  • Experience with Natural Language Processing (NLP)
  • Azure/AWS Cloud experience
  • Hands-on experience with Cloudera Distributed Hadoop (CDH)

Hardware/Operating Systems:

  • Linux, UNIX
  • GCP
  • Distributed, highly-scalable processing environments

Databases:

  • NoSQL, Hbase, Cassandra, MongoDB, Cosmos, In-memory, Columnar, other emerging technologies
  • Build Systems – TFS, Github
  • Ability to integrate tools outside of the core Cloud ecosystem

Physical Demands/Working Conditions

  • Prolonged sitting or standing at computer workstation including use of mouse, keyboard, and monitor.
  • Requires ability to provide after-hours support.
  • Occasional Travel: The job may require travel from time- to-time, but not on a regular basis.

HCA Healthcare’s Information Technology Group (ITG) delivers healthcare IT products and services to HCA Healthcare's portfolio of business and partners, including Parallon, HealthTrust and Sarah Cannon.

Company Info.

HCA Healthcare

HCA Healthcare is an American for-profit operator of health care facilities that was founded in 1968. It is based in Nashville, Tennessee, and, as of May 2020, owns and operates 186 hospitals and approximately 2,000 sites of care, including surgery centers, freestanding emergency rooms, urgent care centers and physician clinics in 21 states and the United Kingdom.

Get Similar Jobs In Your Inbox

HCA Healthcare is currently hiring Hadoop Engineer Jobs in Nashville, TN, USA with average base salary of $90,000 - $190,000 / Year.

Similar Jobs View More