Java Programming, Python Programming, AWS, GoLang, Apache Hadoop, Apache Kafka, Scala Programming, Google BigQuery, Apache Hive, Vertica, Snowflake, Apache Airflow, Docker, Kubernetes-K8s, Google Cloud Platform (GCP), Apache Impala, Amazon Athena, Amazon RedShift, Amazon EMR, Amazon Elastic Compute Cloud-EC2, Amazon RDS, Tez, Calcite, Parquet, Arrow, Exxasol, Drill, Google Borg, Mesos, Amazon Simple Storage Service (S3), Apache Flink
We are looking for a Principal Data Engineer to build reliable, distributed data pipelines and intuitive data products that allow our stakeholders to easily leverage data in an effective manner. As part of this team, you will work on diverse data technologies such as Snowflake, Airflow, Spark, Presto, Flink, Kafka & others to build insightful, scalable and robust data pipelines.
The ideal candidate will have a strong background in distributed data processing, have great and demonstrable data intuition, best of breed data modelling, and share our passion for continuously improving the ways we use data to make the Tableau’s Data Culture better.
This role will be a software engineering leader part of the engineering team in the Enterprise Data Platform And Analytics organization (EDPA) reporting to the Data Engineering Manager.
Who you are:
10+ years experience in building large scale products, data platforms, distributed systems in a high caliber environment
You have a strong background in distributed data processing, software engineering components, and data modeling concepts. Passionate about intuitive data models and an expert in distributed data processing patterns
You LOVE data of all sorts, big and small! You enjoy helping teams push the boundaries of what business insights can be extracted out of data
You excel at taking vague requirements and crystallizing them into scalable data solutions. You tackle large technical projects/problems that are not clearly defined. You anticipate roadblocks and have strategies to de-risk timelines. You orchestrate work that spans multiple teams and keep your stakeholders informed
You are always on the lookout for opportunities to simplify, automate tasks, and build reusable components that can be leveraged across multiple use cases and teams
You have expertise and deep skills with complex SQL - optimized queries and tuning for performance
You have advanced knowledge of at least one programming language, and are happy to learn more. One of your core languages is Java, Python, Go, and Scala.
You identify and solve major architectural problems by going deep in your field or broad across different teams. Extend, improve, or, when needed, build solutions to address architectural gaps or technical debt
You have demonstrated expertise in one of the big data analytics technologies we currently use such as Apache Hadoop (HDFS and YARN), Apache Hive, Impala, Drill, Spark, Tez, Presto, Calcite, Parquet, Arrow, etc. Under the hood experience with similar systems such as Snowflake, Exxasol, Vertica, Apache Impala, Drill, Google Borg, Google BigQuery, Amazon EMR, Amazon RedShift, Docker, Kubernetes, Mesos etc. Conceptually familiar with Public Cloud Resources such AWS cloud resources (S3, EC2, RDS, Athena, etc) or GCP Cloud Resources
You understand requirements beyond the written word. You understand business impact as well as technical impact and you are able to communicate with stakeholders in business units.
You have strong communication skills to effectively partner with data scientists and engineering stakeholders. You love to innovate and push partners to deliver on novel metrics and solutions
Your attention to details leads to a delightful user experience.
You understand the Data Lifecycle and concepts such as lineage, governance, privacy, retention, anonymity, etc
Act as a thought partner and subject matter expert to the Enterprise data platform engineering team, understand the challenges, and make opinionated recommendations that empower all of us with data solutions to efficiently scale Tableau infrastructure and tools
Maintain and rethink existing datasets and pipelines to service a wider variety of use cases
Enable smart analytics by building robust, reliable, and useful data sets that can power various analytic techniques like regression, classification, clustering etc
You have an eye for detail and like to spark joy in internal partners with high-quality data products that are well documented, modeled, and easy to understand
You relate to and embody many of the aspects of the self-service Tableau Data Culture . You love to operate independently while collaborating and giving/receiving strong, candid feedback to your fellow team members
You believe that you can achieve more on a team that the whole is greater than the sum of its parts. You rely on others' candid feedback for continuous improvement
You love to mentor other data engineers and lead the community of data engineers at Tableau to establish best of breed best practices and communities.
You are motivated. You do good work, and lots of it, and not because your boss is watching… You are and espouse the company builder mindset and are focused on getting your best work done and delivering impactful results!
You are a Recruiter! Tableau hires company builders and, in this role, you will be asked to be on the constant lookout for the best talent to bring onboard to help us continue to build one of the best companies in the world!
Apply online at Tableau - JR80177
Tableau Software helps people see and understand data. Offering a revolutionary new approach to business intelligence, Tableau allows you to quickly connect, visualize, and share data with a seamless experience from the PC to the iPad. Create and publish dashboards and share them with colleagues, partners, or customers—no programming skills required.
Zürich, Switzerland
2-4 year