Job Description

Our data engineering team has a strong focus on delivering high-quality, reliable data to platforms and people within O’Reilly as well as building high-performance, scalable and extensible systems. We are intentional in our search for teammates who are helpful, respectful, communicate openly, and are always willing to do what’s best for our users. We keep a close eye on our pipelines and processes to make sure we’re delivering useful, timely improvements to aid decision-making and data visualization within O’Reilly. The team is broadly distributed across the US in multiple cities and timezones and constantly encourages each other to deliver work that instills pride and fulfillment.

About the Job

We are looking for a thoughtful and experienced data engineer to help grow a suite of systems and tools written primarily in Python. The ideal candidate will have a deep understanding of modern data engineering concepts and will have shipped or supported code and infrastructure with a user base in the millions and datasets with billions of records. The candidate will be routinely implementing features, fixing bugs, performing maintenance, consulting with product managers, and troubleshooting problems. Changes you make will be accompanied by tests to confirm desired behavior. Code reviews, in the form of pull requests reviewed by peers, are a regular and expected part of the job as well.

Job Details

In a normal week, you might:

  • Develop a new feature from a user story using Python and PostgreSQL or BigQuery
  • Collaborate with product managers to define clear requirements, deliverables, and milestones
  • Team up with other groups within O’Reilly (e.g. data science or machine learning) to leverage experience and consult on data engineering best practices
  • Review a pull request from a coworker and pair on a tricky problem
  • Provide a consistent and reliable estimate to assess risk for a project manager
  • Learn about a new technology or paper and present it to the team
  • Identify opportunities to improve our pipelines through research and proof-of-concepts
  • Help QA and troubleshoot a pesky production problem
  • Participate in agile process and scrum ceremonies

Why you'll love working on our team:

  • You'll be working for a company that embraces and pursues new technology
  • You'll be working with a company that trusts and engages its employees
  • We believe in giving engineers the tools and hardware that they need to do their job
  • Bi-weekly virtual team hangouts and space to learn new skills (we’re a learning company after all!)
  • Great company benefits (health/dental/vision insurance, 401k, etc.)
  • We care deeply about work-life balance and treat everyone like human beings first

About You

What we like to see for anyone joining our data engineering teams:

  • Proficiency in building highly scalable ETL and streaming-based data pipelines using Google Cloud Platform services and products
  • Proficiency in large scale data platforms and data processing systems such as Google BigQuery and Amazon Redshift
  • Excellent Python and PostgreSQL development and debugging skills
  • Experience building systems to retrieve and aggregate data from event-driven messaging frameworks (e.g. RabbitMQ and Pub/Sub)
  • Strong drive to experiment, learn and improve your skills
  • Respect for the craft - you write self-documenting code with modern techniques
  • Great written communication skills—we do a lot of work asynchronously in Slack and Google Docs
  • Empathy for our users—a willingness to spend time understanding their needs and difficulties is central to the team
  • Desire to be part of a compact, fun, and hard-working team

Not required, but for bonus points:

  • Experience with Google Cloud Dataflow/Apache Beam
  • Experience with Django RESTful endpoints
  • Experience working in a distributed team
  • Knowledge and experience with machine learning pipelines
  • Contributions to open source projects
  • Knack for benchmarking and optimization

Minimum Qualifications

  • 2+ years of professional data engineering (or equivalent) experience
  • 1+ year experience of working in an agile environment

Company Info.

O’Reilly Media Inc.

O'Reilly Media is an American learning company established by Tim O'Reilly that publishes books, produces tech conferences, and provides an online learning platform. Its distinctive brand features a woodcut of an animal on many of its book covers.

  • Industry
    Information Technology
  • No. of Employees
    819
  • Location
    Sebastopol, CA, USA
  • Website
  • Jobs Posted

Get Similar Jobs In Your Inbox

O’Reilly Media Inc. is currently hiring Data Engineer Jobs in United States with average base salary of $120,000 - $190,000 / Year.

Similar Jobs View More