Image Description

Senior Data Engineer (CA or Remote)

Job Description

Posted 15 days ago

Summary

We're looking for a Senior Data Engineer experienced with streaming data processing and generating reporting metrics . Our ideal candidate has worked with stream processing tools and frameworks (such as Kafka, Flink, Storm, etc), analytics datastores (such as Redshift, BigQuery, Druid, ClickHouse, Snowflake), is comfortable managing infrastructure with Kubernetes, and enjoys working in a remote but collaborative environment.

This position is ideally in San Francisco. Remote can be considered but you must be based in the US time zone.

Team, Culture, and Product

We're a 7-person engineering team in a 33-person company and are looking to rapidly and responsibly grow our team. We're remote-heavy, spread across three continents with hubs in Berlin and San Francisco. We aim to help mobile app developers make data-driven decisions about their marketing in a secure, privacy-centric, user-friendly way.

Technology

We run our web and data processing services on Amazon EKS (Elastic Kubernetes Services). Most of our web services are in Go and most of our data processing services are in Java. We also have legacy data collection and backend services that are written in Ruby.

We've started using Kafka and Flink to process events in some of our newer pipelines. We use S3, Redshift, and DynamoDB for storage and serve aggregated reports that support our APIs and dashboard from Postgres.

Our user-facing dashboard is a Ruby on Rails web application with a React JS frontend.

None of this is set in stone - we're very open to experiments and prototyping new approaches with other technology. We hope you'll be able to provide some experience-based guidance and fresh perspective here!

Responsibilities

- Participate in major architecture and software design decisions about our reporting pipeline

- Develop and maintain services for generating and surfacing reports to our customers in a performant, low-latency manner

- Prototype new data pipelines and datastores

- Work cross-functionally with design and product to set requirements

- Set up monitoring, testing, and integrity alerts across our existing and new data pipelines

Requirements

Required Skills

- Experience with a stream processing framework such as Flink, Storm, or Spark

- Experience with big data analytics datastores such as Redshift, BigQuery, Druid, ClickHouse, or Snowflake

- Experience with Git and Github or similar

- Experience with Java and SQL

- Excellent organization and communication skills

Nice-to-haves

- Experience with data platforms such as dbt or Dremio

- Experience with mobile marketing

- Experience with Ruby or Go

Job Summary

Image Description
Website
Location

Similar Jobs

The largest community on the web to find and list jobs that aren't restricted by commutes or a specific location.

Pennylane is hiring an Engineering Manager. You would man...

Paubox is hiring for a Front-End React Developer. You sho...

Sensor Tower is hiring a DevOps engineer. You will be res...