Data Stream Processing Software Engineer
Ship innovations in streaming data processing at publicly traded company data scales but at startup scale shipping velocities!
Responsibilities
- Leverage Apache Flink to develop robust and efficient stream processing jobs and tackle challenges related to data skew, out-of-order events, and other intricacies of streaming data.
- Architect and implement near real-time ingestion pipelines for data lakes.
- Maintain and optimize the data infrastructure for streaming systems.
- Collaborate closely with cross-functional team members to ensure seamless data processing.
Requirements
- Enthusiasm for working in a startup environment with a small, dedicated team.
- Proficiency in at least one programming language, such as Java, Scala, Kotlin, or Python.
- Hands-on experience with big data systems such as Apache Spark, Apache Flink, and Apache Kafka.
- Knowledge of common data storage formats such as Avro, Proto, and Parquet.
- Experience with at least one cloud platform, such as AWS or GCP.
- Familiarity with Kubernetes and experience with tools such as Helm, Prometheus, and Grafana.
Preferred Qualifications
- Familiarity with data analytics tools such as Apache Hive, ClickHouse, Presto, Trino, or AWS Redshift.
- Experience data lake formats such as Apache Iceberg, Hudi, and Paimon.
- Experience with DevOps practices, IoC, and Terraform.
- Background in Machine Learning or Recommendation Systems.