Imply is looking for experienced database and distributed systems engineers to develop a next generation analytics platform focused on interactivity and streaming data. The star of our platform is
Apache Druid (incubating), an open-source, high-performance, real-time analytics database. As a distributed engineer at Imply, you will be heavily involved in the development and technical direction of the open-source Druid project.
You might work on:
- Implementing cutting-edge compression algorithms, storage formats, and other database optimizations.
- Building distributed ingestion systems that can handle throughput rates in the tens of millions of records per second.
- Designing systems for orchestration, monitoring, and performance analysis of Druid clusters.
- Integrating with major cloud vendors to build managed service offerings.
Required experiences and skills
- Bachelor’s degree in computer science, engineering, or a related field (or equivalent professional work experience).
- 5+ years of professional experience in software development with a recent focus on designing and building highly scalable, fault-tolerant distributed systems.
- Experience developing high concurrency, performance-oriented Java systems and using standard tools to tune, profile and debug JVMs.
- Familiarity working with products in the big data ecosystem (Apache Druid, Hadoop, Spark, Kafka, etc.).
- Experience building and deploying resilient, highly available production-grade systems to a public cloud provider such as AWS, Azure, or GCP.
- Working knowledge of Linux environments and proficiency with Python and shell scripting.
- Strong communication skills. Explaining complex technical concepts to designers, support, and other engineers is no problem for you.
- A history of open-source contributions is a plus. Being a committer on a data-related project is a big plus.