March 12
🔄 Hybrid – Bay Area
• As a software engineer at Onehouse, you will contribute directly to Apache Hudi and the surrounding open source ecosystem, while deploying and operating these technologies at massive scale for our customers. • Accelerate our open source <> enterprise flywheel by working on the guts of Apache Hudi's transactional engine and optimizing it for diverse Onehouse customer workloads. • Act as a SME to deepen our teams' expertise on database internals, query engines, storage and/or stream processing. • Build systems that enable users to manage petabytes of data with a fully managed cloud service. • Build functionality that enables data systems to be cloud native (self managed), scalable (auto scaling) and secure (different levels of access control). • Build scalable job management on Kubernetes to ingest, store, manage and optimize petabytes of data on cloud storage. • Design systems that help scale and streamline metadata and data access from different query/compute engines. • Exhibit full ownership of product features, including design and implementation, from concept to completion. • Be passionate about designing for future scale and high availability, while possessing a deep understanding of common failure patterns and their remediations. • Uphold a high engineering bar around the code, monitoring, operations, automated testing, release management of the platform.
• 3+ years of experience as a software engineer with experience developing distributed systems. • Strong, object-oriented design and coding skills with Java. • Experience with inner workings of distributed (multi-tiered) systems, algorithms, and relational databases. • Deal well with ambiguous/undefined problems; ability to think abstractly; articulate technical challenges and solutions. • Speed and hustle → Ability to prioritize across feature development and tech debt. • Ability to solve complex programming/optimization problems. • Ability to quickly prototype optimization solutions and analyze large/complex data. • Clear communication skills. • Nice to haves (but not required): • Experience working on database systems, Query Engines or Spark codebases. • Experience working on cloud based (data focused) services. • Deep understanding of Spark, Flink, Presto, Hive, Parquet internals. • Hands-on experience with open source projects like Hadoop, Hive, Delta Lake, Hudi, Nifi, Drill, Pulsar, Druid, Pinot, etc.
Apply Now