2 days ago
🏢 In-office - Bay Area
Apache
AWS
Azure
Cassandra
Cloud
Distributed Systems
Google Cloud Platform
GraphQL
Hadoop
HDFS
Java
Kafka
Kubernetes
Pandas
Postgres
Python
Ray
Spark
Go
• Design and develop infrastructure for data pipelines • Design abstractions over datastores (Cassandra, PostgreSQL) • Develop file system abstractions over AWS S3 • Design connectors to various external data stores • Develop distributed system components for processing • Maintain industry-leading APIs for AI/ML applications • Work on frameworks for performance and scalability • Collaborate across teams in a highly collaborative environment • Write clean code following test-driven methodology • Deliver commitments promptly following agile methodology
• Bachelor's degree in Computer Science or related fields • Strong understanding of Computer Science fundamentals • High proficiency in coding with Java, C++, C#, or Python • Strong competency in object-oriented programming, data structures, algorithms, and software design patterns • Experience with version control systems such as Git • Experience with large-scale distributed systems • Experience with any public cloud platform (AWS, Azure, GCP) • Familiarity with distributed computing technologies (e.g., Hadoop, Spark, Kafka) • Familiarity with modern data science/analysis libraries (e.g., Pandas, Koalas) • Good verbal and written communication skills • Thrive in a fast-paced environment and value ownership • Intellectually curious and open to challenges
• Excellent benefits • Competitive compensation package • Generous equity plan
Apply Now