[Remote] Senior Software Engineer- Open Source Analytics
Note: The job is a remote job and is open to candidates in USA. Snowflake is a company focused on empowering enterprises to become data-driven with cutting-edge technology. As a Senior Software Engineer on the Open Source Analytics team, you will innovate and build capabilities in the open data lake ecosystem, collaborating with the open source community and contributing to projects like Apache Iceberg and Apache Polaris. Responsibilities • Pioneer new and innovative technical capabilities in the Open Source Analytics community. You will define and build next-generation capabilities on top of critical lakehouse building blocks like interoperable table formats, data catalogs, file formats, and query engines • Design and implement features and enhancements for Apache Iceberg and Apache Polaris focusing on scalability, performance and usability such as Iceberg DML/DDL transactions, schema evolution, partitioning, time travel, and more • Collaborate with the Open source community by contributing code, participating in discussions and reviewing pull requests to ensure high quality contributions • Architect and build systems that integrate open source technologies seamlessly with Snowflake - enabling our customers to build and deploy massive data lake architectures across platforms and across cloud providers • Collaborate with Snowflake’s open-source team and the Apache Iceberg community to contribute new features and enhance the Iceberg table format and REST specification • Work on core data access control and governance features for Apache Polaris • Contribute to our managed Polaris service, Snowflake Open Catalog, enabling customers to seamlessly manage and expand their data lake through Snowflake as well as other external query engines like Spark and Trino • Build tooling and services that automate data lake table maintenance, including compaction, clustering, and data retention for enhanced query performance and efficiency Skills • 5+ years of experience designing and building scalable, distributed systems • Strong programming skills in Java, Scala, or C++ with an emphasis on performance and reliability • Deep understanding of distributed transaction processing, concurrency control, and high-performance query engines • Experience with open-source data lake formats (e.g., Apache Iceberg, Parquet, Delta) and the challenges associated with multi-engine interoperability • Experience building cloud-native services and working with public cloud providers like AWS, Azure, or GCP • A passion for open-source software and community engagement, particularly in the data ecosystem • Familiarity with data governance, security, and access control models in distributed data systems • Contributing to open-source projects, especially in the data infrastructure space • Designing or implementing REST APIs, particularly in the context of distributed systems • Managing large-scale data lakes or data catalogs in production environments • Working on highly-performant and scalable query engines such as Spark, Flink, or Trino Benefits • Medical, dental, vision, life, and disability insurance • 401(k) retirement plan • Flexible spending & health savings account • At least 12 paid holidays • Paid time off • Parental leave • Employee assistance program • Other company benefits Company Overview • Snowflake is a cloud data platform that provides a data warehouse as a service designed for the cloud. It was founded in 2012, and is headquartered in San Mateo, California, USA, with a workforce of 5001-10000 employees. Its website is Company H1B Sponsorship • Snowflake has a track record of offering H1B sponsorships, with 428 in 2025, 281 in 2024, 154 in 2023, 182 in 2022, 113 in 2021, 98 in 2020. Please note that this does not guarantee sponsorship for this specific role. Apply tot his job