Senior Data Infrastructure Engineer, are you there? Hey there! We're looking for a Senior Data Infrastructure Engineer, for a venture-backed AI startup building high-performance forecasting systems powered by large-scale data infrastructure. If you are deeply analytical, hands-on, and experienced designing OLAP schemas and ingestion pipelines in columnar databases like ClickHouse, keep reading. 100% Remote Salary in USD APAC/LatAm ( 4-hour overlap with Sydney, AEST) Full Time Advanced English About our client You will be working for a well-funded AI startup building forecasting systems that trade on prediction markets and process extremely large volumes of time-series and event data. Their infrastructure powers analytics, calibration metrics, backtesting, and portfolio systems — and they need a senior engineer to own and build the OLAP layer from scaffolding to production-grade performance. About Athyna Athyna is a Global Talent Platform, helping align world-class talent with the best employment opportunities across the globe. We believe in the future of work. That’s why we are a 100% globally distributed team across 5 continents. We believe in people, and we value freedom, integrity, quality, and sustainability. We’re proudly carbon neutral as an organization and 5% of all revenue funds impact-driven startups & climate tech. We truly care about everyone on our team, and we want to make the world a better place for all of us. In your day-to-day, you will Design and own the ClickHouse OLAP layer powering analytics, backtesting, and portfolio systems Design schemas for large-scale time-series and event data (trade history, market events, prediction outcomes) Choose appropriate MergeTree engine types, partition strategies, and sort keys aligned with access patterns Build ingestion pipelines from Postgres-based systems (Convex / Supabase) into ClickHouse Define CDC boundaries: real-time vs batch sync vs backfill Replace read-time deduplication ( argMax() patterns) with proper materialized view pipelines Implement incremental aggregation using State / Merge combinators where appropriate Build a query backend that integrates with a Python analytics engine Diagnose and optimize slow queries from first principles (partition pruning, predicate pushdown, memory management) Own operational concerns: schema migrations in version control, monitoring, profiling, cost management, and alerting This is a hands-on, Staff-level IC role — not a management position. Skills you have (Mandatory) Deep experience designing OLAP schemas for time-series or event data at scale (hundreds of millions to billions of rows) Strong expertise in ClickHouse OR equivalent columnar databases (BigQuery, Druid, DuckDB, Pinot, Snowflake, etc.) Strong understanding of ingest-time vs query-time computation trade-offs Experience implementing materialized views and incremental aggregation strategies Hands-on experience building data ingestion pipelines (CDC, batch sync, backfills) Ability to diagnose slow analytical queries from first principles Strong knowledge of Postgres and transactional systems Comfortable reading and integrating with both Python and TypeScript codebases Staff/Principal-level IC mindset (100% hands-on, architectural depth) Skills that will make you stand out Production-level ClickHouse with MergeTree engine variants Experience in trading systems, prediction markets, crypto, or fintech Familiarity with Supabase or Convex Experience building query layers serving both dashboards and heavy analytical workloads AI-assisted development workflows You’ll be entitled to Competitive USD compensation Flexible working hours High-impact role working directly with founders Strong long-term growth potential Opportunity to work on technically dense problems at the intersection of AI, finance, and data infrastructure