Job Description:
• Optimize and scale our PostgreSQL (Supabase) infrastructure
• Design indexing, partitioning, and query strategies for large-scale media datasets
• Improve performance across ingestion, enrichment, and retrieval pipelines
• Build internal tools for querying and auditing large datasets
• Create customer-ready dataset sample packs
• Design and automate dataset exports and delivery pipelines (S3, secure transfers, custom formats)
• Work across backend, ML, and product teams to support new features
Requirements:
• Strong PostgreSQL expertise (indexing, partitioning, performance tuning)
• Experience working with large datasets (100M+ records preferred)
• Deep understanding of storage systems (S3 or similar object storage)
• Strong backend experience (TypeScript, Python, or similar)
• Comfort building internal tooling and automation scripts
• Ability to move between database, backend, and infrastructure work
• Bonus:
• Experience with data pipelines (ETL, transformation layers)
• Experience with vector databases (pgvector, FAISS, Pinecone)
• Experience delivering structured datasets to enterprise customers
• DevOps experience (CI/CD, infra automation)
• Experience working with media-heavy systems
Benefits:
• $150,000 – $350,000 equity
• Relocation support