Architect batch + stream pipelines for market data. Provide reusable SDKs in Python and Go. Implement and tune S3, column-oriented and time-series data storage. Own partitioning, compression, TTL, versioning and cost optimization. Develop internal libraries for schema management, data contracts, validation and lineage. Contribute to shared libraries and services. Embed monitoring, alerting, SLAs, SLOs and CI/CD. Champion automated testing, data quality dashboards and incident runbooks. Partner with Data Science, Quant Research, Backend and DevOps.