Architect and scale a high-performance data lakehouse on GCP, leveraging technologies like StarRocks, Apache Iceberg, GCS, BigQuery, Dataproc, and Kafka. Design, build, and optimize distributed query engines such as Trino, Spark, or Snowflake to support complex analytical workloads. Implement metadata management in open table formats like Iceberg and data discovery frameworks for governance and observability using Iceberg compatible catalogs. Develop and orchestrate robust ETL/ELT pipelines using Apache Airflow, Spark, and GCP-native tools (e.g., Dataflow, Composer). Collaborate across departments, partnering with data scientists, backend engineers, and product managers to design and implement