Bachelor's degree in Computer Science, Engineering, or a related field. Proven experience as a Data Engineer, with a focus on big data technologies. Strong proficiency in programming languages such as Python, Scala, or Java. Extensive experience with data warehousing, ETL processes, and data modeling. Experience with major cloud providers (e.g., AWS, GCP, Azure) and their data storage and processing services. Hands-on experience with big data frameworks like Apache Spark for distributed processing. Excellent problem-solving skills and the ability to work independently and as part of a team. Strong communication and interpersonal skills. Experience with healthcare data and a good understanding of healthcare data standards (e.g., FHIR, HL7). Familiarity with machine learning concepts and LLM fine-tuning processes. Experience with data orchestration tools (e.g., Apache Airflow).