Key Responsibilities:
* Design complex data engineering solutions and define standards.
* Architect and optimize secure, scalable pipelines for real-time and batch processing.
* Integrate diverse data sources, implement fault-tolerant systems, and establish continuous integration and delivery practices.
* Organize large datasets, ensure data quality, and design data lake/warehouse solutions for accessibility.
* Monitor pipeline performance, troubleshoot issues, and implement observability and alerting systems.
The ideal candidate will have strong programming skills in Python, Java, or Scala, deep SQL expertise, and experience with NoSQL and big data technologies.
A bachelor's degree in computer science, bioinformatics, or a related field is required, along with at least 8 years of relevant experience in biotechnology, pharmaceuticals, or other life sciences industries, and 3+ years of cloud platform experience.