Job Overview
As a Data Architect - Large Scale Data Transformation, you will play a critical role in developing and maintaining our world-class data platform. This platform supports various teams by providing scalable, efficient solutions for complex data processing and storage.
Your responsibilities will include designing and implementing the data storage model, defining data governance rules, ensuring high-level designs with stakeholders, and building processes for optimal extraction, transformation, and loading of data from various sources. You will also collaborate closely with our cloud infrastructure team and proactively ensure code reviews to help unblock the team's work.
We are looking for an experienced professional with over 6 years of engineering experience, including 3 years of focus on large-scale data processing and transformation. You should have hands-on experience with AWS services like Data Pipelines, Glue, Redshift, S3, or equivalent cloud technologies, as well as exceptional coding and design skills in Spark, Python, and Java.
Required Skills and Qualifications
* Designed and built scalable, data-centric systems
* Hands-on experience with AWS services like Data Pipelines, Glue, Redshift, S3
* Exceptional coding and design skills in Spark, Python, and Java
* Knowledge of dbt, HBase, Apache Spark, Spark streaming, Kafka, Scala, and similar technology stacks
* Experience in business intelligence data modelling and familiarity with data warehouse concepts such as star schema and data de-normalizing patterns
Benefits
* Successfully processed, transformed, and transported large-scale data
* Applied architectural design principles to create scalable, efficient solutions
* Demonstrated an analytical and strategic mindset driving successful project delivery