about the company.
Our client is a highly prestigious, world-class sports and entertainment organization with a legacy spanning over a century. They operate a unique, highly profitable integrated business model that encompasses premium membership services, large-scale sports entertainment, high-volume transactional systems, and significant philanthropic contributions. Operating at an enterprise scale, this organization drives technical innovation while offering employees unparalleled career stability, highly competitive compensation, and the opportunity to work with cutting-edge technology.
...
about the team.You will be joining the Central IT & Data Engineering department, the core engine room that powers the enterprise's data-driven initiatives. This team is responsible for designing, building, and maintaining the cloud data platforms that handle terabytes to petabytes of consumer and high-volume transactional data. You will work alongside highly skilled Technical Managers (Architects) and Cloud DevOps engineers in a mature data environment. Because the organization has dedicated Data Science and Business Analytics units, the team's pure focus remains on large-scale data engineering and pipeline optimization.
about the job.As an ETL/ELT Developer, you will be the hands-on technical expert responsible for moving, transforming, and optimizing massive datasets. Your core responsibilities include:
- Pipeline Development: Design, build, test, and operationalize highly scalable ETL/ELT data pipelines using PySpark and SQL on the AWS platform.
- Architecture Implementation: Build and manage data flows following the Databricks Medallion Architecture to ensure data reliability and quality.
- Performance Optimization: Tune and troubleshoot complex PySpark code to process huge volumes of data (terabyte to petabyte scale) efficiently and cost-effectively.
- Orchestration: Automate, schedule, and monitor end-to-end data workflows using tools like Apache Airflow.
- Collaboration: Partner with Data Architects to execute infrastructure designs and deliver pristine, business-ready data to downstream Analytics and Data Science teams.
skills & experiences required.- Experience: 3 to 6 years of hands-on experience in data engineering, specifically focused on building cloud-based ETL/ELT pipelines.
- Core Technical Skills: Strong production-level proficiency in Databricks PySpark and Advanced SQL for large-scale data processing.
- Platform Expertise: Commercial experience with Databricks (a must-have) and a deep understanding of Medallion Architecture principles.
- Cloud Knowledge: Hands-on experience with AWS data analytics services.
- DataOps: Familiarity with pipeline orchestration (Apache Airflow), automated testing, and monitoring tools (Datadog, CloudWatch).
- Language: Strong communication skills with a good command of written and spoken English and Chinese (Cantonese).
show more
about the company.
Our client is a highly prestigious, world-class sports and entertainment organization with a legacy spanning over a century. They operate a unique, highly profitable integrated business model that encompasses premium membership services, large-scale sports entertainment, high-volume transactional systems, and significant philanthropic contributions. Operating at an enterprise scale, this organization drives technical innovation while offering employees unparalleled career stability, highly competitive compensation, and the opportunity to work with cutting-edge technology.
about the team.You will be joining the Central IT & Data Engineering department, the core engine room that powers the enterprise's data-driven initiatives. This team is responsible for designing, building, and maintaining the cloud data platforms that handle terabytes to petabytes of consumer and high-volume transactional data. You will work alongside highly skilled Technical Managers (Architects) and Cloud DevOps engineers in a mature data environment. Because the organization has dedicated Data Science and Business Analytics units, the team's pure focus remains on large-scale data engineering and pipeline optimization.
...
about the job.As an ETL/ELT Developer, you will be the hands-on technical expert responsible for moving, transforming, and optimizing massive datasets. Your core responsibilities include:
- Pipeline Development: Design, build, test, and operationalize highly scalable ETL/ELT data pipelines using PySpark and SQL on the AWS platform.
- Architecture Implementation: Build and manage data flows following the Databricks Medallion Architecture to ensure data reliability and quality.
- Performance Optimization: Tune and troubleshoot complex PySpark code to process huge volumes of data (terabyte to petabyte scale) efficiently and cost-effectively.
- Orchestration: Automate, schedule, and monitor end-to-end data workflows using tools like Apache Airflow.
- Collaboration: Partner with Data Architects to execute infrastructure designs and deliver pristine, business-ready data to downstream Analytics and Data Science teams.
skills & experiences required.- Experience: 3 to 6 years of hands-on experience in data engineering, specifically focused on building cloud-based ETL/ELT pipelines.
- Core Technical Skills: Strong production-level proficiency in Databricks PySpark and Advanced SQL for large-scale data processing.
- Platform Expertise: Commercial experience with Databricks (a must-have) and a deep understanding of Medallion Architecture principles.
- Cloud Knowledge: Hands-on experience with AWS data analytics services.
- DataOps: Familiarity with pipeline orchestration (Apache Airflow), automated testing, and monitoring tools (Datadog, CloudWatch).
- Language: Strong communication skills with a good command of written and spoken English and Chinese (Cantonese).
show more