Location:
CN-Shenzhen-HyQ
Shift:
Standard - 40 Hours (China)
Scheduled Weekly Hours:
40
Worker Type:
Permanent
Job Summary:
The AVP under Data Engineering in the Chief Data Office, is a critical role in designing and maintaining scalable data pipelines and infrastructure that support enterprise-wide analytics and reporting. You will collaborate closely with data analysts, business stakeholders, and the Site Reliability Engineering (SRE) team to ensure data integrity, performance, and availability. The ideal candidate will have deep expertise in SparkSQL, Python, HDFS, and Trino, with a strong foundation in data architecture and distributed computing.
Job Duties:
Job Duties
- Lead a data transformation team of 6–8 engineers, fostering delivery excellence and technical leadership
- Collaborate with analysts and business stakeholders to gather requirements and deliver reliable data solutions
- Design, build, and maintain scalable data pipelines using SparkSQL and Python
- Manage and optimize data storage and retrieval within Hadoop Distributed File System (HDFS)
- Utilize Trino for distributed SQL querying and business intelligence support
- Ensure data quality, integrity, and security across all systems
- Monitor and optimize pipeline performance and query execution, conduct performance tuning and profiling
- Troubleshoot and resolve data-related issues to minimize business impact
- Manage production incidents and coordinate with the SRE Lead for incident investigation and postmortem analysis
Key Responsibilities
- Drive technical excellence and delivery within the data engineering team
- Architect and implement robust ETL pipelines and data models
- Maintain high-performance, scalable data infrastructure
- Support business intelligence and reporting through efficient data access and transformation
- Ensure operational reliability and collaborate on incident response
Requirements
Mandatory Skills and Qualifications
- Minimum 8 years of experience in data engineering, including 3+ years in a leadership role
- Strong expertise in SparkSQL for large-scale data processing
- Proven hands-on experience with HDFS for distributed data management
- Deep understanding of Trino for distributed SQL execution
- Proficiency in Python for scripting, automation, and data manipulation
- Solid grasp of ETL architecture, data modeling, and pipeline design
- Familiarity with distributed computing principles and big data ecosystems
- Basic proficiency in English and Mandarin (written and verbal)
Preferred Skills
- Experience with Tableau for data visualization and dashboard development
- Exposure to cloud platforms such as AWS or Huawei HCS
- Knowledge of SQL performance tuning and query optimization techniques
- Background in data governance, data quality, or metadata management
- Experience with capital markets, trading systems, or investment workflows
Company Introduction:
ITD SZ
港交所科技(深圳)有限公司,是2016年12月28日于深圳市前海自贸区成立的外商独资企业。
作为港交所的技术子公司,港交所科技(深圳)有限公司主要是为集团及其附属公司提供计算机软件、计算机硬件、信息系统、云存储、云计算、物联网和计算机网络的开发、技术服务、技术咨询、技术转让;经济信息咨询、企业管理咨询、商务信息咨询、商业信息咨询、信息系统设计、集成、运行维护;数据库管理、大数据分析;以承接服务外包方式提供系统应用管理和维护、信息技术支持管理、数据处理等信息技术和业务流程外包服务。