Overview
Director, Delivery Lead - Data role at Mizuho.
This position leads delivery for data platform initiatives and data migration programs, partnering with cross-functional teams and external vendors.
Responsibilities
- Support the design, development, and implementation of a scalable, reliable, and high-performance data platform, translating business needs into technical specifications.
- Manage timelines, milestones, and deliverables to ensure successful project execution.
- Take full ownership of the data mapping process, a critical component for the success of data migration and integration within the transformation program.
- Analyze source systems to define data flows from domain systems into the new platform and data migration requirements from legacy systems.
- Manage the creation, review, and sign-off of detailed Source-to-Target Mapping (STM) specifications, ensuring accuracy and completeness.
- Collaborate closely with Data Engineers, Data Architects, and Business Analysts to define and implement complex data transformation logic and business rules.
- Establish and execute a robust testing strategy to guarantee data quality, consistency, and integrity throughout the migration lifecycle.
- Act as the primary technical point of contact for external vendors involved in the data platform build and data mapping activities.
- Rigorously manage vendor performance, Statements of Work (SOWs), and deliverables to ensure alignment with project goals, timelines, and quality standards.
- Provide technical leadership to coordinate delivery outcomes with SI Partners, country technology teams, and other stakeholders.
- Champion best practices in data governance, quality, and security.
- Foster a culture of excellence, collaboration, and innovation within the data delivery team, utilizing Agile/Scrum methodologies to maximize productivity.
Requirements
- Bachelor’s degree in computer science, Engineering, Information Systems, or a related field.
- 8+ years of hands-on experience in data integration, data warehousing, or data platform development.
- 3+ years of experience in a lead role, successfully delivering large-scale, complex data projects, particularly involving data migration and system integration.
- Expertise in Data Mapping: proven, extensive experience creating detailed source-to-target data mapping documents based on complex business requirements.
- Cloud Platform Proficiency: deep knowledge and practical experience with at least one major cloud platform: AWS (e.g., S3, Glue, Redshift), Azure (e.g., ADLS, Data Factory, Synapse), or GCP (e.g., Cloud Storage, Dataflow, BigQuery).
- Modern Data Stack: strong understanding of modern data warehousing technologies like Snowflake, BigQuery, or Redshift.
- ETL/ELT Experience: hands-on experience with data pipeline and orchestration tools such as dbt, Airflow, Talend, or Informatica.
- Vendor Management: demonstrated experience managing relationships, contracts, and performance of third-party technology vendors.
- Communication Skills: exceptional ability to communicate complex technical concepts to both technical and non-technical audiences.
Preferred
- Professional certifications (e.g., AWS Certified Data Analytics, Google Professional Data Engineer, Databricks Certifications).
- Programming experience with Python or advanced SQL for data processing.
- Knowledge of data governance frameworks (e.g., DAMA-DMBOK).
- Experience working in an Agile/Scrum environment, with Scrum Master experience being a plus.
- Experience in Corporate / Commercial Banking or other Financial Services.
Seniority level
Employment type
Job function
- Business Development and Sales
We’re unlocking community knowledge in a new way.
Experts add insights directly into each article, started with the help of AI.
#J-18808-Ljbffr