Responsibilities

  • Architect and implement end-to-end ETL pipelines using IBM DataStage/CP4D to meet evolving business needs.
  • Ensure information integrity, data boundary compliance, and secure handling of sensitive data.
  • Strong knowledge of ETL concepts to create scalable and reusable components with thorough documentation and version control.
  • Collaborate with stakeholders to assess existing legacy ETL jobs and plan migration to CP4D/GCP-native services.
  • Develop and execute migration strategies, including dependency mapping, data flow analysis, and compatibility validation.
  • Work on Terraform, Tekton pipelines, and Google BigQuery for orchestrating cloud-based transformations.
  • Proactively monitor production pipelines, resolve failures, and build a knowledge base to document recurring issues.
  • Conduct tuning and optimization, schedule workflows using Astronomer/Airflow, and manage FTP/SFTP-based data exchange.
  • Participate in Agile sprint planning, story grooming, and roadmap execution for delivering product features.
  • Lead test planning, validation, and deployment activities in collaboration with QA and DevOps teams.
  • Troubleshoot data quality issues and work closely with data integration teams to ensure resolution.