Role Overview:We are seeking a highly skilled and detail-oriented Informatica and Teradata Architect to lead the data integration and warehousing architecture efforts as part of a large-scale Data Lakehouse transformation program for a leading banking client. The ideal candidate will possess deep expertise in Informatica (PowerCenter and IDQ) and Teradata, with a strong grasp of enterprise-grade data integration, ETL modernization, and data warehouse optimization. This role requires close collaboration with data architects, engineers, QA teams, and banking domain stakeholders.Key Responsibilities:Lead the architecture and design of Informatica-based ETL frameworks to migrate and modernize existing data pipelines from legacy systems to the new Lakehouse platform.Design efficient Teradata data models to support gold-layer consumption and reporting needs.Develop strategies for ETL workload optimization, code migration, and performance tuning on both Informatica and Teradata platforms.Define standards, best practices, and governance around ETL design, parameterization, error handling, and metadata tracking.Work with SNB and oversight partner to map legacy DWH logic to new Cloudera-based Bronze/Silver architecture and Teradata-based Gold Layer.Provide technical oversight and guidance to developers and data engineers implementing integration and transformation logic.Collaborate with testing and validation teams to ensure data accuracy, lineage, and completeness through all layers.Support impact analysis and dependency checks across ~50 tables and 150+ ETL packages per source system.Required Skills and Experience:10+ years of experience in data warehousing, data integration, and enterprise ETL architecture.Strong hands-on experience with Informatica PowerCenter, Informatica IDQ, and Teradata (SQL, utilities, and performance optimization).In-depth understanding of ETL lifecycle, data quality frameworks, and integration patterns in a banking context.Demonstrated ability to lead ETL modernization efforts and architect scalable, reusable data pipelines.Experience in migrating ETL logic from on-prem systems to Lakehouse environments (e.g., Cloudera, Spark-based processing).Experience in tuning large-volume ETL jobs and Teradata queries to meet SLA-driven performance goals.Familiarity with data governance principles, error logging, and metadata management.Preferred Attributes:Experience with Cloudera Data Platform (CDP), Spark, Hive, or Iceberg tables.Knowledge of DevOps/DataOps for Informatica deployments and integration with CI/CD tools.Exposure to data masking, synthetic data generation, and regulatory data controls.Good understanding of banking use cases in areas like risk reporting, compliance, credit, and finance.