Design, develop, and support ETL pipelines for large-scale data processing
Work extensively with on-prem Oracle databases using SQL and PL/SQL
Explore, analyze, and maintain existing ODI mappings and workflows
Migrate legacy ODI / batch jobs / cron jobs to Apache Airflow and AWS Glue
Develop and manage complex data workflows involving:
File ingestion (CSV, JSON, XML, fixed-width files)
File extraction and transformation
Database-to-database data migration
Data transfer and bulk data loads
Build and optimize PL/SQL procedures, packages, and batch programs
Develop Airflow DAGs with proper scheduling, dependencies, retries, and monitoring
Implement AWS Glue ETL jobs (PySpark) for complex transformation logic
Handle data movement between on-prem Oracle and AWS cloud environments
Ensure data quality, reconciliation, validation, and error handling
Monitor, troubleshoot, and optimize ETL job performance
Collaborate with DBAs, architects, and application teams during migration and modernization initiatives