Responsibilities:
• Design, develop, and implement ETL/ELT processes to extract data from various sources,
transform it according to business requirements, and load it into data warehouses or data
lakes.
• Collaborate with cross-functional teams to understand data requirements and design scalable
data solutions.
• Optimize and tune existing ETL/ELT processes for performance and reliability.
• Develop and maintain documentation for ETL/ELT processes, including data lineage and
metadata management.
• Monitor and troubleshoot data pipeline issues, ensuring data integrity and availability.
• Stay updated with emerging technologies and trends in ETL/ELT, big data, and data streaming.
Qualifications:
• Bachelor’s degree in computer science, Information Technology, or related field.
• Strong proficiency in SQL and experience with relational databases (e.g., PostgreSQL, MySQL,
Oracle).
• Hands-on experience with ETL/ELT tools and frameworks such as Matillion, Stitch Data,
Fivetran, Informatica, Apache Spark, Apache Airflow, Talend.
• Expertise in dbt (core/cloud) for data modeling and transformation.
• Knowledge of big data technologies and frameworks such as Hadoop, Hive, Spark.
• Experience with data streaming technologies such as Apache Kafka, Apache Flink.
• Solid understanding of data warehousing concepts and methodologies.
• Excellent problem-solving skills and attention to detail.
• Strong communication and collaboration skills.
• Ability to work effectively in a fast-paced environment and manage multiple tasks
simultaneously.