Job Description:
We are looking for a skilled Data Engineer to join our team. The ideal candidate will have a strong background in GCP data services, big data technologies, and data engineering best practices. This role involves designing, developing, and maintaining data pipelines, optimizing data processing performance, and collaborating with data scientists and analysts to ensure data quality and integrity.
Key Responsibilities:
Data Pipeline Development: Design, develop, and maintain data pipelines using Google Cloud Platform (GCP) data services (Big Query, Dataflow, Dataproc) and Informatica Cloud data integration tools.
ETL Processes: Extract, transform, and load (ETL) data from various sources into GCP data warehouses and data lakes.
Performance Optimization: Optimize data processing performance and scalability to handle large volumes of data efficiently.
Data Quality: Collaborate with data scientists and analysts to ensure data quality, integrity, and accuracy.
Technical Expertise: Utilize proficiency in SQL and Python programming to develop data solutions and address technical challenges.
Qualifications:
Experience: 3-5 years of experience with big data technologies such as Hadoop, Spark, and Hive.
GCP Proficiency: Strong proficiency in GCP data services including Big Query, Dataflow, and Dataproc.
Data Modeling: Knowledge of data modeling and data warehousing concepts.
Programming Skills: Proficiency in SQL and Python programming.
Data Integration: Familiarity with Informatica Cloud data integration tools.
We regret to inform you that this job opportunity is no longer available