Data Engineer
Stefanini Group is hiring!
Stefanini is looking for a Data Engineer, Dearborn, MI (Hybrid)
For quick apply, please reach out Anmol Tyagi at anmol.tyagi@stefanini.com/ 248-263-8628
Key Responsibilities:
Data Pipeline Development: -
Design, build, and maintain scalable and robust data pipelines on GCP using tools such as Apache Airflow, Cloud Composer, and Cloud Dataflow.-
Implement data integration solutions to ingest data from various sources, including cloud storage, and third-party APIs.-
Data Warehousing:-
Develop and optimize data warehouse solutions using BigQuery and other GCP services.-
Ensure data accuracy, consistency, and security within the data warehouse environment.-
Monitor and troubleshoot data pipeline and warehouse issues to maintain system reliability.
Cloud Platform Expertise: -
Utilize GCP services such as Cloud Storage, Cloud Run, and Cloud Functions to build scalable and cost-effective data solutions.-
Implement best practices for cloud infrastructure management, including resource provisioning, monitoring, and cost optimization.
Collaboration and Communication: -
Work closely with data scientists, analysts, and business stakeholders to understand data requirements and deliver high-quality data solutions.-
Collaborate with cross-functional teams to design and implement data models, ETL processes, and reporting solutions.
Automation and Optimization:-
Develop automated workflows using Apache Airflow and Astronomer to streamline data processing and improve efficiency.-
Continuously optimize data pipelines for performance, scalability, and cost-effectiveness.
Documentation and Training:-
Create and maintain comprehensive documentation for data pipelines, data models, and infrastructure components.-Provide training and support to team members and stakeholders on data engineering best practices and GCP services.