Sr Data Engineer
Minneapolis, MN
3 Months contract to hire
Job Summary
The Senior Data Engineer plays a key role in designing, building, and maintaining data pipelines and infrastructure using Google Cloud Platform (GCP) BigQuery.
The incumbent will collaborate with data analysts, data scientists, and other engineers to ensure timely access to high-quality data for data-driven decision-making across the organization.
The Senior Data Engineer is a highly technical person that hands-on codes data processing solutions and scalable data pipelines to support analytics and exploratory analysis.
This role ensures new business requirements are decomposed and implemented in the cohesive end-to-end designs that enable data integrity and quality, and best support BI and analytic capability needs that power decision-making at Surescripts.
This includes building data acquisition programs that handle the business's growing data volume as part of the Data Lake and Data Warehouse ecosystem and maintaining a robust data catalog.
This role continually improves the business's data and analytic solutions, processes, and data engineering capabilities.
The incumbent embraces industry best practices and trends and, through acquired knowledge, drives process and system improvement opportunities.
Responsibilities
- Design, develop, and implement data pipelines using GCP BigQuery for data ingestion, transformation, and loading.
- Optimize data pipelines for performance, scalability, and cost-efficiency.
- Ensure data quality through data cleansing, validation, and monitoring processes.
- Develop and maintain data models and schemas in BigQuery to support various data analysis needs.
- Automate data pipeline tasks using scripting languages like Python and tools like Dataflow.
- Collaborate with data analysts and data scientists to understand their data requirements and translate them into technical solutions.
- Comfortable with DevOps to ensure seamless integration of data pipelines with CI / CD workflows.
- Monitor and troubleshoot data pipelines and infrastructure to identify and resolve issues.
- Stay up-to-date with the latest advancements in GCP BigQuery and related technologies.
- Document data pipelines and technical processes for future reference and knowledge sharing.
Qualifications
Basic Requirements :
- Bachelor's degree or equivalent experience in Computer Science, Mathematics, Information Technology or related field.
- 5+ years of experience as a data engineer or a similar role.
- Strong understanding of data warehousing concepts and data modeling principles.
- Proven experience with designing and implementing data pipelines using GCP BigQuery.
- Strong SQL and scripting languages like Python (or similar) skills.
- Experience with data quality tools and techniques.
- Excellent communication and collaboration skills.
- Ability to work independently and as part of a team.
- Strong problem-solving and analytical skills.
- Passion for data and a desire to learn and adapt to new technologies.
- Experience with other GCP services like Cloud Storage, Dataflow, and Pub / Sub.
- Experience with cloud deployment and automation tools like Terraform.
- Experience with data visualization tools like Tableau or Power BI or Looker.
- Experience with healthcare data.
- Experience with machine learning, artificial intelligence and data science concepts.
- Experience with data governance and healthcare PHI data security best practices.