Sr/Lead Data Engineer (Python/Spark/Jupyter Notebooks/Delta Lake/Data Vault 2.0)

Yoh, A Day & Zimmermann Company
Avenel, New Jersey
$77-$110 an hour
Full-time

Sr / Lead Data Engineer (Python / Spark / Jupyter Notebooks / Delta Lake / Data Vault 2.0)

Location : Remote (MST / CST / EST preferred)

Pay rate : $77-$110 / HR W2

Duration : 6-month increments, if going well will extend out yearly

Notes :

  • Client has built a modern data platform and needs senior data engineers to work on various projects supporting Client’s business Storage initiatives Data Applications Code reviews Azure Synapse Analytics Delta Lake initiatives
  • Data Group 65-70 resources including Data Engineers, Data Analysts, BA’s, QA, Scrum Masters, etc broken out into 6 teams Do NOT use ETL tools, utilize Data Vault 2.

0 methods for data transfer

  • Looking for VERY senior resources, up to hands-on lead level Experienced with Assertion based Architecture Engineers vs coders Coding is done in Jupyter Notebooks on Delata Lakes Need resources who can articulate design and build highly scalable solutions before jumping into coding Do NOT want resources who need to be told what to do Need critical thinkers who can troubleshoot and debug Independent workers, self starters, who speak up and raise impediments and offer solutions
  • Required skills : Python Jupyter Notebooks Delta Lake Spark, PySpark, Spark SQL Serverless data infrastructure Data Vault 2.

0 methodology experience Great Expectations data quality validation Automated Testing

Bonus skills : Kakfa streaming HUGE plus if they have solid background here Scala

Key Responsibilities :

  • Design, develop, and maintain data pipelines using Python, PySpark, and Spark SQL to process and transform large-scale datasets.
  • Implement Delta Lake architecture to ensure data reliability, consistency, and integrity for large, distributed datasets.
  • Utilize serverless data infrastructure (e.g., AWS Lambda, Azure Functions, Databricks) to build scalable and cost-efficient data solutions.
  • Collaborate with Data Scientists and Analysts by creating reusable Jupyter Notebooks for data exploration, analysis, and visualization.
  • Optimize and manage data storage and retrieval processes, ensuring high performance and low latency.
  • Implement best practices for data security, governance, and compliance within the data infrastructure.
  • Work closely with cross-functional teams to understand data requirements and deliver solutions aligned with business objectives.
  • Continuously monitor, troubleshoot, and improve the performance of data processing pipelines and infrastructure.

Qualifications :

  • 10-15+ years of experience in data engineering or related fields.
  • Strong programming skills in Python with experience in data processing frameworks like PySpark.
  • Extensive hands-on experience with Apache Spark and Spark SQL for processing and querying large datasets.
  • Expertise with Delta Lakes for implementing scalable data lakehouse architectures.
  • Experience with Jupyter Notebooks for prototyping and collaboration with data teams.
  • Familiarity with serverless data technologies such as AWS Lambda, Azure Functions, or similar platforms.
  • Proficient in working with cloud platforms such as AWS, Azure, or Google Cloud.
  • Experience with data pipeline orchestration tools (e.g., Apache Airflow, Prefect, or similar).
  • Solid understanding of data warehousing, ETL / ELT pipelines, and modern data architectures.
  • Strong problem-solving skills and ability to work in a collaborative environment.
  • Experience with CI / CD pipelines and DevOps practices is a plus.

Preferred Qualifications :

  • Experience with Databricks for data engineering workflows.
  • Familiarity with modern data governance practices and tools like Apache Atlas or AWS Glue.
  • Knowledge of machine learning workflows and how data engineering supports AI / ML model s.

Note : Any pay ranges displayed are estimations. Actual pay is determined by an applicant's experience, technical expertise, and other qualifications as listed in the job description.

All qualified applicants are welcome to apply.

Yoh, a Day & Zimmermann company, is an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, or status as a protected veteran.

Visit to contact us if you are an individual with a disability and require accommodation in the application process.

For California applicants, qualified applicants with arrest or conviction records will be considered for employment in accordance with the Los Angeles County Fair Chance Ordinance for Employers and the California Fair Chance Act.

All of the material job duties described in this posting are job duties for which a criminal history may have a direct, adverse, and negative relationship potentially resulting in the withdrawal of a conditional offer of employment.

7 hours ago
Related jobs
Promoted
Hispanic Technology Executive Council
Jersey City, New Jersey

As a Data and Operational reporting Lead for Custody Technology you will work together with product development and operational teams and members of the development team to provide solutions to modernizing and optimizing the current reporting platform and optimize data integration with data warehous...

Promoted
PwC
Bayonne, New Jersey

Demonstrates thought leader-level abilities with, and/or a proven record of success directing efforts: Demonstrating thorough knowledge in designing and implementing MDM solutions across multiple sectors; Sourcing, cleansing, deduping/match merge for master data; Demonstrating proven experience in i...

Promoted
Cloud Analytics Technologies, LLC
Newark, New Jersey

Lead Global Insights and Data Science. ...

Promoted
Dell Technologies
Jersey City, New Jersey

They must have strong experience using a variety of data analysis methods and data science libraries to build and implement new models. Assess the effectiveness and accuracy of new data sources and data gathering techniques. Research and develop new data models and algorithms to apply to data sets. ...

tanishasystems
Jersey City, New Jersey

Job Title: Data Engineer ( Python + AWS)Location : Jersey City (3 day onsite per week)Experience Level : 9+Position ID: 384412 Job DescriptionExecutes creative software solutions, design, development, and technical troubleshooting with ability to think beyond routine or conventional approaches to ...

Axis Group
Berkeley Heights, New Jersey

As a Senior Data Engineer, you will collaborate with cross-functional teams to gather requirements, architect data solutions, and ensure data quality and reliability. Design, develop, and maintain robust and scalable data pipelines for extracting, transforming, and loading (ETL) data from various so...

JPMorgan Chase & Co.
Jersey City, New Jersey

Experience with data lakes, delta lakes, data validation/ingesting/pipelines. As a Senior Lead Software Engineer at JPMorgan Chase within the Corporate Technology, Treasury and Chief Investment Office of Finance Risk Data and Controls business, you are an integral part of an agile team that works to...

SMBC Group
Jersey City, New Jersey

Employ communication skills to articulate a clear vision of transformation deliverables and encourage buy-in across all stakeholder groups, including business leadership, process owners and infrastructure partners. ...

Mondelēz International
Remote Worker, New Jersey, USA
Remote

In collaboration with the Data Governance Office (DGO) and Global Process & Governance team, lead efforts to establish governance processes for Master Data and emerging data areas, such as reference data. Accountable for full scope of data migration, data process definition, stablishing of data gove...

Seven Seven Softwares
Newark, New Jersey

Experience in AWS data lake/data warehouse/business analytics-Experience in system analysis, design, development, and implementation of data ingestion pipeline in AWS-Knowledge of ETL/ELT-End-to-end data solutions (ingest, storage, integration, processing, access) on AWS-Architect and implement CI/C...