Search jobs > New York, NY > Senior data engineer

Senior Data Pipeline Engineer

Clever Real Estate
New York, NY, United States
Full-time

Position : Senior Data Pipeline Engineer Location : 100% Remote, 40-hour Contract Position About Clever At Clever Real Estate , we envision a world where every person makes real estate moves that bring them closer to their life's goals.

To achieve that vision, we connect people with the most trustworthy advice and the best solutions for every step of their real estate journey.

We've built the leading online education platform in real estate and are rated the #1 real estate company on TrustPilot and our solutions have helped people save over $160 million on real estate fees.

Clever is venture-backed, profitable, and still evolving. If you're looking for a challenging role with deep ownership, smart co-workers, and real-world impact read on to learn more! About the Role 100% Remote Clever is seeking a highly skilled Senior Data Pipeline Engineer to join our team.

This role involves designing, developing, and maintaining robust data pipelines for efficient extraction, transformation, and loading (ETL / ELT) of data from various sources into our data ecosystem, which includes internal, B2B, and consumer-facing tools.

This position requires a strong understanding of data engineering principles, distributed computing systems, and experience with big data technologies.

Key Responsibilities : Design and Maintenance : Build and maintain scalable and reliable data pipelines to ingest, process, and transform large volumes of data from diverse sources.

Collaboration : Work with cross-functional teams, including data scientists, software engineers, product owners, and business analysts, to understand data requirements and translate them into technical solutions.

Optimization : Optimize data pipelines for performance, scalability, and cost-efficiency, ensuring timely delivery of high-quality data to downstream systems and applications.

Quality Assurance : Implement data quality checks, error handling mechanisms, and monitoring solutions to ensure the integrity and reliability of data pipelines.

Innovation : Stay current with emerging technologies and industry trends in data engineering and apply best practices to continuously improve our data pipeline infrastructure.

Leadership : Provide technical leadership and mentorship to junior members of the data engineering team, fostering a culture of collaboration, innovation, and excellence.

What to Expect First 30 Days : Gain a comprehensive understanding of all the data sources we utilize. Begin developing high-level models and diagrams to map out where the data is coming from and where it is stored.

Create diagrams of data flow and storage architecture. First 90 Days : Ensure the delivery of content products on a weekly basis.

Complete 50% of the data modeling and implementation tasks. Establish a strong collaborative relationship with Product Managers, actively planning for the next 6 to 12 months.

Research, vet, and select appropriate data build tools and data warehouse solutions (e.g., Redshift, Snowflake, BigQuery).

First 6 Months : Develop and release a minimum of 4 to 5 content series 6 is above and beyond! . Build a functional data pipeline that addresses the needs and organizes Clever's data effectively.

Qualifications Bachelor's or Master's degree in Computer Science, Engineering, or a related field, or equivalent experience.

5+ years of experience in data engineering, with a focus on building and maintaining data pipelines. Proficiency in object-oriented programming languages and experience with frameworks / libraries for data processing (e.

g., Apache Spark, Apache Beam). Experience with high volume data warehouses (e.g., Redshift, Snowflake, BigQuery). Strong understanding of distributed computing concepts and experience with big data technologies (e.

g., Apache Kafka, Apache Flink). Experience with cloud platforms, particularly AWS, and related services such as AWS Glue and AWS Data Pipeline.

Familiarity with data pipeline orchestration tools such as Apache Airflow for workflow management and scheduling. Experience with modern data modeling tools such as DBT (Data Build Tool) for transforming data into analytics-friendly formats.

Excellent problem-solving skills, attention to detail, and ability to work effectively in a fast-paced environment. Strong communication skills and ability to collaborate with cross-functional teams to deliver high-quality solutions.

Technologies We Use Directus Google Colab Jira for issue tracking Github Actions, BitRise, AWS, and Heroku for infrastructure and CI / CD integration Logistics We are looking for a full-time Senior Data Pipeline Engineer who can work 40 hours per week.

Clever Real Estate operates on a flexible work schedule with core business hours from 9 AM to 5 PM CT, Monday through Friday.

We accommodate variations in schedules as long as team members have a significant overlap ( 6 hours) with the core business hours.

Equal Employment Opportunity Employer Statement : Clever Real Estate provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state or local laws.

6 days ago
Related jobs
Promoted
DoubleVerify
New York, New York

Senior Software Engineer (Data) - Social Integrations. As a Senior Software Engineer (Data), you own new initiatives, design and build world-class platforms to measure and optimize ad performance. Senior Software Engineer II - Social Integrations. Deep understanding of distributed system technologie...

Promoted
Justworks
New York, New York

Your Success ProfileWhat You Will Work OnDesign, build, and maintain data pipelines and models that power tax computation, withholding, and reporting for Justworks customersBuild reliable and resilient data pipelines for shared datasets with a focus towards ensuring that data is correct, complete, a...

Promoted
HealthVerity
New York, New York

As a data engineer on the data platform team, you will be supporting and enhancing the platform that supports HealthVerity’s Petabyte-scale core data asset. Engineer efficient, adaptable and scalable data pipelines to process structured and unstructured data. You will work closely with other enginee...

Datadog
New York, New York

You will be working on large-scale distributed systems, horizontally-scalable datastores and a variety of data-processing frameworks to build our next-generation platform and capabilities for Data Science at Datadog. Senior Software Engineer - Data Science. Ensure that our data science-driven featur...

Goldman Sachs
New York, New York

Our engineering teams build scalable software and systems, architect low latency infrastructure solutions, proactively guard against cyber threats, and leverage machine learning alongside financial engineering to continuously turn data into action. In-depth knowledge of relational and columnar SQL d...

Datadog
New York, New York

Senior Software Engineer - Data Reliability. Work with developers to design data models and advise on datastores options. We're on a mission to build the best platform for our engineers to deliver stateful services at high scale. We provide High Performance Transaction Systems to all the Datadog dev...

Global Channel Management, Inc
New York, New York

Senior Cloud Data Platform Engineer needs 5+ years of implementing data applications or data platforms with BigData/Hadoop, Python/Java/Spark full stack, etc. Senior Cloud Data Platform Engineer requires:. Extensive experience in designing, engineering and managing data lake ingestion, validation, t...

Altice USA
Queens, New York

As a Full Stack Engineer at Altice USA, you will play a key role in designing, developing, and maintaining our self-service data applications. ...

Paramount Global
New York, New York

As a Senior Data Engineer, you will be part of a multi-functional Global Analytics and Data Science team, collaborating closely with executives, data scientists, and data analysts on various data initiatives, infrastructure, and automation of data workflows. This role will be responsible for growing...

MongoDB
New York, New York

The Data Pipelines Engineering team is responsible for building ETL pipelines that populate the Internal Data Platform, which drives analytics that help the company run more efficiently. Work with Data Analysts and Data Scientists to understand and make available the data that is important for their...