What you'll do :
- Develop Data Pipelines using Scala / Java and Spark EMR to move data into our consolidated data lake or data warehouse (Snowflake)
- Consolidate / join datasets to create easily consumable, performant, and consistent information
- Look for ways to improve processes and take initiative to implement them
- Evaluate new technology and advise on our data lake ecosystem
- Collaborate with cross-functional teams to design and implement impactful solutions to department and business problems
- Creates and maintains documentation for business end users and other data analysts.
- Determine where in our infrastructure we should house our data based on the use case and data model.
What you'll bring :
- Degree (Masters preferred) in Computer Science, Information Systems, Data Science, or related field and 4+ years of experience in data engineering, or an equivalent combination of education and experience.
- Expert and Hands-On development in distributed computing environment using Hadoop, Scala / Java, Spark and Apache Airflow
- SQL knowledge and understanding
- Experience architecting solutions in collaboration with engineering development and data analysts team.
- Experience working with third party APIs for data collection
- Self-motivated; able to work independently to complete tasks and collaborate with others to identify and implement solutions.
- Expertise and Hand-On working knowledge of AWS
- Experience with Git / Github or other version control systems
- Familiarity with Snowflake or DataBricks a plus.
LI-SK
LI-Hybrid
Actual compensation offered will be based on factors such as the candidate’s work location, qualifications, skills, experience and / or training.
Your recruiter can share more information about the specific salary range for your desired work location during the hiring process.
We want our employees and their families to thrive.
In addition to comprehensive benefits we offer holistic mind, body and lifestyle programs designed for overall well-being.
30+ days ago