What you'll do...
Position : Senior Data Engineer
Job Location : 603 Munger Ave, Dallas, TX 75202
Duties : Data Strategy : Understands, articulates, and applies principles of the defined strategy to routine business problems that involve a single function.
Data Transformation and Integration : Extracts data from identified databases. Creates data pipelines and transform data to a structure that is relevant to the problem by selecting appropriate techniques.
Develops knowledge of current data science and analytics trends. Data Source Identification : Supports the understanding of the priority order of requirements and service level agreements.
Helps identify the most suitable source for data that is fit for purpose. Performs initial data quality checks on extracted data.
Data Modeling : Analyzes complex data elements, systems, data flows, dependencies, and relationships to contribute to conceptual, physical, and logical data models.
Develops the Logical Data Model and Physical Data Models including data warehouse and data mart designs. Defines relational tables, primary and foreign keys, and stored procedures to create a data model structure.
Evaluates existing data models and physical databases for variances and discrepancies. Develops efficient data flows. Analyzes data-related system integration challenges and proposes appropriate solutions.
Creates training documentation and trains end-users on data modeling. Oversees the tasks of less experienced programmers and stipulates system troubleshooting supports.
Code Development and Testing : Writes code to develop the required solution and application features by determining the appropriate programming language and leveraging business, technical, and data requirements.
Creates test cases to review and validate the proposed solution design. Creates proofs of concept. Tests the code using the appropriate testing approach.
Deploys software to production servers. Contributes code documentation, maintains playbooks, and provides timely progress updates.
Problem Formulation : Translates business problems within one's discipline to data related or mathematical solutions. Identifies what methods (for example, analytics, big data analytics, automation) would provide a solution for the problem.
Shares use cases and gives examples to demonstrate how the method would solve the business problem. Applied Business Acumen : Provides recommendations to business stakeholders to solve complex business issues.
Develops business cases for projects with a projected return on investment or cost savings. Translates business requirements into projects, activities, and tasks and aligns to overall business strategy.
Serves as an interpreter and conduit to connect business needs with tangible solutions and results. Recommends new processes and ways of working.
Data Governance : Establishes, modifies, and documents data governance projects and recommendations. Implements data governance practices in partnership with business stakeholders and peers.
Interprets company and regulatory policies on data. Educates others on data governance processes, practices, policies, and guidelines.
Provides recommendations on needed updates or inputs into data governance policies, practices, or guidelines.
Minimum education and experience required : Bachelor’s degree or the equivalent in Computer Science or a related field plus 3 years of experience in software engineering or related experience OR Master’s degree or the equivalent in Computer Science or a related field plus 1 year of experience in software engineering or related experience.
- Skills required : Must have experience with : Designing, developing, implementing and tuning distributed data processing pipelines that process large volumes of data and data management on Azure Databricks by using Pyspark, Databricks, Azure cloud storage and Python;
- Designing, developing, implementing and tuning distributed data processing pipelines that process large volume of data and data management on BigQuery by using Pyspark, Dataproc, BigQuery, GCS and Python;
- Orchestrating and monitoring data pipelines using Airflow; designing and implementing Data warehousing solutions using Databricks and building visualization dashboards using BI tools;
- Testing automation and code coverage using PyTest and Python; CICD process implementation using Git, SonarQube and Hygieia;
- Creating Epics, user stories, required documentation, tracking the user stories for project management using JIRA, Confluence and Agile delivery methodology;
- Preparing technical specification documents, functional designs, business process diagrams, interfaces, system dependencies, component interactions, data types / structures and code standards using Confluence;
- Creating scalable and easily extendable Logical and Physical data models; Data analysis using SQL queries and BI Tools; Data pipelines coding in Python programming language;
and Designing and developing data pipelines using ETL tools, Stored Procedures and SQL scripts. Employer will accept any amount of experience with the required skills.
LI-DNP #LI-DNI
Wal-Mart is an Equal Opportunity Employer.