• Login Name       Password       Remember me       LOGIN LOGIN    FORGOT PASSWORD


Data Engineer ( 12-month contract )

Job Ref
Job Type
Employer Type
Date Added 17 Aug 2021
Expiry Date 14 Sep 2021
* There has been 1 application to this job.
* This job has been viewed 3426 times.

Western Cape

Market related


Role details:
Data Engineers build and support data pipelines and datamarts built off those pipelines. Both must be scalable, repeatable and secure. The Data Engineer helps to facilitate gathering data from a variety of different sources, in the correct format, assuring that it conforms to data quality standards and assuring that downstream users can get to that data timeously. This role functions as a core member of an agile team. These professionals are responsible for the infrastructure that provides insights from raw data, handling and integrating diverse sources of data seamlessly. They enable solutions, by handling large volumes of data in batch and real-time by leveraging emerging technologies from both the big data and cloud spaces. Additional responsibilities include developing proof of concepts and implements complex big data solutions with a focus on collecting, parsing, managing, analysing and visualising large datasets. They know how to apply technologies to solve the problems of working with large volumes of data in diverse formats to deliver innovative solutions. Data Engineering is a technical job that requires substantial expertise in a broad range of software development and programming fields. These professionals have a knowledge of data analysis, end user requirements and business requirements analysis to develop a clear understanding of the business need and to incorporate these needs into a technical solution. They have a solid understanding of physical database design and the systems development lifecycle. This role must work well in a team environment.
- - - - - - - - - - - - - -
Design and develop data feeds from an on-premise environment into a datalake environment in an AWS cloud environment. Design and develop programmatic transformations of the solution, by correctly partitioning, formatting and validating the data quality. Design and develop programmatic transformation, combinations and calculations to populate complex datamarts based on feed from the datalake. Provide operational support to datamart datafeeds and datamarts. Design infrastructure required to develop and operate datalake data feeds. Design infrastructure required to develop and operate datamarts, their user interfaces and the feeds required to populate the datalake.

IT Degree/Diploma (3 years) AWS Certification at least to associate level
Business Intelligence (5+ years) Extract Transform and Load (ETL) processes (5+ years) Cloud AWS (2+ years) Agile exposure, Kanban or Scrum (2+ years) Desirable: Retail Operations (5+ years) Big Data (2+ years) Experience working as a Technical Lead in the space
Knowledge: Creating data feeds from on-premise to AWS Cloud (2 years) Support data feeds in production on break fix basis (2 years) Creating data marts using Talend or similar ETL development tool (4 years) Manipulating data using python and pyspark (2 years) Processing data using the Hadoop paradigm particularly using EMR, AWS’s distribution of Hadoop (2 years) Devop for Big Data and Business Intelligence including automated testing and deployment (2 years) Skills: Talend (1 year) AWS: EMR, EC2, S3 (1 year) Python (1 year) Business Intelligence Data modelling (3 years) SQL (3 years) Desirable PySpark or Spark (1 year)