Linux / Unix
- Big Data
- Powershell / Bash
- Cloud Data Hub (CDH)
- CDEC Blueprint
- Expertise in data modelling with Oracle SQL.
- Exceptional analytical
complex data sets.
- Thorough testing and data validation to ensure the accuracy of data transformations
Knowledge of data formats such as Parquet, AVRO, JSON, XML, CSV, etc.
- Experience with Data Quality
APIs.
- Experience building data pipelines using AWS Glue or Data Pipeline, or similar platforms.
Our client requires the services of a Data Engineer/Scientist ( Expert) - Midrand/Menlyn/Rosslyn/Home as possible ROLE: Data Engineers are responsible for building and maintaining Big Data Pipelines using using GROUP Data Platforms. Data Engineers are custodians of data and must ensure that data is shared in line Boto3 ETL Docker Linux / Unix Big Data Powershell / Bash GROUP Cloud Data Hub (CDH) GROUP CDEC Blueprint Knowledge of data formats such as Parquet, AVRO, JSON, XML, CSV. Experience working with Data Quality Tools
Our client requires the services of a Data Scientist/Engineer (Entry) – Midrand/Menlyn/Rosslyn/Home Office
PE010814-LM-1 The purpose of the Data Engineer is to leverage their data expertise and data related technologies technologies, in line with the company's Data Architecture Roadmap, to advance technical thought leadership for the purpose data products, and support data initiatives. In addition, Data Engineers enhance the data infrastructure intelligence by providing clean, usable data to stakeholders. They also create data pipelines, Ingestion, provisioning and solutions around big data that support the Bank's strategy to become a data driven organisation. Job
The purpose of the Data Engineer is to leverage their data expertise and data related technologies, in
in line with the companys Data Architecture Roadmap, to advance technical thought leadership for the
purpose data products, and support data initiatives. In addition, Data Engineers enhance the data infrastructure
intelligence by providing clean, usable data to stakeholders. They also create data pipelines, Ingestion, provisioning
solutions around big data that support the Bank's strategy to become a data driven organisation.
as possible ROLE: Data Engineers are responsible for building and maintaining Big Data Pipelines using using GROUP Data Platforms. Data Engineers are custodians of data and must ensure that data is shared in line Boto3 ETL Docker Linux / Unix Big Data Powershell / Bash GROUP Cloud Data Hub (CDH) GROUP CDEC Blueprint Business Intelligence (BI) Experience Technical data modelling and schema design (“not drag and drop”) Boto3 ETL Docker Linux / Unix Big Data Powershell / Bash GROUP Cloud Data Hub (CDH) GROUP CDEC Blueprint
JHB001726-KK-1 Join our IT Hub South Africa as a Data Scientist-AI Platform, focusing on innovative AI
ETL
- Docker
- Linux / Unix
- Big Data
- PowerShell / Bash
Advantageous
Expertise in data modelling with Oracle SQL
- Analytical skills for large and complex data sets
-
/>- Thorough testing and data validation
- Strong written and verbal communication
- Ability
(Confluence, JIRA, etc.)
- Knowledge of Cloud Data Hub (CDH) and CDEC Blueprint
- Development
Knowledge of data formats (Parquet, AVRO, JSON, XML, CSV, etc.)
- Experience with Data Quality Tools
Linux / Unix
- Big Data
- PowerShell / Bash
- Cloud Data Hub (CDH)
- CDEC Blueprint
data formats such as Parquet, AVRO, JSON, XML, CSV, etc.
- Experience working with Data Quality
REQUIREMENTS:
- Demonstrated expertise in data modelling with Oracle SQL.
- Exceptional analytical
complex data sets.
- Perform thorough testing and data validation to ensure the accuracy of data transformations
multi-task.
- Experience building data pipelines using AWS Glue or Data Pipeline, or similar platforms.
Reference: JHB001787-GuguN-1 Job Description: The Data Scientist will focus on developing cost variance variance use cases as part of Exxaro Resources' Data & AI Strategy. This position entails working onsite statistical and machine learning techniques to drive data-driven decision making. - Present findings to both What We're Looking For: - Experience:4-6 years in data science roles. - Skills: Proficiency in Python, analysis and modeling techniques. Experience with data visualization tools (e.g., Tableau, Power BI). -