Platform Manager: Data Scientist (0659) KG
Mediro ICT
Date: 3 weeks ago
City: Midrand, Gauteng
Contract type: Full time

WHAT WILL BE YOUR ROLE AND RESPONSIBILITIES?
Implements analytical algorithms for mixed integer linear programming problems.
Responsible for large scale calculations of core datasets used in a production environment.
Deep understanding of mathematical models and statistical frameworks for machine learning optimisation
Comfortable in large scale parallel computing environments
Is responsible for discovering insights and identifying business opportunities using algorithmic, statistical, and mining techniques
Plays a strategic role in creative ideas for leveraging the business’ vast collection of data into new ground-breaking IT solutions
Establishes accurate and scalable analytics systems
Interprets the results of statistical and predictive experiments and regression analysis and integrates this into complex business processes
Conducts customer specification studies, gathers requirements, conducts system architectural design and turns the requirements into a final product
Provides complete application lifecycle development, deployment, and operations support for Big Data solutions and infrastructure
Collaborates with various team members and facilitates the development, automation, and seamless delivery of analytics solutions into Big Data clusters
Imports and exports data using Sqoop from HDFS to Relational Database Systems and vice-versa
Codes and tests the Standardisation, Normalisation, Load, Extract and AVRO models to filter / massage the data and its validation
Installs, configures, and uses ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume
WHICH QUALIFICATIONS/EXPERIENCE DO WE NEED FOR THE ROLE?
10+ years extensive experience within the data science industry in a fast-paced and complex business setting and with leading teams
Extensive experience in Machine Learning solving business problems
Master’s degree (related to data engineering, statistics, mathematical modelling)
Preferable PhD Degree (related to data engineering, statistics, mathematical modelling)
ESSENTIAL SKILLS REQUIREMENTS:
Strong background in mathematics and good analytical and problem-solving skills.
Data Science and Mathematical competency. General AWS experience in the DataScience / big data context. Ability to code in Python and R
ADVANTAGEOUS SKILLS REQUIREMENTS:
Ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop , Pig, HDFS , HBase , Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume
Between 7 - 10 Years
Implements analytical algorithms for mixed integer linear programming problems.
Responsible for large scale calculations of core datasets used in a production environment.
Deep understanding of mathematical models and statistical frameworks for machine learning optimisation
Comfortable in large scale parallel computing environments
Is responsible for discovering insights and identifying business opportunities using algorithmic, statistical, and mining techniques
Plays a strategic role in creative ideas for leveraging the business’ vast collection of data into new ground-breaking IT solutions
Establishes accurate and scalable analytics systems
Interprets the results of statistical and predictive experiments and regression analysis and integrates this into complex business processes
Conducts customer specification studies, gathers requirements, conducts system architectural design and turns the requirements into a final product
Provides complete application lifecycle development, deployment, and operations support for Big Data solutions and infrastructure
Collaborates with various team members and facilitates the development, automation, and seamless delivery of analytics solutions into Big Data clusters
Imports and exports data using Sqoop from HDFS to Relational Database Systems and vice-versa
Codes and tests the Standardisation, Normalisation, Load, Extract and AVRO models to filter / massage the data and its validation
Installs, configures, and uses ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume
WHICH QUALIFICATIONS/EXPERIENCE DO WE NEED FOR THE ROLE?
10+ years extensive experience within the data science industry in a fast-paced and complex business setting and with leading teams
Extensive experience in Machine Learning solving business problems
Master’s degree (related to data engineering, statistics, mathematical modelling)
Preferable PhD Degree (related to data engineering, statistics, mathematical modelling)
ESSENTIAL SKILLS REQUIREMENTS:
Strong background in mathematics and good analytical and problem-solving skills.
Data Science and Mathematical competency. General AWS experience in the DataScience / big data context. Ability to code in Python and R
ADVANTAGEOUS SKILLS REQUIREMENTS:
Ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop , Pig, HDFS , HBase , Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume
Between 7 - 10 Years
See more jobs in Midrand, Gauteng