Data Scientist (Senior) - 0659 EvdB
Mediro ICT
Date: 3 days ago
City: Johannesburg, Gauteng
Contract type: Full time

Implements analytical algorithms for mixed integer linear programming problems.
Responsible for large scale calculations of core datasets used in a production environment.
Deep understanding of mathematical models and statistical frameworks for machine learning optimization
Comfortable in large scale parallel computing environments
Is responsible for discovering insights and identifying business opportunities using algorithmic, statistical, and mining techniques
Plays a strategic role in creative ideas for leveraging the business’ vast collection of data into new ground-breaking IT solutions
Establishes accurate and scalable analytics systems
Interprets the results of statistical and predictive experiments and regression analysis and integrates this into complex business processes
Conducts customer specification studies, gathers requirements, conducts system architectural design and turns the requirements into a final product
Provides complete application lifecycle development, deployment, and operations support for Big Data solutions and infrastructure
Collaborates with various team members and facilitates the development, automation, and seamless delivery of analytics solutions into Big Data clusters
Imports and exports data using Sqoop from HDFS to Relational Database Systems and vice-versa
Codes and tests the Standardisation, Normalisation, Load, Extract and AVRO models to filter / massage the data and its validation
Installs, configures, and uses ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume
10+ years extensive experience within the data science industry in a fast-paced and complex business setting and with leading teams
Extensive experience in Machine Learning solving business problems
Master’s degree (related to data engineering, statistics, mathematical modelling)
Preferable PhD Degree (related to data engineering, statistics, mathematical modelling)
Ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume
Strong background in mathematics and good analytical and problem-solving skills.
Data Science and Mathematical competency. General AWS experience in the DataScience / big data context. Ability to code in Python and R
Between 7 - 10 Years
Responsible for large scale calculations of core datasets used in a production environment.
Deep understanding of mathematical models and statistical frameworks for machine learning optimization
Comfortable in large scale parallel computing environments
Is responsible for discovering insights and identifying business opportunities using algorithmic, statistical, and mining techniques
Plays a strategic role in creative ideas for leveraging the business’ vast collection of data into new ground-breaking IT solutions
Establishes accurate and scalable analytics systems
Interprets the results of statistical and predictive experiments and regression analysis and integrates this into complex business processes
Conducts customer specification studies, gathers requirements, conducts system architectural design and turns the requirements into a final product
Provides complete application lifecycle development, deployment, and operations support for Big Data solutions and infrastructure
Collaborates with various team members and facilitates the development, automation, and seamless delivery of analytics solutions into Big Data clusters
Imports and exports data using Sqoop from HDFS to Relational Database Systems and vice-versa
Codes and tests the Standardisation, Normalisation, Load, Extract and AVRO models to filter / massage the data and its validation
Installs, configures, and uses ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume
10+ years extensive experience within the data science industry in a fast-paced and complex business setting and with leading teams
Extensive experience in Machine Learning solving business problems
Master’s degree (related to data engineering, statistics, mathematical modelling)
Preferable PhD Degree (related to data engineering, statistics, mathematical modelling)
Ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume
Strong background in mathematics and good analytical and problem-solving skills.
Data Science and Mathematical competency. General AWS experience in the DataScience / big data context. Ability to code in Python and R
Between 7 - 10 Years
See more jobs in Johannesburg, Gauteng