Data Scientist (Senior) - 0659 EvdB

Mediro ICT


Date: 3 days ago
City: Johannesburg, Gauteng
Contract type: Full time
Implements analytical algorithms for mixed integer linear programming problems.

Responsible for large scale calculations of core datasets used in a production environment.

Deep understanding of mathematical models and statistical frameworks for machine learning optimization

Comfortable in large scale parallel computing environments

Is responsible for discovering insights and identifying business opportunities using algorithmic, statistical, and mining techniques

Plays a strategic role in creative ideas for leveraging the business’ vast collection of data into new ground-breaking IT solutions

Establishes accurate and scalable analytics systems

Interprets the results of statistical and predictive experiments and regression analysis and integrates this into complex business processes

Conducts customer specification studies, gathers requirements, conducts system architectural design and turns the requirements into a final product

Provides complete application lifecycle development, deployment, and operations support for Big Data solutions and infrastructure

Collaborates with various team members and facilitates the development, automation, and seamless delivery of analytics solutions into Big Data clusters
Imports and exports data using Sqoop from HDFS to Relational Database Systems and vice-versa

Codes and tests the Standardisation, Normalisation, Load, Extract and AVRO models to filter / massage the data and its validation

Installs, configures, and uses ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume

10+ years extensive experience within the data science industry in a fast-paced and complex business setting and with leading teams

Extensive experience in Machine Learning solving business problems

Master’s degree (related to data engineering, statistics, mathematical modelling)

Preferable PhD Degree (related to data engineering, statistics, mathematical modelling)

Ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume

Strong background in mathematics and good analytical and problem-solving skills.

Data Science and Mathematical competency. General AWS experience in the DataScience / big data context. Ability to code in Python and R

Between 7 - 10 Years
Post a CV