Advanced Knowledge and Experience:
*The Python programming language (3 years experience)
*The Hadoop ecosystem: MapReduce, MapReduce Streaming, HDFS, Oozie (2 years experience)
*Unix systems, preferably Debian/Ubuntu. Comfortable in a Linux terminal and with bash scripting (2 years experience)
*Python modules,(1-2 years experience)
*MongoDB (1 year experience).
*Apache Spark using the PySpark API: (0,5 -- 1 year experience)
*Git for version control
*JIRA for task management
*Experience with python modules for Data Science / Data processing
*Experience with (Hadoop) Job scheduling using a tool such as Oozie, Azkaban or Luigi. You know
how to build pro-active monitoring systems (alerts in case a job fails).
*Python engineering skills: familiarity with debugging tools, Pylint for static
code checking and you work according to Test-Driven Development principles.