Perfil del Candidato:
* Parallel distributed processing based on the MapReduce paradigm with Spark or Apache Hadoop.
* Experience or knowledge with any of the main Big Data distributions: Cloudera, Hortonworks, MapR etc.
* Distributed processing of real-time data with Storm or Spark Streaming.
* Experience or knowledge of the tools available in the Hadoop ecosystem: Hive, Flume, Zookeeper, Sqoop, etc.
* Knowledge of NoSQL databases: Cassandra, MongoDB, HBase, Redis, Aerospike, etc.
* Cluster Resources Management: Mesos, YARN.
* Experience with recommendation engines and machine learning algorithms: Apache Mahout, MLlib, etc.
* CEP engines such as Esper CEP or Siddhi.
* Experience with virtualization and cloud deployment AWS, JCloud, Docker.
* Experience with search engines based on Lucene such as Elasticsearch.