Big Data Engineer Reference number: 2182

  • Online since: 26-01-2022
  • 1 year + extension
  • Start: 28-02-2022 | End: 31-12-2022
  • Work experience: demonstrable relevant 3+ years
  • Job type: Employee/Contract
  • Category: Data science

Job description

As a Data Engineer, you will play a key role preparing the infrastructure and data that will be used to deliver high quality data products. You will help us design, develop and maintain data pipelines that will deliver insights. By using a DevOps approach, you will make sure the overall system is running on all times by automating tasks so you can spend time on creating and not deploying. You will also make sure the system is appropriately tested and monitored by using adapted methods and tools. You will collaborate with the other data engineers and data scientists of the Advanced Analytics team to create the simplest possible effective data landscape to improve delivery speed of future AI use cases.

What you'll do

  • Conceive and build data architectures
  • Participate in the short/mid/long term vision of the overall system
  • Simplify & optimize existing pipelines if needed
  • Execute ETL (extract/transform/load) processes from complex and/or large data sets
  • Ensure data are easily accessible and that their exploitation is performing as requested, even in highly scalable circumstances
  • Participate to the architecture and planning of the big data platform to optimize the ecosystem’s performances
  • Create large data warehouses fit for further reporting or advanced analytics
  • Collaborate with machine learning engineers for the implementation, deployment, scheduling and monitoring of different solutions
  • Ensure robust CI/CD processes are in place
  • Promote DevOps best practices in the team

Profile

  • You’re quality oriented
  • You are multi-disciplined & able to work with divers APIs and understand multiple languages well enough to work with them
  • You are an excellent problem analyser and solver
  • You’re open minded , collaborative, team player, ready to adapt to the changing needs
  • Curiosity about new techniques and tools, eagerness to always keep learning
  • You’re committed to deliver, pragmatic and solution oriented
  • Experience in telecom and/or financial sector is a plus
  • Experience with an agile way of working is a plus
  • Languages : English (very good in reading, writing, speaking) is a must

Required skills

  • Demonstrable relevant (3+ years of) experience with big data platforms (Hadoop, Cloudera, EMR, Databricks, ...)
  • Data pipeline management
  • Cluster management
  • Workflow management ( Oozie, Airflow)
  • Database management of SQL and noSQL databases
  • Large file storage (HDFS, Data Lake, S3, Blob storage,..)
  • Strong knowledge of Scala and Python
  • Strong knowledge & experience in Spark (Scala and Pyspark)
  • Strong knowledge of CI/CD concepts
  • Stream processing such as Kafka, Kinesis, Elasticsearch
  • Good knowledge of a cloud environment
  • High level understanding of data science concepts
  • Knowledge of Data Visualisation framework like Qlik Sense is a plus
Apply now
Tatiana Paiu Recruiter
02 481 52 16
Interested? Send us your resumé

To apply for this job, please complete the form below and join your resume. This instantly places your information into our database. Once we have received your information, we will be in touch by e-mail or phone. If you have not heard from us after 3 working days, please call us!

Thank you for your interest in working with Harvey Nash and we look forward to assisting you in your job search!