Back to Job Search

Job Description

My client is looking for an experienced Data Engineer to assist on what will initially be a 6-month long contract assignment.

The client in question is partnering with their IT department to deliver an ecosystem of curated, enriched and protected sets of data - created from global, raw, structured and unstructured sources. The company has over 300 sources and a rapidly growing book of work. We are utilising the latest technologies to solve business problems and deliver value and truly unique insights.

Key Responsibilities:
Data Engineer:

  • Ability to process and rationalize structured data, message data and semi/unstructured data and ability to integrate multiple large data sources and databases into one system
  • Proficient understanding of distributed computing principles and of the fundamental design principles behind a scalable application
  • Strong knowledge of the Big Data ecosystem, experience with Hortonworks/Cloudera platforms
  • Practical experience in using HDFS
  • Practical expertise in developing applications and using querying tools on top of Hive, Spark (PySpark)
  • Strong Scala skills
  • Experience in Python, particularly the Anaconda environment and Python-based ML model deployment
  • Experience of Continuous Integration/Continuous Deployment (Jenkins/Hudson/Ansible)
  • Experience with using GIT/GITLAB as a version control system.
  • Experience in working in Teams using the Agile Methods (SCRUM) and Confluence/JIRA
  • Good communication skills (written and spoken), ability to engage with different stakeholders and to synthesise different opinions and priorities.

Nice to Haves

  • Knowledge of at least one Python web framework (preferably: Flask, Tornado, and/or twisted)
  • Basic understanding of front-end technologies, such as JavaScript, HTML5, and CSS3 would be a plus
  • Good understanding of global markets, markets macrostructure and macroeconomics
  • Knowledge of Elastic Search Stack (ELK)
  • Experience with Google Cloud Platform (Data Proc / Dataflow)

Domain Knowledge:

  • Knowledge of and experience using data models and data dictionaries in a Banking and Financial Markets context. Knowledge of Trade Finance or Securities Services particularly useful.
  • Knowledge of one or more of the following domains (including market data vendors):
    • Party/Client
    • Trade
    • Settlements
    • Payments
    • Instrument and pricing
    • Market and/or Credit Risk

Need Experience using below languages/tools:

  • Java
  • HQL, SQL
  • Querying tools on top of Hive, Spark (PySpark)
  • Scala
  • Python, particularly the Anaconda environment
  • GIT/GITLAB as a version control system
If you are interested in this role please apply below or contact me for more information. 

Eames Consulting is acting as an Employment Business in relation to this vacancy.