Graph Database Engineer – Neo4j – REMOTE – 5064

Our direct client, a growing life sciences software company, is seeking a passionate Data Engineer who has experience processing terabytes of data. The ideal candidate will have experience in creating and automating scalable, fault-tolerant and reproducible data pipelines using Amazon AWS technologies. The incumbent will also work with graph databases so experience with Neo4j or similar databases is required.

What You’ll Do:

  • Build & operate automated ETL pipelines that process terabytes of text data nightly
  • Develop service frontends around our various backend datastores (AWS Aurora MySQL, Elasticsearch, S3)
  • Perform technical analyses and requirements specification with our product team on data service integrations
  • Help customers bring their data to the platform

What You Know:

Must Haves:

  • Python 3 or Java programming experience, preferably both
  • Knowledge of Graph databases such as Neo4j or Neptune
  • Day-to-day experience using AWS technologies such as Lambda, ECS Fargate, SQS, & SNS
  • Experience building and operating cloud-native data pipelines
  • Experience extracting, processing, storing, and querying of petabyte-scale datasets
  • Familiarity with building and using containers
  • Familiarity with event-based microservices

Nice-to-Haves:

  • Prior experience with Elasticsearch (custom development and/or administration) is a huge plus
  • Prior work with text and natural-language processing

Aside from the technologies, our client is seeking an incumbent who considers themselves a self-starter, well-organized, and strives to grow personally and professionally, beyond just expanding technical abilities. In this role, you will be encouraged to experiment with new technology and share knowledge with the team. If this sounds like you, apply today for immediate consideration.

Related Post