Default_best_data_analytics_tools_promo_image_with_saturated_0 (1)

Locations

  • Bulgaria
  • Georgia
  • Lithuania
  • Mexico
  • Poland
  • Romania

Company Background

The customer is a global leader in diagnostics and drug development, employing over 70,000 professionals and serving clients in more than 100 countries. With over $14 billion in annual revenue, they are committed to advancing healthcare and empowering patients, providers, and researchers through data-driven solutions. Their mission is to improve health and improve lives by delivering clear and confident answers in a complex medical landscape.

Project Description

The project involves building a new internal system to support Laboratory Information Management (LIM). The goal is to modernize and streamline data handling processes, improve data accessibility, and enhance performance across large-scale data workflows. The team is responsible for both designing and implementing a high-performance solution, working closely with the client on architecture decisions, performance tuning, and best practices in data engineering.

Technologies

  • Python
  • Databricks
  • Apache Spark
  • Hive
  • AWS EMR
  • S3
  • Oracle SQL
  • DataStage
  • CI/CD tools
  • Mainframe systems

What You'll Do

  • Design and implement scalable data processing pipelines using Spark, Hive, and Python
  • Collaborate with stakeholders on system architecture, performance tuning, and design decisions
  • Optimize SQL and Spark queries to ensure fast and efficient data access
  • Develop ETL processes and manage data flow across systems using tools like DataStage
  • Contribute to the CI/CD pipeline setup, automation, and deployment strategies
  • Participate in code reviews, documentation, and cross-functional planning meetings
  • Support the data modeling process within a data warehouse environment
  • Work collaboratively with the client’s engineering and architecture teams

Job Requirements

  • 5+ years of experience in Data Engineering or related role
  • 5+ years of hands-on Python development experience
  • Experience with Databricks, Spark, Hive, AWS EMR/S3
  • Proficiency in Oracle SQL and query tuning
  • Familiarity with CI/CD tools and modern DevOps practices
  • Strong understanding of SDLC and software engineering principles
  • Experience with data modeling and ETL in large-scale environments
  • Exposure to mainframe systems is a plus
  • English level: B1+ (spoken and written)

What Do We Offer

The global benefits package includes:

  • Technical and non-technical training for professional and personal growth;
  • Internal conferences and meetups to learn from industry experts;
  • Support and mentorship from an experienced employee to help you professional grow and development;
  • Internal startup incubator;
  • Health insurance;
  • English courses;
  • Sports activities to promote a healthy lifestyle;
  • Flexible work options, including remote and hybrid opportunities;
  • Referral program for bringing in new talent;
  • Work anniversary program and additional vacation days.

Didn't find anything suitable?

We're always starting new projects and we'd love to work with you. Please send your CV and we'll get in touch.

We will be glad to see you!

First Name is required. Maximum 50 characters.
Last Name is required. Maximum 50 characters.
Email is required. Please enter a valid email address (e.g. recipient@domaine.org).
Maximum 100 characters. Add the name of our employee (e.g. John Smith)
Maximum 2000 characters.
Please attach file in the allowed format .pdf, .doc(x), .txt, .rtf Please attach file less than 3 Mb
Formats (3 MB): doc, docx, pdf, txt, rtf
Please Add Comment or Attach File.

An error occurred sending your message.
Try again or contact us via webinforequest@coherentsolutions.com.

Thanks for your application!
We will reply soon.

Share vacancy