Back to Jobs

Data Engineer (Bio-Technology background is a must have)

Remote, USA Full-time Posted 2025-07-27

Data Engineer (Bio-Technology, Python, AWS, SQL, Terraform, Airflow, and Pandas libraries)
Location: 100% Remote
Duration: 6+ Months Contract

    You will:
  • Lead the architecture, design, delivery & deployment of core data platforms, data warehouse and data modeling needs.
  • Exceptional understanding on various data topics w.r.t. data engineering i.e. building data pipelines, data modeling , data warehousing etc.
  • Understanding and experience with cloud software development. Specifically AWS Cloud.
  • Contribute to the design and execution of data governance, data quality frameworks.
  • Have a passion and attention to detail for all aspects of data from ingestion, validation/quality, transformation, modeling, storage etc.
  • Interface with various teams from product, laboratory, web services, data science etc.
  • Minimum Requirements:
  • B.S. / M.S. in a quantitative field (e.g. Computer Science, Engineering, Mathematics, Physics, Computational Biology) with at least 6 years of related industry experience, or Ph.D. with at least 4 years of related industry experience
  • Substantial experience in architecting and delivering secure, scalable cloud-based data warehouses / data lakes on AWS, Azure, or Google Cloud Platform
  • Exceptional experience with data modeling principles , patterns and industry trends.
  • Very comfortable in designing and reorganizing facts and dimensions tables, complex data models, SCDs, etc.
  • Solid object-oriented and/or functional programming experience, specifically in Python and GO.
  • Expert with data pipelining and workflow engines, like Apache Airflow, Spark etc., and proven ability to choose the correct frameworks as well as tools depending on the requirements.
  • Experience with provisioning on AWS Cloud, e.g. with Terraform or cloudformation. Leveraging CI in a cloud environment for automation.
  • In depth Experience with relational databases, query authoring, and performance tuning.
  • Ability to take a high-level requirement and decompose that into clear engineering objectives, which can be further evolved into detailed specifications.
  • High emotional quotient to work with potential ambiguity, ask the right questions and engage to drive resolution from requirements to solutions.
  • The following are highly welcome:
  • Proven track record of building and operating scalable data infrastructure, managing data models for hundreds to thousands of tables.
  • Experience with various data products, involvement in build vs buy decisions, designing a solution with limited resources and/or timelines .
  • Experience with DevOps, e.g. CI/CD pipelines, containerized deployment, infrastructure as code, Terraform.
  • Experience building microservices and web applications.
  • Experience with supporting data science / machine learning data pipelines.

Apply Job!

Apply to this Job

Similar Jobs