OUR
INSIGHTS

Blog

Astrix Partner Blog – How to implement the right lab automation stack?

CATEGORY
Blog

DATE
December 2, 2024

Share this...
Share on facebook Share on twitter Share on linkedin Youtube

Astrix works with thought leaders throughout the Life Science industry. This includes a variety of technology partners with whom we collaborate on projects and industry thought leadership.  As part of this we feature high quality content from our partners to help our customers and website visitors learn about new technologies, process and strategy that may help move their R&D operations through a digital transformation!  This partner blog is featured by Ganymede; The lab data and instrument connectivity company

Here’s a 5 step approach for bringing high throughput to your R&D.

The Lab Automation Sandwich

Once biotech organizations have some form of data management/strategy and standardized workflows, they need to build their automation tech stack. In less technical terms, we like to think about this stack as an automation sandwich composed of multiple layers. Yes, a sandwich—not the most glamorous of metaphors, but one that we find accurate and illustrative of how lab automation fits together. Bear with us, we’ll do our best to refrain from any (too many) puns about lunch.

lab automation

There are five layers to every lab automation sandwich:

  1. Systems of Record
  2. Design of Experiment Systems
  3. Physical Automation
  4. Data Automation 
  5. Systems of Record (yes, again! We’ll explain why below.)

In this post we’ll walk through each of these layers step-by-step, discussing why they matter and how each contributes to an automation-ready tech stack.

Step One: Establish Systems of Record

At the top and bottom of every automated lab, there is a system of record. Think of this as the bottom layer of bread that lays the foundation for the whole sandwich. This crucial layer of software serves as a shared repository of knowledge within a lab. It’s where scientists generate and annotate experiments, and where everyone within the lab has access to experimental data, analyses, and more. Critically, this layer is often the source of key pieces of metadata that serve as the unifying thread throughout each automated workflow.

Examples of systems of record include:

  • Electronic lab notebooks (ELNs), such as Scinote and Benchling
  • Laboratory information management systems (LIMS), such as STARLIMS and ThermoFisher’s SampleManager
  • Manufacturing execution systems (MES) such as Apprentice and MasterControl

These days, most labs already have some piece of system of record software in place. While there might still be record-keeping happening by hand on paper, that information will eventually make it into an ELN, LIMS or MES through manual entry (although whether that information is entered correctly or with full context is another question).

Regardless, without a system of record, labs can’t take advantage of automation. So if you’re new to lab automation, start with finding the right system of record for your organization. And if you already have one in place, it’s time to move on to step two.

Step Two: Set Up Design of Experiment Systems

A lab’s data is only as good as the experiment that generated it. Once a lab has implemented a system of record, the next step is to find the right design of experiment software.

With this software layer, labs can lay out experimental plans in a comprehensive manner, accounting for variables, each step in the workflow, and other important criteria. If the ELN or LIMS system is where a lab’s data goes for sharing and analysis, design of experiment software is where labs ensure that data is generated in a high-quality, repeatable, and statistically powerful way (we’re looking at you p-value hackers).

By codifying a given experiment into a set of repeatable processes that lab operators can understand, design of experiment software enables automation engineers and scientists to translate that SOP into a machine-readable method/protocol. A good example of this kind of software is Synthace or JMP.

Step Three: Invest in Physical Automation

Now that your lab has a place to share data AND generate experimental designs, it’s time to actually do some lab work. This is the step where real data gets generated through real workflows and processes. Welcome to the workflow automation layer, where there’s actual, well, automation.

If you’re still thinking of our sandwich analogy, the lab automation layer is the thickest part of the whole meal—it’s all the options you have for the middle part of the whole stack. Rather than a single piece of software, this step is actually composed of lots of different elements of software and hardware and protocols, including:

  • Robotic elements, such as liquid handlers that automate sample prep or more complex works cells that combine liquid handlers with analytical instruments and robots
  • Laboratory execution systems (LESs), like Cellario, Momentum, and Apprentice’s Tempo LES that enable scientists and automation engineers to define workflows and run experiments on them
  • Digital orchestration platforms, like Artificial and Green Button Go that manage work between manual and automated sources
  • And more

Step Four: Automate Your Data

Once a lab is generating data from experiments running at scale, some big questions arise about the storage, management, and quality of that data. All too often, data comes off instruments and becomes siloed through old school processes, such as downloading data to a USB or entering data in Excel. Even in cases where data is automatically uploaded from an instrument to cloud storage or to an ELN or LIMS, it’s essentially a raw data dump. Sometimes the files are locked in a proprietary OEM format, and they often lack detailed metadata that provides the context for the experiment. The result is that most labs have data scattered all over the place.

Let’s think about 3 common considerations for labs when they purchase traditional, physical lab automation systems to illustrate this point further.

  1. Higher experimental throughput – how can you expect to meaningfully introduce higher throughput with the low throughput data processing?
  2. Flexible enough to match workflows – are the processes really that flexible if the data pipelines for handling them are static and hard coded?
  3. Standardize complex workflows – how standardized can the protocols be with manual data management steps in between, let alone the lack of traceability?

That’s why labs should think about data automation as their fourth layerIn this step, platforms like Ganymede apply automation to the data itself. This can look like automatically:

  • Pulling data from instruments, such as flow cytometers
  • Cleaning that data and annotating it with crucial metadata
  • Storing and structuring data in the cloud
  • Streamlining data management in and around analysis
  • Automating the entry of key results into your ELN/LIMS/MES
  • Customizing and updating your data pipelines to adapt to your physical workflows
  • And more

Without data automation, you’re unable to handle the reams of data generated by your robotic hardware and orchestration systems.  By automating data, labs can make discoveries faster and operate their labs much more efficientlyEvery speck of data is FAIR—findable, accessible, interoperable, and reusable. . At Ganymede, we’ve actually found that automating data saves scientists 20 percent of their time at the bench and helps life science companies move 10 percent faster to IND or to market

Step Five: Feed Back Into Your Systems of Record

Every sandwich needs a top layer and a bottom layer to hold everything together. That’s why systems of record come back to serve as another slice of bread in this analogy. Once data management is automated, all that FAIR, beautiful data can be fed back into your lab’s ELN, LIMS, or MES. With well-organized, high-quality data at their fingertips, everyone—from the IT department to developers to lab managers to bench scientists—is working from the same version of the truth.

Lab Automation: The Future of Science

Science is more complex and interdisciplinary than ever before. Life science labs today are working across biology, chemistry, and more. The data they generate is some of the most valuable in the world, with the potential to cure diseases and bring new hope to patients. These ground-breaking labs are also facing the same pressures as many other industries: the mandate to do more with less, move faster than their competitors, and deliver value to patients and the market.

Data automation can help companies on all fronts. When data is intelligently generated, organized, and analyzed, labs have a better view of their science. And with a better view of their science, they have a better chance of seeing the next big discovery clearly. Labs should invest in their ideal sandwich-style tech stack, thoughtfully integrating each of the five layers described in this post. And if you’ve read this before lunch, we hope we didn’t make you too hungry, and if you’re reading this after lunch, we think we have a pretty good idea of what you’ll be having for dinner 😀.

About Ganymede

Ganymede connects lab instruments, apps like ELN/LIMS, and other data sources in a single FAIR data cloud. Go beyond SDMS; our data infrastructure and analysis automation allows for automating data capture, entry, and analysis for everything from flow cytometry to HPLC to LIMS to Veeva. Think of it as a data backend-in-a-box for labs, driving scientist time savings, data integrity and compliance, and forming a rich data foundation for AI and scientific insights.

About Astrix

Astrix is the unrivaled market-leader in creating & delivering innovative strategies, solutions, and people to the life science community.  Through world class people, process, and technology, Astrix works with clients to fundamentally improve business & scientific outcomes and the quality of life everywhere. Founded by scientists to solve the unique challenges of the life science community, Astrix offers a growing array of strategic, technical, and staffing services designed to deliver value to clients across
their organizations.

Astrix work with Ganymede

Astrix provides comprehensive  services to maximize lab efficiency and improve time to result for Ganymede.  Core services include:

  • Implementation: Our team can help get your lab up and running quickly on Ganymede with our proven implementation methodology.
  • Upgrades: We can help you stay current with the latest features and functionality of Ganymede by upgrading your system as new releases become available.
  • Strategic Planning: Our team can help develop a long-term strategy for using Ganymede to its fullest potential in your lab.
  • Validation: We have the experience and expertise to help you validate your Ganymede system in accordance with industry standards.
  • Training: We offer comprehensive training on all aspects of using Ganymede in your lab.

LET´S GET STARTED

Contact us today and let’s begin working on a solution for your most complex strategy, technology and staffing challenges.

CONTACT US
Web developer Ibiut