Better Labels, Better Models: What a New Dataset Reveals About AI in Clinical Trials

By
PERLE TEAM
5.30.2025

A new dataset paper, Automatically Labeling $200B Life-Saving Datasets: A Large Clinical Trial Outcome Benchmark, by Gao, Pradeepkumar, Das, Thati, and Sun, offers an important lens on the role of labeling in high-stakes AI development-particularly in healthcare, where the cost of getting it wrong is high.

The authors used a mix of heuristics, financial data, and large language models to generate outcome labels for over 125,000 clinical trials. But what stands out isn't the scale-it's the method of refinement. Instead of accepting noisy auto-labels as-is, they built a feedback loop where LLMs actively revised, aligned, and improved labels based on a trusted human-verified subset.

It's a powerful signal for where AI data practices are headed.

What Clinical Trial Outcomes (CTO) Got Right: Labeling as Iteration, Not a One-Off Step

Rather than treating labels as static inputs, the team designed an iterative, LLM-in-the-loop process that revisits and revises labels through multiple passes. By aligning noisy labels to a small gold-standard set and using models to critique and regenerate outputs, they were able to achieve a 91 F1 score across trial phases-without relying on full manual annotation.

That shift-from labeling as a front-loaded task to a dynamic feedback loop-mirrors the emerging best practices we're seeing across domains.

Label Quality Is a Leverage Point, Not a Bottleneck

In clinical trials, small differences in label precision can have outsize effects on downstream models-biasing predictions, undermining trust, or failing to surface critical insights. The CTO paper highlights a key tension: as AI moves into more regulated and sensitive domains, we can't afford to treat labeling as a background task.

The stakes are too high for "mostly correct" data to be good enough.

Human Oversight, Strategically Applied

While the CTO pipeline was largely automated, it didn't cut humans out of the loop. Instead, expert-verified labels acted as the anchor-the objective standard against which automated outputs were iteratively improved. This kind of strategic human involvement helps ground machine-driven pipelines and mitigate drift over time.

Final Thoughts

CTO, by Gao et al., makes a compelling case that we're entering a new era of dataset design-one where automation is powerful, but not standalone. As models grow more capable, the quality of their inputs becomes an even greater determinant of performance, particularly in high-stakes applications.

At Perle, we see this firsthand: the most robust systems come from pairing automation with domain expertise-especially where ambiguity and risk are highest.

In AI, good labels aren't just a starting point-they're an edge.

References
Gao, C., Pradeepkumar, J., Das, T., Thati, S., & Sun, J. (2024). Automatically Labeling $200B Life-Saving Datasets: A Large Clinical Trial Outcome Benchmark. arXiv preprint arXiv:2406.10292.

Get in touch

Learn how
Perle can help 

No matter how specific your needs, or how complex your inputs, we’re here to show you how our  innovative approach to data labelling, preprocessing, and governance can unlock Perles of wisdom for companies of all shapes and sizes. 

You can unsubscribe from these communications at any time. For more information on how to unsubscribe, our privacy practices, and how we are committed to protecting and respecting your privacy, please review our Privacy Policy.

By clicking submit, you consent to allow Perle to store and process the personal information submitted above to provide you the content requested.