AI/ML in Antibody Discovery: What Works, What Doesn’t, What’s Next

Traditional antibody discovery is a war of attrition: screen thousands of candidates, fail late, and repeat. AI/ML changes that calculus by shifting failures from late and expensive to early and cheap. The result is fewer wet-lab cycles and a higher probability that the molecules you carry forward actually behave like drugs, while at the same time integrating learnings from failures at every stage. 

What Works, What Doesn’t, What’s Next

Traditional antibody discovery is a war of attrition: screen thousands of candidates, fail late, and repeat. AI/ML changes that calculus by shifting failures from late and expensive to early and cheap. The result is fewer wet-lab cycles and a higher probability that the molecules you carry forward actually behave like drugs, while at the same time integrating learnings from failures at every stage. 

 

Where It Works Today: Intercepting Failure Early

Developability triage: 


Most molecules fail not because they don’t bind, but because they’re sticky, aggregation-prone, or immunogenic. ML models trained on sequence patterns and combined with AI structure prediction, can flag these liabilities before a single experiment is run. The result is fewer late-stage surprises by avoiding problematic patterns from the start.

Optimization of the lead sequence:

Getting from a hit to a balanced lead typically means multiple rounds of make-test-measure. ML helps on many fronts: protein language models propose variants that follow antibody “grammar”, inverse folding ensures they’re structurally viable and Bayesian optimization maximizes the return of information in every cycle. Panels of these computationally proposed variants, additionally triaged for developability in silico, are both smaller and compress the number of iteration cycles to as few as 1-2. The same logic applies to humanization: rather than relying solely on traditional grafting, AI/ML proposes and ranks variants that increase humanness without sacrificing binding or introducing new liabilities.

NGS selection:

A typical NGS repertoire derived from animal tissues contains millions of sequences, and without computational support, selection rests on simple rules and gut instinct, requiring large candidate panels just to avoid leaving the best binders behind. AI/ML-enabled annotation, including clustering, early developability scoring or even predicted binding modes, turns that choice into a data-driven one. In practice, this can reduce the number of candidates down to 96 while preserving or improving both diversity and hit quality.
None of these are moonshots. They’re incremental compressions of time and failure rate, applied at the points where discovery programs traditionally bleed the most in ways that can often remain invisible until much later.

Integrating AI/ML for Smarter Antibody Discovery

Practical, production-grade AI/ML workflows for candidate selection, developability triage, and optimization — covering what works today, what doesn’t, and how integrated data infrastructure turns models into compounding advantage.

Where it’s still brittle: De Novo Design

The boldest application of AI/ML is the emerging design of binders from scratch, in silico. Progress in this space is real and the pace of published results has accelerated sharply over the past two years with diffusion-based generative models producing novel candidates in regions of sequence space that incremental mutation would never reach. 

The benchmark isn’t “can it propose a binder?” It is “Does it perform like a reliable drug discovery engine across target classes?” By that standard, de novo design is still an active area of research. Scaffolds are often well-trodden, binding strength is frequently modest, and hits generally need further optimization.

 

What’s Next: The Data Advantage

We are past the point of debating whether AI/ML belongs in antibody discovery. It does and most serious programs already treat it as a core tool rather than an experiment. With the most accessible gains largely captured, the next leap will require more effort. 
As adopting a protein language model or plugging in a structure prediction tool becomes a procurement decision, the focus shifts to who has built the better data foundation. A model trained on your own consistent assay conditions, your own failure modes, your own progression decisions across targets, is a fundamentally different tool than one trained on public benchmarks. That advantage is not available off the shelf. It accumulates slowly, program by program, and it is almost invisible until suddenly it isn’t.

The next breakthrough in that sense, will be an organizational one: the decision to treat every experiment as a data asset rather than a result.