Blog

Why High-Throughput Screening Hits Don't Reproduce

Feb 9, 2026

The primary screen identified 847 hits. Dose-response narrowed it to 127 confirmed actives. You cherry-picked the best 20 for follow-up. Six months later, you've tested them all in orthogonal assays. Exactly two show real activity against your target. The other 18 were artifacts, aggregators, or assay interference. A 90% false positive rate—and you're not alone.


This is the reproducibility crisis in high-throughput screening, and it's one of the most expensive problems in early drug discovery.

Key Takeaways

  • Most HTS hits are false positives: Industry estimates suggest 80-95% of primary hits fail to reproduce in orthogonal assays

  • Compound aggregation is the dominant artifact: Colloidal aggregates nonspecifically inhibit proteins and confound most biochemical screens

  • PAINS filters catch only a fraction: Computational filters miss most problematic compounds and flag many legitimate ones

  • Orthogonal validation is essential: No single assay type can distinguish true binders from artifacts

  • Protein quality matters: Unstable or impure target protein increases false positive rates dramatically

The False Positive Problem

The Numbers Are Sobering

A typical HTS campaign:

  • Screen 500,000 compounds

  • Primary hits: 0.1-1% (~500-5000 compounds)

  • Confirmed in dose-response: ~10-25% of primary hits

  • Validated in orthogonal assay: ~10-20% of confirmed hits

  • True target engagement: Often <5% of confirmed hits


The math: From 5000 primary hits, you might get 50-100 true actives. The rest consumed months of chemistry resources pursuing artifacts.

Why This Matters

Cost: Hit-to-lead programs cost $1-5 million. Most are chasing compounds that never had a chance.


Time: False positives delay real leads by months or years.


Missed opportunities: The best compounds might have been deprioritized because an aggregator looked more potent.


Career impact: Projects fail not because the target was undruggable, but because the hits were never real.

The Five Sources of False Positives

Source 1: Colloidal Aggregation

The mechanism: Many organic molecules—including drugs, clinical candidates, and especially early leads—spontaneously form colloids in aqueous buffer. These aggregates are densely packed particles ranging from 50 to over 800 nm in radius that sequester 10⁵ to 10⁶ protein molecules, leading to their partial denaturation and typically their inhibition.


The prevalence: Studies suggest that aggregation has been reported to occur for up to 95% of primary hits from HTS campaigns. This is the dominant source of false positives in biochemical screens.


The insidious nature: Aggregation-based inhibition looks real:

  • Dose-dependent (more compound = more aggregates = more inhibition)

  • Saturates at high concentration

  • Shows apparent IC50 values

  • Can be potent (sub-micromolar apparent IC50)


What doesn't distinguish it from true inhibitors:

  • Apparent potency

  • Hill slope (often ~1)

  • Time-dependence (if any)


What does distinguish it:

  • Detergent sensitivity: Add 0.01-0.1% Triton X-100 or Tween-80; aggregate-based inhibition disappears

  • Counter-enzyme susceptibility: Aggregators inhibit unrelated enzymes (e.g., AmpC β-lactamase as a counter-screen)

  • Dynamic light scattering (DLS): Particles visible above critical aggregation concentration


Even cell-based assays aren't immune: Recent research on COVID-19 drug repurposing showed that of 41 tested drugs, 17 behaved as classic aggregators, forming particles by DLS and inhibiting counter-screening enzymes. Colloidal aggregation contributed to false positives in cell-based antiviral screens.

Source 2: Pan-Assay Interference Compounds (PAINS)

What PAINS are: PAINS are chemical compounds that give false positive results in many different assay types due to:

  • Covalent reactivity with proteins

  • Redox cycling

  • Metal chelation

  • Fluorescence interference

  • Photoreactivity


Common PAINS classes:

  • Rhodanines

  • Quinones

  • Catechols

  • Isothiazolones

  • Hydroxyphenyl hydrazones

  • Curcuminoids

  • Enones


The scope: Studies estimate that a typical academic screening library contains roughly 5-12% PAINS. There are approximately 400 structural classes of PAINS.


The limitation of PAINS filters: Large-scale analysis found that "the same PAINS substructure was often found in consistently inactive and frequently active compounds, indicating that the structural context in which PAINS occur modulates their effects."


PAINS filters are problematic: they are "oversensitive and disproportionately flag compounds as interference compounds while failing to identify a majority of truly interfering compounds." In other words, they create both false positives and false negatives in compound triage.


Better approaches: Quantitative Structure-Interference Relationship (QSIR) models for specific interference mechanisms (thiol reactivity, redox activity, luciferase activity) show 58-78% external balanced accuracy—better than PAINS filters but still imperfect.

Source 3: Technology-Specific Artifacts

Different assay formats have different vulnerabilities:


Fluorescence-based assays:

  • Compound autofluorescence (false activation)

  • Fluorescence quenching (false inhibition)

  • Inner filter effects at high concentration

  • Compound fluorescence overlapping with probe


Luciferase reporter assays:

  • Luciferase inhibition (independent of pathway)

  • Luciferase stabilization

  • Compound effects on reporter expression

  • Redox-active compounds


AlphaScreen/HTRF:

  • Singlet oxygen scavengers (false inhibition)

  • Light absorbers

  • Compound aggregation

  • Metal chelation affecting donor/acceptor beads


SPR (Surface Plasmon Resonance):

  • Non-specific binding to chip surface

  • Aggregation on sensor

  • Refractive index changes

  • Compound precipitation in flow


Each technology has specific failure modes that generate false positives unique to that platform.

Source 4: Protein Quality Issues

The problem nobody wants to discuss: Poor protein quality dramatically increases false positive rates:

Protein Issue

How It Causes False Positives

Aggregated protein

Binds compounds nonspecifically

Partially unfolded

Exposes normally buried sites

Inactive fraction

Shifts apparent IC50, flat SAR

Degraded protein

Cleavage products may have altered binding

Wrong oligomeric state

Different binding properties

Missing cofactor

Dead enzyme, artifacts dominate


The uncomfortable reality: Production of recombinant proteins is often hampered by instability and propensity to aggregate. Protein samples of poor quality are associated with reduced reproducibility.


When your target protein is 30% aggregated, 20% inactive, and slowly degrading during the screen—your data is compromised before a single compound is tested.


What to do:

  • Characterize protein quality before screening (SEC-MALS, DLS, activity assay)

  • Monitor stability over screening duration

  • Use fresh protein batches

  • Include positive control compounds on every plate

  • Track Z' factor for quality control

Source 5: Assay Condition Artifacts

The reproducibility trap: The exact conditions of screening matter enormously:

  • DMSO concentration: Above 1-2%, protein starts denaturing

  • Buffer composition: pH, ionic strength, divalent cations all affect binding

  • Temperature: Room temperature assays drift

  • Incubation time: Equilibrium may not be reached

  • Protein concentration: Below Kd, binding is hard to detect

  • Order of addition: Compounds added to protein vs. protein added to compounds


The drift problem: HTS assay validation guidelines emphasize that "plates should not exhibit material edge or drift effects." But in practice, a 384-well plate takes time to fill, and conditions change:

  • First wells incubate longer than last wells

  • Reagents in reservoirs warm up

  • DMSO evaporates, changing final concentration

  • Enzyme activity decays during plate setup

The Validation Cascade

Why One Assay Is Never Enough

A compound that inhibits your target in a fluorescence assay might be:

  1. A true inhibitor

  2. An aggregator

  3. A fluorescence quencher

  4. A PAINS compound that reacts with your target (but also everything else)

  5. A compound that affects assay reagents


You can't distinguish these with a single assay format.

The Standard Validation Workflow

Primary HTS (~0.5-1% hit rate)
        
Dose-response confirmation (same assay)
         (~10-25% confirm)
Counter-screen (detergent, different enzyme)
         (~50% survive)
Orthogonal assay (different detection)
         (~30-50% validate)
Biophysical confirmation (SPR, ITC, TSA)
         (~50% confirm binding)
Secondary cellular assay
        
Lead series identification

Each step reduces compound numbers by 50-80%.

Orthogonal Assay Selection

The second assay should:

  • Use a different detection technology

  • Measure the same biochemical event

  • Be insensitive to the artifacts of the first assay

Primary Assay

Orthogonal Options

Fluorescence intensity

HTRF, luminescence, SPR

HTRF/TR-FRET

Fluorescence polarization, AlphaScreen, SPR

AlphaScreen

HTRF, luminescence, mass spec

Reporter gene

Direct enzyme assay, protein interaction

Cell viability

Mechanistic assay, target engagement

The Biophysical Imperative

For any hit going into chemistry:

  • Confirm direct binding: SPR, ITC, MST, or equivalent

  • Measure stoichiometry: Is it 1:1? Is it specific?

  • Characterize kinetics: On-rate, off-rate

  • Check concentration dependence: Does binding saturate appropriately?


If a compound inhibits your enzyme but doesn't bind it biophysically—it's not inhibiting your enzyme.

Counter-Screens That Work

The Detergent Test

Simple and effective for aggregators:

  1. Run dose-response in standard buffer

  2. Run dose-response with 0.01% Triton X-100 or Tween-80

  3. Compare IC50 values


Interpretation:

  • IC50 unchanged: Likely real

  • IC50 shifts >5-fold or activity lost: Likely aggregator


Research has shown that this counter-screen reliably detects aggregate-based inhibition because non-ionic detergent disrupts the colloidal aggregates.

The Counter-Enzyme Test

Use an unrelated enzyme as a control: If your compound inhibits both your target kinase AND AmpC β-lactamase—it's not selective. It's probably an aggregator or reactive compound.


AmpC β-lactamase is commonly used because it's extensively characterized for aggregate-based inhibition.

The BSA Displacement Test

For cell-based assays: Run assay with and without 0.1% BSA (bovine serum albumin). Compounds whose activity is much reduced by BSA addition may be aggregators or have promiscuous protein binding.

The DLS Test

Direct observation of aggregation: Measure particle size by dynamic light scattering:

  • Below critical aggregation concentration: No particles

  • Above CAC: Particles appear (50-800 nm)


DLS provides direct evidence of aggregation without inference from enzyme inhibition.

The Target Protein Factor

Protein Quality Requirements for HTS

Before screening, verify:

Check

Method

Acceptable

Purity

SDS-PAGE

>90%

Aggregation

SEC, DLS

<10% aggregate

Activity

Specific assay

>80% of expected

Stability

Activity over time

<10% loss during screen

Identity

Mass spec

Correct MW, intact

When Protein Quality Causes Problems

Symptoms of protein-derived false positives:

  • Flat SAR (many compounds have similar IC50)

  • IC50 values inconsistent with biochemical understanding

  • Hits don't confirm in orthogonal assays

  • Different protein batches give different hit lists

  • Z' factor is borderline or variable


Root causes:

  • Protein aggregation creates non-specific binding sites

  • Inactive protein dilutes the signal

  • Degraded protein has altered activity

  • Missing cofactors create dead enzyme background

The Fresh Protein Rule

For critical screens:

  • Prepare protein fresh or from validated frozen stocks

  • Quality control each batch before use

  • Don't use protein that's been sitting in the fridge for weeks

  • Include positive controls on every plate to monitor target activity

Case Studies in False Positives

Case 1: The Aggregator Epidemic

Screen: Kinase inhibitor HTS, 300,000 compounds Primary hits: 2,400 compounds (0.8%) Dose-response confirmed: 480 compounds (20%) After detergent counter-screen: 96 compounds (20% of confirmed) After SPR: 28 compounds confirmed binding (6% of original confirmed hits)


Lesson: 80% of confirmed hits were aggregators, eliminated by a simple detergent test.

Case 2: The Luciferase Artifact

Screen: Reporter gene assay for pathway activation Primary hits: 150 activators Counter-screen (constitutive luciferase): 120 also activated control reporter True pathway activators: 30 compounds (20% of hits)


Lesson: 80% of "activators" were stabilizing luciferase, not activating the pathway.

Case 3: The Fluorescent Compound

Screen: Fluorescence polarization for protein-protein interaction inhibitor Top hit: Apparent IC50 = 100 nM, beautiful dose-response Orthogonal SPR: No binding detected Investigation: Compound was fluorescent at assay wavelength, creating artifact


Lesson: The most potent hit was completely artifactual.

Case 4: The Protein Quality Problem

Screen: Enzyme inhibitor, fluorescence-based First campaign (old protein): 450 hits, 5% confirmed in orthogonal Second campaign (fresh protein): 180 hits, 35% confirmed in orthogonal


Lesson: Degraded protein had 2.5× more hits, but 7× fewer real ones.

Building a Robust Screening Cascade

Before the Screen

  1. Validate protein quality

    • SEC-MALS for MW and aggregation

    • Activity assay for specific activity

    • Stability test at screening conditions

  2. Develop counter-screens

    • Detergent sensitivity assay

    • Counter-enzyme assay

    • Orthogonal detection method

  3. Establish positive controls

    • Known inhibitor/binder

    • Tool compound if available

    • Titrate to establish expected dose-response

  4. Set quality criteria

    • Minimum Z' factor (typically >0.5)

    • Maximum plate-to-plate variability

    • Signal window requirements

During the Screen

  1. Include controls on every plate

    • Positive control compounds

    • DMSO-only controls

    • Reference inhibitors at fixed concentration

  2. Monitor quality metrics

    • Z' factor per plate

    • Signal-to-background ratio

    • Control compound IC50 consistency

  3. Flag systematic problems

    • Edge effects

    • Drift across plates

    • Batch-to-batch variation

After the Screen

  1. Confirm in dose-response

    • Same assay format

    • Full 10-point curves

    • Triplicate measurements

  2. Counter-screen immediately

    • Detergent test for aggregation

    • Counter-enzyme for promiscuity

    • Autofluorescence if relevant

  3. Orthogonal validation

    • Different detection technology

    • Same biochemical endpoint

    • Biophysical binding confirmation

  4. Prioritize judiciously

    • Structurally tractable scaffolds

    • Clean mechanism of action

    • Confirmed target engagement

The Bottom Line

False positives are not failures of the screen—they're expected outcomes that require systematic elimination. The question isn't whether you'll have false positives (you will); it's whether you'll identify them before investing months in medicinal chemistry.

Prevention Strategy

Implementation

Aggregation

Detergent counter-screen, DLS

PAINS

Structural filters (with caution), QSIR models

Technology artifacts

Orthogonal assay validation

Protein quality

QC before screening, batch-to-batch controls

Assay artifacts

Robust validation, positive controls


The key insight: A hit is not a hit until it's validated by orthogonal methods. Primary screening identifies candidates for investigation—not compounds ready for optimization.

Target Validation for Screening Campaigns

For researchers planning HTS campaigns, understanding your target protein is the first line of defense against false positives. Platforms like Orbion can help with:

  • Aggregation propensity prediction: Identify whether your target is likely to aggregate, affecting assay performance

  • Binding site analysis: Understand the druggable pockets you're screening against

  • Protein stability assessment: Predict regions that may cause instability during screening

  • PTM requirements: Identify modifications that might be essential for proper target behavior


Better target understanding leads to better assay design, which leads to fewer false positives—saving months of wasted effort chasing artifacts.

References

  1. Shoichet BK. (2006). Screening in a spirit haunted world. Drug Discovery Today, 11(13-14):607-615. PMC4646424

  2. Owen SC, et al. (2012). Colloidal aggregation causes inhibition of G protein-coupled receptors. Journal of Medicinal Chemistry, 55(16):7203-7211. PMC3613083

  3. Feng BY & Bhatt R. (2006). A detergent-based assay for the detection of promiscuous inhibitors. Nature Protocols, 1(2):550-553. PMC1544377

  4. Tran H, et al. (2023). Colloidal aggregation confounds cell-based Covid-19 antiviral screens. ACS Chemical Biology, 18(10):2332-2342. PMC10634915

  5. Baell JB & Holloway GA. (2010). New substructure filters for removal of pan assay interference compounds (PAINS) from screening libraries and for their exclusion in bioassays. Journal of Medicinal Chemistry, 53(7):2719-2740.

  6. Jasial S, et al. (2017). How frequently are pan-assay interference compounds active? Large-scale analysis of screening data reveals diverse activity profiles, low global hit frequency, and many consistently inactive compounds. Journal of Medicinal Chemistry, 60(9):3879-3886. Link

  7. Baell JB & Walters MA. (2014). Chemical con artists foil drug discovery. Nature, 513:481-483. Link

  8. Schorpp K, et al. (2020). High-throughput screening to predict chemical-assay interference. Scientific Reports, 10:3839. Link

  9. Iversen PW, et al. (2012). HTS assay validation. In: Assay Guidance Manual. Eli Lilly & Company and NIH NCATS. NCBI Bookshelf

  10. Raynal B, et al. (2022). Measuring protein aggregation and stability using high-throughput biophysical approaches. Frontiers in Molecular Biosciences, 9:890862. Link