Skip to main contentSkip to navigation
ThisIsHowItWorks.in

Where complex ideas unfold at human pace

Primary

  • Atrium
  • Map
  • Pieces
  • Series
  • Search

Secondary

  • Archive
  • Index
  • Library
  • Fragments

Meta

  • About
  • Principles
  • Lexicon
  • Questions
  • Resources

Connect

  • Instagram
  • Discord
  1. Home
  2. /The Hardening of Knowledge
  3. /46 · When AI Started Doing Science: Machines as Researchers
Map

When AI Started Doing Science: Machines as Researchers


December 2020. DeepMind announces AlphaFold 2 has solved the "protein folding problem."

Given an amino acid sequence, predict the 3D structure of the resulting protein.

This problem had stumped biologists for 50 years.

Proteins fold in milliseconds. But predicting HOW they fold—which atoms go where, what shape emerges—required months of lab work per protein. Sometimes years.

AlphaFold solved it in minutes.

With 90%+ accuracy.

The scientific community reaction: Shock.

"This changes everything," said structural biologists who'd spent careers on this problem.

Because AlphaFold didn't just speed up science. It replaced a fundamental scientific activity.

Determining protein structure used to require:

  • Growing crystals (weeks/months, often fails)
  • X-ray crystallography (expensive, specialized)
  • Expert interpretation (PhDs, years of training)

Now: Feed sequence to AI. Get structure. Move on.

This wasn't AI assisting scientists. This was AI doing science.

July 2023. Google's AlphaFold predicts structures for 200+ million proteins—essentially all known proteins.

Work that would've taken human scientists centuries completed in months.

The inflection point had arrived.

For 400 years, science was human minds studying nature. Experimenting, measuring, theorizing, testing.

Now machines are doing it.

AI discovers drugs. Designs experiments. Analyzes data. Generates hypotheses. Even writes papers.

What happens to science when the scientists are machines?

Let's examine how AI entered science, what it can do that humans can't, what it still can't do, whether machine-discovered knowledge counts as "understanding," and what science becomes when it's automated.


THE ENTRY POINTS: How AI Invaded Science

AI'S EXPANDING ROLE IN SCIENCE

PHASE 1: DATA ANALYSIS (1990s-2010s) ┌─────────────────────────────────────────┐ │ Machine learning for pattern recognition│ │ ↓ │ │ Examples: │ │ • Particle physics (find signals in │ │ detector noise) │ │ • Genomics (identify gene patterns) │ │ • Astronomy (classify galaxies) │ │ ↓ │ │ Role: Tool for humans │ └─────────────────────────────────────────┘

PHASE 2: PREDICTION (2010s) ┌─────────────────────────────────────────┐ │ AI predicts outcomes from data │ │ ↓ │ │ Examples: │ │ • Drug binding affinity │ │ • Material properties │ │ • Protein interactions │ │ ↓ │ │ Role: Hypothesis generator │ └─────────────────────────────────────────┘

PHASE 3: DISCOVERY (2020s) ┌─────────────────────────────────────────┐ │ AI discovers new knowledge │ │ ↓ │ │ Examples: │ │ • AlphaFold (protein structures) │ │ • AlphaTensor (matrix multiplication │ │ algorithms) │ │ • Drug discovery (new molecules) │ │ ↓ │ │ Role: Independent researcher │ └─────────────────────────────────────────┘

PHASE 4: AUTOMATION (Emerging) ┌─────────────────────────────────────────┐ │ AI runs entire experiments │ │ ↓ │ │ Examples: │ │ • Robot scientists (design, run, analyze│ │ experiments) │ │ • AI-designed chips │ │ • Automated labs │ │ ↓ │ │ Role: Replaces human scientists │ └─────────────────────────────────────────┘

The progression: Tool → Assistant → Collaborator → Replacement.


CASE STUDY 1: AlphaFold—Solving a Grand Challenge

ALPHAFOLD 2 (DeepMind, 2020)

THE PROBLEM: ┌─────────────────────────────────────────┐ │ Protein Folding Problem: │ │ ↓ │ │ Proteins = chains of amino acids │ │ ↓ │ │ Fold into 3D shapes │ │ ↓ │ │ Shape determines function │ │ ↓ │ │ Challenge: Predict shape from sequence │ └─────────────────────────────────────────┘

TRADITIONAL APPROACH: ┌─────────────────────────────────────────┐ │ 1. Crystallize protein (hard, often │ │ fails) │ │ ↓ │ │ 2. X-ray crystallography (expensive) │ │ ↓ │ │ 3. Solve structure (months, requires │ │ expertise) │ │ ↓ │ │ Cost: $100K-$1M per protein │ │ Time: Months to years │ │ Success rate: ~30-40% │ └─────────────────────────────────────────┘

ALPHAFOLD APPROACH: ┌─────────────────────────────────────────┐ │ 1. Train neural network on known │ │ structures (~170,000 proteins) │ │ ↓ │ │ 2. Learn patterns of folding │ │ ↓ │ │ 3. Predict structure from sequence │ │ ↓ │ │ Cost: Essentially free (after training) │ │ Time: Minutes │ │ Accuracy: ~90% │ └─────────────────────────────────────────┘

THE IMPACT: ┌─────────────────────────────────────────┐ │ 2022: AlphaFold predicts 200M+ protein │ │ structures │ │ ↓ │ │ Made freely available │ │ ↓ │ │ Entire fields transformed: │ │ • Drug discovery │ │ • Molecular biology │ │ • Biotechnology │ │ ↓ │ │ Work that would take 100+ years done in │ │ months │ └─────────────────────────────────────────┘

WHAT IT MEANS: ┌─────────────────────────────────────────┐ │ Structural biology no longer bottleneck │ │ ↓ │ │ Entire career paths obsolete │ │ ↓ │ │ New questions now askable │ │ ↓ │ │ Science accelerated by orders of │ │ magnitude │ └─────────────────────────────────────────┘

One AI system obsoleted decades of human effort.


CASE STUDY 2: Drug Discovery—AI-Designed Molecules

AI DRUG DISCOVERY

TRADITIONAL DRUG DISCOVERY: ┌─────────────────────────────────────────┐ │ 1. Identify target (disease protein) │ │ ↓ │ │ 2. Screen millions of molecules │ │ (5-10 years) │ │ ↓ │ │ 3. Test promising candidates │ │ (animal studies, clinical trials) │ │ ↓ │ │ 4. Most fail │ │ ↓ │ │ Timeline: 10-15 years │ │ Cost: $1-2 billion per approved drug │ │ Success rate: <10% │ └─────────────────────────────────────────┘

AI-ASSISTED DRUG DISCOVERY: ┌─────────────────────────────────────────┐ │ 1. AI predicts which molecules will bind│ │ to target │ │ ↓ │ │ 2. AI designs new molecules (never seen │ │ in nature) │ │ ↓ │ │ 3. AI predicts toxicity, side effects │ │ ↓ │ │ 4. Synthesize only most promising │ │ ↓ │ │ Timeline: 2-3 years (screening phase) │ │ Cost: Much lower │ └─────────────────────────────────────────┘

EXAMPLES: ┌─────────────────────────────────────────┐ │ INSILICO MEDICINE (2019-2021): │ │ • AI designs drug for fibrosis │ │ • 18 months from start to Phase 1 trial │ │ • Normal timeline: 5+ years │ │ ↓ │ │ EXSCIENTIA (2020): │ │ • AI-designed drug for OCD │ │ • 12 months to clinical candidate │ │ ↓ │ │ ATOMWISE (2020): │ │ • AI screens billions of molecules │ │ • Identifies COVID antivirals in weeks │ └─────────────────────────────────────────┘

THE REVOLUTION: ┌─────────────────────────────────────────┐ │ AI doesn't just speed up existing │ │ process │ │ ↓ │ │ It designs molecules humans wouldn't │ │ think of │ │ ↓ │ │ Explores chemical space humans can't │ │ (10^60 possible drug-like molecules) │ │ ↓ │ │ True discovery, not just assistance │ └─────────────────────────────────────────┘

AI invents drugs humans never would have imagined.


CASE STUDY 3: Mathematics—AI Discovers Proofs

AI IN MATHEMATICS

DEEPMIND'S ALPHATENSOR (2022) ┌─────────────────────────────────────────┐ │ Problem: Matrix multiplication │ │ ↓ │ │ Find most efficient algorithm │ │ ↓ │ │ AlphaTensor discovers new algorithms │ │ better than human-found (50+ years) │ │ ↓ │ │ Finds thousands of efficient algorithms │ │ for different matrix sizes │ └─────────────────────────────────────────┘

LEAN THEOREM PROVER + AI: ┌─────────────────────────────────────────┐ │ AI assists mathematicians in proving │ │ theorems │ │ ↓ │ │ Suggests proof strategies │ │ ↓ │ │ Checks logical validity automatically │ │ ↓ │ │ Speeds up proof verification │ └─────────────────────────────────────────┘

AI-GENERATED CONJECTURES: ┌─────────────────────────────────────────┐ │ Ramanujan Machine (Technion, 2021): │ │ • AI generates mathematical conjectures │ │ • Some novel, some known │ │ • Mathematicians verify/prove │ │ ↓ │ │ AI as hypothesis generator in pure math │ └─────────────────────────────────────────┘

THE CONTROVERSY: ┌─────────────────────────────────────────┐ │ If AI finds proof humans can't follow: │ │ ↓ │ │ Do we "understand" the theorem? │ │ ↓ │ │ Is computer-verified proof sufficient? │ │ ↓ │ │ What is mathematics if not human │ │ understanding? │ └─────────────────────────────────────────┘

AI can discover mathematical truths humans might never find.

But is it doing mathematics? Or just computation?


WHAT AI CAN DO THAT HUMANS CAN'T

AI'S SUPERHUMAN CAPABILITIES

CAPABILITY 1: PATTERN RECOGNITION IN HIGH DIMENSIONS ┌─────────────────────────────────────────┐ │ Humans: Can visualize 3 dimensions │ │ ↓ │ │ AI: Operates in thousands of dimensions │ │ ↓ │ │ Finds patterns invisible to humans │ │ ↓ │ │ Example: Gene interactions (20,000 │ │ genes = 20,000 dimensions) │ └─────────────────────────────────────────┘

CAPABILITY 2: EXHAUSTIVE SEARCH ┌─────────────────────────────────────────┐ │ Humans: Intuition guides search │ │ ↓ │ │ AI: Can search billions of possibilities│ │ ↓ │ │ Example: Drug molecules, material │ │ combinations │ │ ↓ │ │ Doesn't get tired, bored, biased │ └─────────────────────────────────────────┘

CAPABILITY 3: INTEGRATION OF VAST DATA ┌─────────────────────────────────────────┐ │ Humans: Read ~1000 papers max │ │ ↓ │ │ AI: Process millions of papers │ │ ↓ │ │ Synthesize connections across all │ │ scientific literature │ │ ↓ │ │ Example: Meta-analysis, cross-field │ │ discovery │ └─────────────────────────────────────────┘

CAPABILITY 4: SPEED ┌─────────────────────────────────────────┐ │ Protein structure prediction: │ │ • Human: Months │ │ • AI: Minutes │ │ ↓ │ │ Enables iterative design impossible for │ │ humans │ └─────────────────────────────────────────┘

CAPABILITY 5: HYPOTHESIS GENERATION ┌─────────────────────────────────────────┐ │ AI generates thousands of hypotheses │ │ ↓ │ │ Humans test most promising │ │ ↓ │ │ Expands possibility space │ └─────────────────────────────────────────┘

AI doesn't replace human insight—it complements it.

But increasingly, it surpasses it.


WHAT AI STILL CAN'T DO (Yet)

CURRENT AI LIMITATIONS

LIMITATION 1: CAUSAL UNDERSTANDING ┌─────────────────────────────────────────┐ │ AI finds correlations, not causation │ │ ↓ │ │ Can predict "X correlates with Y" │ │ ↓ │ │ Can't explain "X causes Y because..." │ │ ↓ │ │ Problem: Science needs causal mechanisms│ └─────────────────────────────────────────┘

LIMITATION 2: EXPERIMENTAL DESIGN ┌─────────────────────────────────────────┐ │ AI good at analyzing results │ │ ↓ │ │ Still weak at designing novel │ │ experiments │ │ ↓ │ │ Requires creativity, domain knowledge, │ │ "what would be interesting to test?" │ │ ↓ │ │ Improving, but not yet human-level │ └─────────────────────────────────────────┘

LIMITATION 3: ASKING NEW QUESTIONS ┌─────────────────────────────────────────┐ │ AI optimizes within defined problem │ │ ↓ │ │ Humans: "What if we asked a completely │ │ different question?" │ │ ↓ │ │ Paradigm shifts still require human │ │ insight │ └─────────────────────────────────────────┘

LIMITATION 4: INTERPRETABILITYThe extent to which a system can explain its own outputs. Interpretability is about tracing causes, not just achieving performance. ┌─────────────────────────────────────────┐ │ Neural networks = "black boxes" │ │ ↓ │ │ Can predict, but can't explain HOW │ │ ↓ │ │ Science wants understanding, not just │ │ predictions │ │ ↓ │ │ Problem: "It works but we don't know │ │ why" │ └─────────────────────────────────────────┘

LIMITATION 5: COMMON SENSE ┌─────────────────────────────────────────┐ │ AI lacks intuition about physical world │ │ ↓ │ │ Makes "stupid" mistakes obvious to │ │ humans │ │ ↓ │ │ Example: Proposes chemically impossible │ │ molecules, physically absurd experiments│ └─────────────────────────────────────────┘

AI powerful, but not (yet) independent.

Still needs human judgment.


THE AUTOMATION FRONTIER: Robot Scientists

FULLY AUTOMATED RESEARCH SYSTEMS

THE VISION: ┌─────────────────────────────────────────┐ │ AI + Robotics = Autonomous scientist │ │ ↓ │ │ System: │ │ 1. Generates hypotheses (AI) │ │ 2. Designs experiments (AI) │ │ 3. Runs experiments (robots) │ │ 4. Analyzes results (AI) │ │ 5. Iterates (loop back to step 1) │ │ ↓ │ │ No human in loop │ └─────────────────────────────────────────┘

EXAMPLES ALREADY EXIST: ┌─────────────────────────────────────────┐ │ ADAM (2009, Aberystwyth University): │ │ • Robot scientist in biology │ │ • Discovered gene functions in yeast │ │ • Autonomously designed experiments │ │ ↓ │ │ EVE (2020, Manchester): │ │ • Successor to ADAM │ │ • Drug discovery against malaria │ │ ↓ │ │ CLOUD LABS (Emerald Cloud Lab): │ │ • Fully robotic labs │ │ • Scientists submit experiments remotely│ │ • Robots execute │ └─────────────────────────────────────────┘

MATERIALS SCIENCE AUTOMATION: ┌─────────────────────────────────────────┐ │ AI designs material compositions │ │ ↓ │ │ Robots synthesize samples │ │ ↓ │ │ Automated testing │ │ ↓ │ │ AI analyzes, suggests next composition │ │ ↓ │ │ Loop runs 24/7, no human needed │ │ ↓ │ │ Speed: 100x-1000x faster than human labs│ └─────────────────────────────────────────┘

THE IMPLICATIONS: ┌─────────────────────────────────────────┐ │ Science becomes industrial process │ │ ↓ │ │ Humans: Set goals, interpret findings │ │ ↓ │ │ Machines: Do research │ │ ↓ │ │ Bottleneck: No longer human effort, but │ │ computational/experimental resources │ └─────────────────────────────────────────┘

Science without scientists is already here.

In narrow domains.


THE EPISTEMOLOGICAL PROBLEM: Do We Understand?

UNDERSTANDING vs. PREDICTION

THE TRADITIONAL VIEW: ┌─────────────────────────────────────────┐ │ Science = Understanding nature │ │ ↓ │ │ Not just predicting, but explaining │ │ ↓ │ │ "Why does X happen? What mechanism │ │ produces X?" │ └─────────────────────────────────────────┘

AI CHALLENGE: ┌─────────────────────────────────────────┐ │ Neural networks predict accurately │ │ ↓ │ │ But: Can't explain HOW │ │ ↓ │ │ Internal workings = incomprehensible │ │ ↓ │ │ Question: Is prediction without │ │ explanation still science? │ └─────────────────────────────────────────┘

THE DEBATE: ┌─────────────────────────────────────────┐ │ POSITION 1: PREDICTION IS ENOUGH │ │ • Science's goal: Predict phenomena │ │ • Understanding = nice, but optional │ │ • If AI predicts perfectly, job done │ │ ↓ │ │ POSITION 2: UNDERSTANDING IS ESSENTIAL │ │ • Science seeks explanations, not just │ │ predictions │ │ • Must know WHY, not just WHAT │ │ • Black-box AI isn't real science │ └─────────────────────────────────────────┘

HISTORICAL PARALLEL: ┌─────────────────────────────────────────┐ │ Babylonian astronomy: │ │ • Predicted eclipses accurately │ │ • No understanding of mechanisms │ │ ↓ │ │ Was that science? │ │ ↓ │ │ Most say: No, pattern recognition only │ │ ↓ │ │ Is AI science same? Pattern recognition │ │ at massive scale? │ └─────────────────────────────────────────┘

THE MIDDLE GROUND: ┌─────────────────────────────────────────┐ │ AI discovers patterns │ │ ↓ │ │ Humans interpret, build theory │ │ ↓ │ │ Division of labor: │ │ • AI = Discovery engine │ │ • Humans = Meaning-makers │ └─────────────────────────────────────────┘

Can we have science we don't understand?

That's the question AI forces.


THE SOCIAL PROBLEM: What Happens to Scientists?

IMPACT ON SCIENTIFIC CAREERS

SCENARIO 1: DISPLACEMENT ┌─────────────────────────────────────────┐ │ AI automates: │ │ • Data analysis │ │ • Literature review │ │ • Hypothesis generation │ │ • Experiment design │ │ • Paper writing (already happening) │ │ ↓ │ │ What's left for humans? │ │ ↓ │ │ Fewer science jobs needed │ └─────────────────────────────────────────┘

SCENARIO 2: AUGMENTATION ┌─────────────────────────────────────────┐ │ AI = Tool, not replacement │ │ ↓ │ │ Scientists 10x-100x more productive │ │ ↓ │ │ Focus on: │ │ • Big questions │ │ • Interpretation │ │ • Theory building │ │ • Creativity │ │ ↓ │ │ Different skills, but still need humans │ └─────────────────────────────────────────┘

SCENARIO 3: BIFURCATION ┌─────────────────────────────────────────┐ │ Science splits: │ │ ↓ │ │ TIER 1: Elite scientists using AI │ │ (Small number, highly productive) │ │ ↓ │ │ TIER 2: Everyone else │ │ (Displaced or relegated to routine work)│ │ ↓ │ │ Inequality in scientific labor market │ └─────────────────────────────────────────┘

THE TRAINING PROBLEM: ┌─────────────────────────────────────────┐ │ If AI does science, how do we train │ │ scientists? │ │ ↓ │ │ Traditional: Learn by doing experiments,│ │ analyzing data │ │ ↓ │ │ But if AI does that, what do students │ │ learn? │ │ ↓ │ │ Risk: Generation that doesn't understand│ │ foundations │ └─────────────────────────────────────────┘

The scientific career path is changing.

Unclear what replaces it.


THE ETHICAL QUESTIONS: Who Controls AI Science?

GOVERNANCE AND ACCESS

QUESTION 1: PROPRIETARY AI ┌─────────────────────────────────────────┐ │ AlphaFold: Open (DeepMind released) │ │ ↓ │ │ But many AI tools: Proprietary │ │ ↓ │ │ Only companies with resources can build │ │ powerful AI │ │ ↓ │ │ Science becomes corporate? │ └─────────────────────────────────────────┘

QUESTION 2: BIAS IN AI ┌─────────────────────────────────────────┐ │ AI trained on existing data │ │ ↓ │ │ Data contains historical biases │ │ ↓ │ │ AI perpetuates biases │ │ ↓ │ │ Example: Medical AI trained on mostly │ │ white male data → doesn't work well for │ │ women, minorities │ └─────────────────────────────────────────┘

QUESTION 3: ACCOUNTABILITY ┌─────────────────────────────────────────┐ │ If AI designs drug that harms patients: │ │ ↓ │ │ Who's responsible? │ │ • AI creator? │ │ • Scientist who used it? │ │ • Hospital/company? │ │ ↓ │ │ Legal/ethical framework unclear │ └─────────────────────────────────────────┘

QUESTION 4: DUAL USE ┌─────────────────────────────────────────┐ │ AI can design beneficial drugs │ │ ↓ │ │ Same AI can design bioweapons │ │ ↓ │ │ How to enable good uses, prevent bad? │ │ ↓ │ │ Governance challenge │ └─────────────────────────────────────────┘

AI science raises questions science isn't equipped to answer.


CONCLUSION: Science Beyond Human Scale

For 400 years, science was human minds interrogating nature.

Galileo with his telescope. Darwin with his barnacles. Curie with her radioactive samples.

Slow. Limited. Human-scale.

By 2020, that was ending.

AlphaFold predicts 200 million protein structures.

AI designs drugs never imagined.

Robot scientists run experiments 24/7.

Science is becoming superhuman.

Not in flashy sci-fi ways—but in capability.

AI can:

  • Search billions of possibilities
  • Find patterns in thousands of dimensions
  • Integrate all scientific literature
  • Generate hypotheses at scale
  • Operate continuously without fatigue

What humans can't do, machines now can.

But we face profound questions:

Epistemological: Is prediction without understanding still science? Or just sophisticated pattern-matching?

Social: What happens to scientists when AI does their work? How do we train the next generation?

Ethical: Who controls these tools? How do we prevent misuse? Who's accountable when AI science fails?

The hardening of science required systematic, rigorous investigation by human minds.

Now the investigation is being handed to machines.

And we don't know if machine-discovered knowledge counts as understanding.

We have science we can't fully comprehend.

Produced by systems we can't fully interpret.

At speeds we can't keep up with.

This is the future of science.

Already here.

The question isn't whether AI will do science.

It already does.

The question is: What is science when scientists are machines?


[Cross-references: For protein folding and AlphaFold details, see Biology Companion #115. For drug discovery revolution, see Chemistry Companion #135-137. For mathematical proof automation, see Mathematics Companion #145-147. For reproducibility and how AI might help, see "The Reproducibility Crisis: When Science Couldn't Replicate Itself" (Core #41). For automation replacing expertise, see "When Expertise Lost Authority: Populism vs. Science" (Core #45). For what comes after this transformation, see "What Comes After Falsification? New Epistemologies" (Core #48) and "The Future of Hardening" (Core #50).]

PreviousWhen Expertise Lost Authority: Populism vs. ScienceNextWhen Science Faced Limits (Again): The New Impossibilities

The Suitcase

Take this piece with you—works offline, no internet needed.

↩ Return to The Hardening of Knowledge⌂ Ascend to The Observatory