Fiveable

☣️Toxicology Unit 1 Review

QR code for Toxicology practice questions

1.3 Toxicity testing methods

1.3 Toxicity testing methods

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
☣️Toxicology
Unit & Topic Study Guides

In Vitro Toxicity Testing

In vitro testing uses isolated cells, tissues, or organs outside the body to evaluate how substances cause harm. These methods serve as a first-pass screen: they help identify potential toxicants and prioritize which compounds need further, more expensive testing. They also reveal mechanisms of toxicity at the cellular level, which whole-animal studies often can't pinpoint as precisely.

Cell Culture Assays

Cell culture assays grow isolated cells in controlled lab conditions and then expose them to test substances. You can measure a range of endpoints: cell viability (did the cells survive?), proliferation (did they keep dividing?), and differentiation (did they mature normally?).

  • Commonly used cell lines include human hepatocytes (liver cells, useful for studying drug metabolism) and human keratinocytes (skin cells, useful for irritation and sensitization testing)
  • These assays can be automated and miniaturized into multi-well plates, making them suitable for high-throughput screening of thousands of compounds at once

Organ-on-a-Chip Models

Traditional cell cultures grow cells in flat layers, which doesn't reflect how cells behave inside the body. Organ-on-a-chip models address this by using microfluidic devices that mimic the structure, flow, and mechanical forces of real organs.

  • Fluid channels simulate blood flow, and different cell types can be arranged in tissue-like architectures
  • Examples include liver-on-a-chip (for hepatotoxicity) and kidney-on-a-chip (for nephrotoxicity)
  • These models provide a more physiologically relevant environment than standard cell culture, though they're still simplified compared to a whole organ

High-Throughput Screening

High-throughput screening (HTS) uses robotic systems to test thousands or even millions of compounds rapidly against specific biological targets or cell-based assays.

  • HTS platforms can run cell-based assays and biochemical assays simultaneously across large compound libraries
  • A major example is the EPA's ToxCast program, which has screened over 9,000 chemicals across hundreds of assay endpoints to build toxicity profiles without relying on animal data
  • Results from HTS feed into prioritization decisions: compounds flagged as potentially toxic move on to more detailed testing

Advantages vs. Limitations

Advantages:

  • Reduced animal use
  • Ability to test large numbers of compounds quickly and cost-effectively
  • Can reveal specific mechanisms of toxicity at the molecular and cellular level

Limitations:

  • Lower physiological relevance than whole-animal studies (isolated cells don't replicate organ-level or systemic interactions)
  • Cannot assess effects that depend on metabolism, immune response, or communication between organ systems
  • Risk of false positives (flagging safe compounds as toxic) and false negatives (missing truly toxic compounds)

In Vivo Toxicity Testing

In vivo testing uses whole, living animals to assess how a substance behaves in a complete biological system. This captures things in vitro methods can't: absorption, distribution, metabolism, excretion, and interactions between organ systems. Regulatory agencies generally require in vivo data before approving new drugs or industrial chemicals.

Animal Models

  • Rodents (mice and rats) are the most commonly used species due to their small size, short reproductive cycles, and well-characterized genetics
  • Rabbits are often used for dermal and ocular irritation studies
  • Non-human primates are reserved for cases where no other model adequately predicts human response, such as certain immunological or reproductive studies
  • Transgenic animal models with specific genes knocked out or inserted allow researchers to study how individual genes contribute to toxic responses

The choice of species depends on the toxicological endpoint and how closely the animal's physiology matches human biology for the relevant organ system.

Dose Selection

Dose selection determines the range of exposures animals receive during a study. Getting this right is critical because too narrow a range can miss important effects.

  • The range should span from the NOAEL (No Observed Adverse Effect Level, the highest dose causing no detectable harm) up to doses that produce clear toxicity
  • Dose levels are typically informed by preliminary range-finding studies, expected human exposure levels, and the substance's toxicokinetics (how the body absorbs, distributes, metabolizes, and excretes it)
  • Most studies use at least three dose groups plus a control group

Route of Administration

The route of administration must match how humans would actually be exposed to the substance, because the same chemical can produce very different effects depending on how it enters the body.

  • Oral (gavage): substance delivered directly to the stomach via a tube; used for drugs and food additives
  • Dermal: applied to the skin; used for cosmetics, pesticides, and industrial chemicals
  • Inhalation: aerosolized or vaporized substance breathed in; used for air pollutants and occupational exposures
  • Injection (intravenous, intraperitoneal, subcutaneous): used when precise dosing or rapid systemic exposure is needed

Endpoints and Biomarkers

Endpoints are the specific effects you measure: mortality, body weight changes, organ weight changes, and histopathological changes (microscopic tissue damage).

Biomarkers are measurable indicators that a biological process has been altered. They can signal toxicity before overt damage appears:

  • Liver injury: elevated ALT and AST enzyme levels in blood
  • Kidney damage: elevated BUN (blood urea nitrogen) and creatinine
  • Gene expression changes or altered metabolite profiles can serve as early-warning biomarkers

Alternative Testing Strategies

Alternative testing strategies follow the 3Rs framework (Reduction, Refinement, Replacement) by aiming to minimize or eliminate animal use while still generating reliable toxicity data.

In Silico Modeling

In silico modeling uses computational tools to predict toxicity based on a substance's chemical structure and physicochemical properties, without any biological testing.

  • QSAR models (Quantitative Structure-Activity Relationship) analyze relationships between chemical features and known toxic effects across large databases of compounds, then predict toxicity for untested chemicals
  • PBPK models (Physiologically Based Pharmacokinetic) simulate how a substance moves through the body over time, predicting tissue concentrations and metabolic fate
  • These models are most useful for prioritizing compounds and flagging likely hazards early in the testing pipeline
Cell culture assays, Frontiers | Cell Culture Based in vitro Test Systems for Anticancer Drug Screening

Read-Across Approaches

Read-across uses existing toxicity data from structurally similar compounds to predict the toxicity of a substance that lacks data. The core assumption is that chemicals with similar structures tend to behave similarly in biological systems.

  • This approach is especially valuable for filling data gaps when testing every compound individually isn't feasible
  • The strength of a read-across prediction depends on how closely the source and target compounds match in structure and in relevant biological activity

Weight of Evidence

Weight of evidence (WoE) integrates data from multiple sources (in vitro, in vivo, in silico, epidemiological) into a single overall assessment of a substance's toxicity.

  • WoE considers the quality, reliability, and relevance of each data source rather than relying on any single study
  • A systematic framework guides how conflicting results are weighed and how conclusions are drawn
  • This approach reduces the uncertainty that comes with depending on individual studies alone

Integrated Testing Strategies

Integrated testing strategies (ITS) combine multiple methods in a tiered approach:

  1. Start with in silico predictions and existing data review
  2. Move to in vitro assays if the first tier raises concerns or leaves gaps
  3. Proceed to in vivo testing only if earlier tiers can't provide sufficient information

This tiered structure optimizes resources and minimizes animal use by ensuring that animal studies are only conducted when truly necessary.

Toxicogenomics

Toxicogenomics applies genomic technologies to toxicology, studying how toxicants alter gene expression across the entire genome. This field helps uncover why a substance is toxic at the molecular level, not just that it's toxic.

Gene Expression Profiling

Gene expression profiling measures mRNA transcript levels in cells or tissues after toxicant exposure, revealing which genes are turned up or down in response.

  • Microarray technology measures thousands of known gene transcripts simultaneously on a chip
  • RNA sequencing (RNA-seq) provides a more comprehensive and quantitative view, including detection of novel transcripts
  • The resulting expression patterns can point to specific molecular pathways driving the toxic response

Pathway Analysis

Raw gene lists aren't very informative on their own. Pathway analysis uses bioinformatics tools to map affected genes onto known biological pathways and networks.

  • This reveals which cellular processes (e.g., oxidative stress response, apoptosis, DNA repair) are disrupted by the toxicant
  • Commonly used tools include Ingenuity Pathway Analysis (IPA) and Gene Set Enrichment Analysis (GSEA)
  • Pathway-level understanding supports the development of Adverse Outcome Pathways (AOPs), which link molecular initiating events to organ-level or organism-level harm

Biomarker Discovery

Toxicogenomics can identify molecular biomarkers, specific gene expression signatures or metabolite changes, that indicate exposure to a particular toxicant or predict the likelihood of an adverse outcome.

  • These biomarkers can be used for monitoring exposed populations or for early detection of organ damage in clinical and occupational settings
  • Examples include characteristic gene expression profiles associated with hepatotoxicity or genotoxicity

Limitations and Challenges

  • Toxicogenomic studies generate massive, complex datasets that require specialized bioinformatics expertise to analyze and interpret
  • Translating findings from animal models or cell lines to human health predictions remains difficult
  • The field still lacks full standardization of methods and data analysis pipelines, which limits reproducibility and cross-study comparisons

Validation of Testing Methods

Before a new testing method can be used for regulatory decisions, it must be validated: formally demonstrated to be reliable and relevant for its intended purpose. This process is what allows alternative methods to eventually replace animal tests.

Regulatory Requirements

Agencies like the FDA (US Food and Drug Administration) and EMA (European Medicines Agency) set specific validation requirements:

  • The method's reliability, reproducibility, and relevance to the intended purpose must be demonstrated
  • Validation studies follow established guidelines (e.g., from the OECD or ICCVAM)
  • Results must be independently reviewed and approved before the method gains regulatory acceptance

Interlaboratory Reproducibility

A method is only useful if different labs can run it and get consistent results. Interlaboratory reproducibility testing addresses this:

  1. Multiple independent laboratories receive the same test substances and the same protocol
  2. Each lab performs the assay independently
  3. Results are compared to identify sources of variability
  4. The method is refined if variability is too high
Cell culture assays, Frontiers | Cell-Type-Specific High Throughput Toxicity Testing in Human Midbrain Organoids

Predictive Value

Predictive value measures how well a test method's results match actual human outcomes. This is assessed by comparing test results against data from human clinical trials or epidemiological studies.

  • High sensitivity means the method correctly identifies truly toxic substances (few false negatives)
  • High specificity means it correctly identifies non-toxic substances (few false positives)
  • A method with strong predictive value reduces the risk of unsafe products reaching humans and prevents unnecessary rejection of safe compounds

Acceptance Criteria

Acceptance criteria are the performance benchmarks a method must meet to be considered valid:

  • Accuracy: how close results are to the true value
  • Precision: how reproducible results are across repeated tests
  • Sensitivity and specificity: the method's ability to correctly classify toxic and non-toxic substances

These criteria are set based on the method's intended regulatory use and the level of certainty required for decision-making.

Emerging Technologies

Several new technologies are pushing toxicity testing toward greater accuracy, efficiency, and human relevance.

Stem Cell-Based Assays

Stem cell-based assays use human stem cells, which can differentiate into virtually any cell type, providing models that are more physiologically relevant than immortalized cell lines.

  • Embryonic stem cell tests assess developmental toxicity by measuring how test substances interfere with differentiation
  • Induced pluripotent stem cell (iPSC)-derived assays can generate patient-specific cell types (neurons, cardiomyocytes, hepatocytes), enabling personalized toxicity assessment
  • These assays are particularly valuable for studying developmental and organ-specific toxicity

Zebrafish Embryo Model

Zebrafish embryos have become a popular alternative model that bridges the gap between in vitro and traditional mammalian in vivo testing.

  • Embryos are transparent, allowing direct observation of organ development and toxicant effects in real time
  • They develop rapidly (most organs form within 48-72 hours) and are small enough for multi-well plate screening
  • They can be genetically manipulated and are useful for assessing developmental toxicity, neurotoxicity, and cardiotoxicity
  • Regulatory bodies in some jurisdictions classify early-stage zebrafish embryos as alternatives to animal testing

Organ-on-a-Chip Advancements

The next generation of organ-on-a-chip technology goes beyond single-organ models:

  • Multi-organ chips connect liver, kidney, heart, and lung compartments through microfluidic channels, simulating systemic interactions
  • Body-on-a-chip models aim to replicate whole-body pharmacokinetics on a single device
  • Integration of immune cells, blood vessel networks, and 3D bioprinted tissue structures is increasing physiological realism
  • These systems could eventually reduce or replace many types of systemic toxicity studies currently requiring animals

CRISPR-Cas9 Gene Editing

CRISPR-Cas9 allows precise, targeted modifications to DNA, making it a powerful tool for toxicological research.

  • Researchers can create cell lines or animal models with specific gene knockouts to determine whether a particular gene is involved in a toxic response
  • CRISPR screens can systematically test thousands of genes to identify molecular targets of toxicants
  • These models help clarify mechanisms of toxicity and can be used to develop more targeted screening assays

Ethical Considerations

Ethical principles shape how toxicity testing is designed, conducted, and regulated. The central tension is between the need for reliable safety data and the obligation to minimize harm to animals.

Animal Welfare

Researchers have a responsibility to minimize suffering throughout the testing process. This includes:

  • Appropriate housing, nutrition, and veterinary care
  • Use of anesthesia and analgesia during painful procedures
  • Humane endpoints: studies are designed so animals are euthanized before experiencing severe suffering, rather than waiting for death as an endpoint

Reduction, Refinement, Replacement (3Rs)

The 3Rs are the guiding ethical framework for animal use in research:

  • Reduction: minimize the number of animals used (e.g., through efficient statistical designs, data sharing between studies)
  • Refinement: minimize suffering in animals that are used (e.g., less invasive procedures, environmental enrichment)
  • Replacement: use non-animal methods whenever scientifically possible (e.g., in vitro assays, in silico models)

These principles are embedded in regulations governing animal research in most countries.

Alternatives to Animal Testing

Developing and validating alternatives is a major priority, but the process is slow. A new in vitro method may take a decade or more to move from development through formal validation to regulatory acceptance. The challenge is demonstrating that the alternative provides equivalent or better predictive value compared to the animal test it would replace.

Public Perception and Acceptance

Public attitudes toward animal testing influence both policy and research funding. Concerns about animal welfare have driven significant investment in alternative methods, particularly in the EU, where cosmetics testing on animals has been banned since 2013. Transparency in research practices and clear communication about why certain tests are still necessary help maintain public trust in the regulatory process.

2,589 studying →