Randomized controlled trials (RCTs) are supposed to be the gold standard of medical evidence. But acne research reveals a painful truth: running a rigorous acne RCT is extraordinarily difficult, and many published trials fall short of even basic scientific standards. A 2009 analysis of 25 acne RCTs found at least 12 potentially serious problems across the studies, including statistical errors, blinding failures, and outcome reporting bias. When researchers actually examined 274 controlled acne trials in total, 21% had at least one major weakness and no offsetting strengths—meaning more than one in five studies couldn’t provide reliable conclusions about whether a treatment actually works.
The problem isn’t one thing. It’s a compounding storm of obstacles: researchers can’t agree on how to measure acne severity (over 25 different methods exist), patients don’t adhere to topical treatments reliably, treatments look or feel different enough that patients figure out which group they’re in (breaking the “blind”), and the simple act of controlling a trial—like preventing patients from using other acne products—conflicts with the need to recruit diverse participants. This creates a situation where even well-intentioned researchers struggle to produce trustworthy evidence about which acne treatments actually work. This article walks through why acne RCTs are so difficult: the reporting and design problems that plague them, the outcome measurement chaos that makes comparing studies nearly impossible, the practical barriers that derail even careful studies, and the evidence gaps that remain as a result.
Table of Contents
- How Statistical and Design Problems Undermine Acne Trial Credibility
- The Blinding Problem—Patients Figuring Out Their Treatment Assignment
- The Acne Grading Chaos—25 Different Severity Scales
- Patient Adherence and Complex Formulations—Real-World Breakdown
- The Conflicting Requirements of Good Study Design
- The Evidence Gap That Remains
- What This Means for Future Acne Research
- Conclusion
- Frequently Asked Questions
How Statistical and Design Problems Undermine Acne Trial Credibility
Many acne trials never properly consider whether they have enough statistical power to detect a real treatment benefit. Inadequate statistical power means a trial might conclude “this treatment doesn’t work” even when it actually does—the study was just too small or poorly designed to catch the effect. Guidelines from the Journal of the American Academy of Dermatology note that many acne trials are insufficiently powered to exclude potentially useful treatment benefits, yet still publish conclusions claiming no effect exists. This is a major credibility problem: if your study can’t reliably detect a real difference, you shouldn’t claim a treatment is ineffective.
Non-inferiority studies present another design trap specific to acne research. These studies test whether a new treatment is “not worse” than an existing one, rather than better. But researchers often design these trials without pre-specifying a non-inferiority margin—the threshold that defines what counts as “not worse.” Without this upfront definition, researchers can interpret results flexibly after the data arrives, a practice that inflates the apparent success of mediocre treatments. The SD-ACNE trial documented these specific margin-setting failures in contemporary acne research, showing that the problem persists despite being well-known in dermatology and biostatistics.

The Blinding Problem—Patients Figuring Out Their Treatment Assignment
Double-blind trials are supposed to prevent both patients and researchers from knowing who gets which treatment, eliminating bias. In acne research, however, this ideal frequently collapses. Trials reported as “double-blind” often compare treatments that look or feel completely different, or that produce different side effects. A topical antibiotic might cause different skin texture or odor than a retinoid, or one formulation might be an obvious cream while another is a gel. When patients can tell which treatment they’re receiving based on appearance, texture, or side effects, they’re no longer truly blinded. Their expectations shift, their adherence behavior changes, and their placebo response is no longer the same across groups—exactly the bias that blinding is meant to prevent.
However, not all trials can easily solve this problem. True matching placebos that look and feel identical while delivering entirely different active ingredients require specialized pharmaceutical manufacturing. Smaller academic studies or trials of natural compounds often lack the resources for this. This creates a catch-22: you can produce a properly blinded trial at high cost and complexity, or you can run a simpler trial with obvious blinding failures. Many researchers choose the latter, which is why BMC Trials’ analysis found blinding failures scattered across acne RCTs even when researchers claimed double-blind design. The solution isn’t always available within realistic research budgets.
The Acne Grading Chaos—25 Different Severity Scales
Before any acne trial can begin, researchers must decide how to measure whether acne improves. This should be straightforward. Instead, it’s a nightmare of fragmentation. Over 25 different methods exist for assessing acne severity, and more than 19 different methods exist just for counting lesions. No standardized ruler exists. Some systems count inflamed lesions only; others count noninflammatory ones.
Some assess severity by area of face involved; others by absolute lesion count. Some grade on a 0-4 scale; others use 0-10. The consequence is that published acne trials are essentially incomparable. A trial showing a 50% lesion reduction using one grading system cannot be directly compared to a trial showing a 40% reduction using a different system. Moreover, current grading systems are subjective and influenced by clinician experience, visual acuity, lighting conditions, and patient skin type. According to the 2023 Annals of Dermatology review on acne grading scales, most don’t assess post-inflammatory hyperpigmentation, scarring, or patient quality-of-life impacts—outcomes that patients actually care about. The FDA’s own guidance on acne drug effectiveness acknowledges this mess but offers no simple fix, because fixing it would require dermatology to converge on standardized methods that don’t currently exist.

Patient Adherence and Complex Formulations—Real-World Breakdown
Acne treatments are topical creams, gels, and cleansers that patients apply twice daily at home. Unlike pills that a clinic can directly monitor, adherence is invisible and often poor. Patients skip doses, forget to apply treatments, or get discouraged when results take weeks to appear. Poor adherence to topical acne treatments is a significant issue that complicates trial design and leads to poor quality-of-life outcomes, according to dermatology reviews in PubMed Central. In a trial setting, some patients will skip applications or use treatments inconsistently, yet still be counted as part of the treatment group—which dilutes the observed treatment effect and makes the drug look less effective than it actually is.
Many clinical trials also use formulations with multiple ingredients: a retinoid plus niacinamide plus zinc, for example. This creates a testing nightmare. You can’t tell whether improvements come from the retinoid, the niacinamide, the zinc, the formulation’s pH, the emollient, or some synergistic combination. Assessment of individual component efficacy and mechanism of action becomes impossible. A trial showing that a five-ingredient formulation works tells you almost nothing about which ingredients matter or how they work. This is particularly limiting for researchers trying to understand acne pathophysiology or develop better treatments.
The Conflicting Requirements of Good Study Design
Here’s the central paradox: RCTs need broad eligibility criteria to be “externally valid”—meaning the results apply to real patients with diverse skin, genetics, and acne presentations. But RCTs also need tight control to isolate the effect of the treatment being tested. This often means prohibiting patients from using concomitant topical medications during the trial. A patient can’t use additional moisturizers, other acne treatments, or sunscreen formulations not approved by the study.
In real life, no dermatologist tells a patient with severe acne to stop using moisturizer to reduce inflammation. But in a clinical trial, allowing patients to use additional products muddies the results—you no longer know if improvement came from the test drug or the moisturizer. The SD-ACNE trial explicitly documented these conflicting study design requirements, showing that the necessity of restricting concomitant topical medications to maintain scientific control directly conflicts with the need for broad inclusion criteria to ensure external validity. Researchers can include a diverse population, or they can have tight control, but often not both. Most acne trials try to split the difference and end up with reduced diversity and still-questionable control.

The Evidence Gap That Remains
After reviewing the literature, the 2023 Journal of the American Academy of Dermatology guidelines identified important evidence gaps that remain even after decades of acne research. There is insufficient evidence regarding microbiology and endocrinology testing approaches, systemic antibiotics beyond tetracyclines, physical modalities like light therapy, complementary therapies, dietary interventions, and cost-effectiveness comparisons of treatments.
Additionally, most acne research focuses on mild-to-moderate acne, creating a significant gap in clinical evidence for severe acne and for cosmeceutical efficacy—meaning we have far less data on severe inflammatory acne or natural skincare products than we do on standard treatments. This skew toward mild-to-moderate disease reflects practical trial design: it’s easier to recruit and retain patients with mild acne, which typically shows improvement faster. But it leaves severe acne patients with less evidence-based guidance, and it leaves open important questions about which populations benefit most from different approaches.
What This Means for Future Acne Research
The problems in acne RCTs aren’t unsolvable, but solving them requires resources, coordination, and a willingness to move slowly. Some solutions exist: central image grading by blinded experts rather than on-site clinicians can reduce subjective bias. Standardized outcome measures adopted across trials can make results comparable. Pre-registration of trial protocols and non-inferiority margins can prevent selective reporting. High-quality placebo formulations can improve blinding. But each of these solutions adds cost and complexity, which many academic research teams cannot afford.
The acne research community is gradually improving. Recent large trials have adopted more rigorous reporting standards and clearer outcome definitions. But the 21% of trials with major weaknesses and no strengths shows that the shift is incomplete. For patients, the practical implication is clear: don’t treat published acne research as settled truth. When a study claims a treatment works, ask whether the trial was properly blinded, whether the outcome measures are transparent, and whether the sample size was adequate. The published evidence is often better than no evidence, but it’s shakier than it appears.
Conclusion
Randomized controlled trials for acne are difficult because the disease itself is difficult to measure consistently, the treatments are easy for patients to abandon or misuse, the studies often can’t maintain proper blinding, and researchers face irreconcilable tradeoffs between scientific rigor and real-world applicability. A substantial proportion of published acne trials contain serious methodological problems that call their conclusions into question. This isn’t a failure of the researchers—it’s an honest reflection of how hard it is to run clean, comparable clinical trials in dermatology.
Understanding these challenges matters because it helps patients and clinicians interpret acne research critically rather than assuming all published trials are equally credible. It also explains why dermatologists often rely on clinical judgment and patient response rather than strict adherence to trial results: they understand the evidence base has real gaps and limitations. As acne research continues, the field is slowly adopting more standardized measures and rigorous reporting, but large-scale improvement will require funding and coordination beyond what individual academic teams can provide.
Frequently Asked Questions
If acne trials are so flawed, should I ignore published research about acne treatments?
No—published research is still useful, but you should read it critically. Look for trials that clearly describe their blinding method, pre-specify their primary outcome, report how many patients dropped out, and use established acne grading systems. Larger, more recent trials in established dermatology journals tend to be more rigorous than smaller, older studies.
Why don’t acne trials just use photos and computer analysis to avoid subjective grading?
Some newer trials do use image analysis, but it introduces different problems: lighting, camera angle, and software bias can skew results in their own ways. Perfect objectivity doesn’t exist. Central grading by blinded human experts is often more reliable than fully automated analysis, but it’s also more expensive.
Is there an acne grading system that most dermatologists use?
Unfortunately no. The Global Acne Grading System is common, but it’s not universal. Many trials develop their own severity definitions, which is why comparing results across trials remains difficult. Standardization has been recommended but not widely adopted.
Why does adherence matter so much in acne trials?
Because poor adherence dilutes the observed treatment effect. If half your study population barely uses the treatment, the overall group doesn’t show the benefit that careful, consistent users would see. This makes drugs look less effective than they actually are in committed patients.
Can dermatologists tell if a treatment works better than just observing patients?
Often, yes—observing many patients over time gives clinicians real-world experience that balances the limitations of any individual trial. But individual trial results remain the closest we get to isolating a drug’s true effect, even when flawed.
Will acne research ever have truly standardized outcome measures?
Possibly, but it would require buy-in from dermatology organizations, funders, and pharmaceutical companies—a level of coordination that hasn’t happened yet. Some efforts are underway, but change is slow.
You Might Also Like
- Why Ziana Gel Combines Tretinoin and Clindamycin for Acne
- Why Vegans Still Get Acne — What They’re Missing
- Why Tretinoin Compounded with Clindamycin Is Prescribed for Acne
Browse more: Acne | Acne Scars | Adults | Back | Blackheads



