QMENTA News

Multi-Site Neuroimaging Trials: The State of Operations in 2026

Written by Imaging Team | May 5, 2026 3:42:32 PM

Managing imaging across more than ten trial sites is one of those jobs where the problems you prepared for aren’t the ones that actually cost you. The scanner variability, the protocol drift, the site that’s been uploading in the wrong orientation for six months — none of it showed up in your risk register. It showed up in your data.

The neuroimaging trial space has a documentation problem that gets misread as a technology problem, which gets misread as a science problem. Sponsors spend real money on protocol development, site training, and regulatory strategy, and then lose ground to things like a technologist at a community hospital in eastern Europe who adjusted the field of view because it looked better to them. Not malicious. Not even wrong, from their vantage point. Just not what the protocol said.

Phase III trials run at roughly $55,716 per day in direct costs, according to a 2024 Tufts CSDD study pulling from 447 actual protocol budgets.2 That figure becomes relevant fast when you’re doing the arithmetic on a three-month site qualification delay. Finance sees the holding cost. What’s harder to get into a spreadsheet is the downstream hit to enrollment velocity, the interim analysis that slips, the extended monitoring. Those don’t show up as line items until the post-mortem.

The deviation problem is worse than most protocols assume

Baseline data on multi-center neuroimaging studies suggests protocol deviations touch something like half of all imaging time points.1 The number sounds alarming until you understand what’s being counted — it’s not catastrophic failures, mostly. Missing localizer scans. Sequence parameters that drifted slightly. Orientation errors. The kind of thing that, individually, a radiologist might wave through and which, aggregated across forty sites and eighteen months, quietly destroys your statistical power.

There’s a version of this that shows up in biostatistics review meetings when someone points out that a meaningful slice of the data is unusable. At that point you’re past the correctable stage. The variance is baked in.

Scanner homogeneity is a related issue that doesn’t get enough airtime in protocol planning. Two Siemens Verio 3T systems, same model year, same installation team, will not produce the same field homogeneity.3 The physics here is not controversial — the research on this goes back years — but trial sponsors continue to treat same-model scanners as functionally equivalent for protocol purposes. They aren’t, and phantom studies are the established mechanism for catching the drift before it affects patient data rather than after. This isn’t a new lesson. It keeps having to be learned.

What reviewing hundreds of scans a week actually produces

There’s a straightforward workload argument against manual QC at multi-site scale that gets underplayed. A radiologist clearing MRI sequences for protocol compliance across forty active sites, each sending weekly visits, is moving through a few hundred image series per week. The literature on inter-rater reliability at high volume is consistent: the same reviewer making the same call on Monday morning and Friday at 4 p.m. is not reliable. This isn’t a quality criticism; it’s just a description of how humans work under sustained cognitive load.

The documentation side of this is what regulators care about. Email chains and spreadsheet trackers don’t produce audit trails. They produce stories about what probably happened, reconstructed after the fact, which is not what FDA inspectors are looking for when they review pivotal trial imaging endpoints. They want time-stamped records. They want to see the chain of review, the flags, the corrective actions, the confirmation that the actions worked. Manual QC doesn’t generate that, not at the volume and traceability level that’s now expected.

Some actual numbers, with the caveats attached

QMENTA has published operational results from an MS trial showing a 98.1% initial QA pass rate, with 1.6% of visits requiring reacquisition, and MRI-related queries accounting for 5% of 750 total queries across the study.1 These are good numbers. They’re also QMENTA’s numbers about our own platform, which is worth acknowledging when you cite them. The direction they point is credible and consistent with what automated QC infrastructure should produce; the precision should be understood as vendor-reported.

The 5% query figure is the one worth dwelling on. In manually managed trials, MRI queries don’t stay at 5% of the queue — they tend to dominate it. Pulling them down to that level changes the operational reality for the monitoring team in ways that compound over a multi-year study. That’s not a marginal efficiency gain.

On AI reading speed: a 2024 paper in Neuroradiology measured reading times for MS brain MRIs in 35 consecutive patients, with and without AI assistance.4 Mean time without AI was 9.05 minutes; with AI it dropped by 2.83 minutes. About 31%. The gains were in follow-up scans, not baseline reads — the AI contribution was essentially zero at baseline, a finding the authors noted and which tends to get omitted when the headline number circulates. It’s a meaningful result. It’s also one study, 35 patients, one institution.

QMENTA’s published uptime figure is 99.977%.1 Translating that to operational reality: a four-hour system outage in a global trial with continuous enrollment is missed appointments, sites improvising, backlogs. The uptime number exists because the alternative has consequences.

Phantom harmonization gets less attention than any of this. Running standardized phantom protocols across a hundred sites and actually maintaining them — centrally reviewing the data, trending it, catching drift before it affects patient scans — is unglamorous work that doesn’t produce a press release. It also doesn’t show up as a gap in your audit findings, which is the goal.

The regulatory piece

Algorithm version control is the requirement that consistently comes up too late in planning. Statistical analysis plan lock and AI algorithm version lock happen together. The FDA and EMA want documentation that the algorithm processing images at site one in month one is the same algorithm at site fifty in month eighteen. This is not ambiguous and it’s not new, but it still surfaces as a surprise in programs that added AI components to imaging workflows without fully thinking through the regulatory implications.

External audits of imaging core labs are adversarial by design. QMENTA has reported zero nonconformities across multi-year, multi-site studies.1 That audit record is worth asking for directly when evaluating imaging infrastructure partners, rather than relying on what’s in the sales materials.

The part technology doesn’t solve

AI tools handle certain categories of error well and have hard limits on others. Version-locked models — which is what regulatory compliance requires — can’t adapt when something unexpected starts appearing in the image data mid-trial. Edge cases, unusual pathology, protocol amendments: these land on a radiologist’s desk regardless of what the automated system is doing. That’s not a criticism of the technology; it’s a description of what it currently is.

Site engagement is a separate problem entirely and doesn’t have a software solution. A platform can flag a deviation the moment it occurs. Whether the site coordinator, investigator, and technologist respond — and whether that response holds across a two-year trial with staff turnover and competing priorities — comes down to training, relationships, and the occasional difficult conversation about performance. The data infrastructure makes that conversation more grounded. It doesn’t replace it, and programs that expect it to tend to be surprised by what slips through anyway.

Sources

  1. QMENTA internal operational data and published case study results, qmenta.com. Figures are vendor-reported.
  2. Getz, K. Quantifying the Value of a Day of Delay in Drug Development. Tufts CSDD / Applied Clinical Trials, June 2024. The $1.5M/month figure is derived from $55,716/day × 30. Also cited in Contract Pharma, September 2024.
  3. Tabelow, K. et al. Standardized Brain MRI Acquisition Protocols Improve Statistical Power in Multicenter Quantitative Morphometry Studies. PMC7391934; Teipel, S. et al. Reliability of brain volumes from multicenter MRI acquisition. PMC6872009.
  4. Sieber, V. et al. Automated assessment of brain MRIs in multiple sclerosis patients significantly reduces reading time. Neuroradiology 66, 2171–2176 (2024). DOI: 10.1007/s00234-024-03497-7.
  5. AI in Medical Imaging Statistics 2026. SQ Magazine, April 2026.