An Ångström
An Ångström (Å) is a unit of length equal to 10−10 metres; one ten-billionth of a metre. It sits at a comfortable scale for the atomic world, with the diameter of a hydrogen atom, the length of a chemical bond, all measured in Ångström.
It is not an International System of Units (Système International d’Unités) “SI” unit. In fact, it has been formally deprecated in favour of the nanometre (1 Å = 0.1 nm), and standards bodies such as NIST and the BIPM discourage its use. Yet, in structural biology and chemistry, crystallography, and materials science, the Ångström persists. I would say, partly out of stubbornness, but mostly out of convenience. Saying a protein structure was solved at 2.1 Å feels natural in a way that 0.21 nm does not.
So we keep using it. And because we keep using it, we inherit its quirks and history.
Preamble: Missing Z-values in Medical Research
Before I share with you the particular quirk I wrote this blog about, let us discuss a different but hopefully apparently similar anomaly.
In medical statistics, Barnett & Wren (2019) examined decades of Medline papers and noticed something odd: missing Z-values. More precisely, there was a suspicious scarcity of results just shy of conventional significance thresholds.

This is the statistical analogue of a cliff edge: results pile up just beyond p < 0.05, and vanish tangentially at p 0.05 (corresponding to z-values of +/- 1.96). You have probably seen this before, a critical value of 0.05 is by and large the default of statistical thresholds, in pretty much every major scientific field.
Two well known mechanisms explain this:
- Publication bias: non-significant results are less likely to be published
- P-hacking / “researcher degrees of freedom”: analyses are nudged until they cross significance thresholds
van Zwet & Cator (2021) describe this as the “significance filter” where only results that pass a threshold survive to be seen. The consequence is a distorted distribution: not a smooth continuum, but one with suspicious gaps and spikes. This raises the broader questions:
When humans report numbers, do we see the underlying truth? Or just the thresholds we (and what we believe others) care about? Regardless, it does not help the overwhelming nature of scientific discovery being abundant with true positives, and very few negatives (at least in terms of significance).
The RCSB-PDB
With that question in mind, let us turn to structural biology.
Using the RCSB Protein Data Bank API, I retrieved 199,761 protein structures solved by X-ray diffraction. For each, I extracted the reported refined resolution, the canonical measure of structural quality for the model, as reported by the authors who deposited the structural model.
At first glance, everything looks reassuringly normal.

- The mean resolution is 2.11 Å
- The distribution is unimodal, centred around 2 Å
- There is a long tail toward poorer resolutions
This aligns with intuition: 2.0 Å is often considered a “good” structure, high enough to resolve side chains reliably, low enough to be experimentally tractable. The 2.0 Å is very much analogous to a p < 0.05, it is used not only in determining structure, but also in measuring algorithm performance in areas such as drug discovery where, much like the p-value, it is sometimes abused.
The Illusion of Smoothness
When we bin the data coarsely (0.2 Å bins, Figure A), the distribution looks smooth, almost Gaussian-like around its peak. But smoothness is a function of resolution, not of the X-ray data in this case, but of the visualisation.
Zooming In
Now we increase the granularity (Figure B): 0.01 Å bins, focusing on the 2.0–2.5 Å range. Do you see it too? The smooth distribution fractures.

Instead of a continuous curve, we see distinct spikes:
2.00 Å
2.05 Å
2.10 Å
2.15 Å
2.20 Å
… and so on
What’s Going On?
For me, several hypotheses present themselves:
- Rounding and Reporting Conventions. Crystallographic refinement pipelines often produce resolutions with limited precision. Authors may: round to 2 decimal places, round to “nice” increments, report conservative values.
- Software Defaults and Pipelines. If widely used tools output resolution in fixed increments, this introduces systematic quantisation across the entire field.
- Psychological Thresholds. Just like p=0.05, structural biology has its own soft thresholds: “~2.0 Å” = good, “<2.5 Å” = acceptable, “<3.0 Å” = usable. If a structure refines to 2.04 Å, is it reported as 2.04, or nudged to 2.00 or 2.05? Even without misconduct, human preference for round numbers can shape distributions.
- Selection and Filtering. Structures just above key thresholds may be: deprioritised for deposition, less likely to be written up, filtered during curation. This would mirror the significance filter in statistics.
Same, Same, but Different?
Now when we reconsider the Z-value distribution from the medical literature. We see (well, I do): depletion, inflation just beyond thresholds (e.g. p values of 0.05, 0.01, versus 2.0 Å) and asymmetry driven by selection.
So is it bias, and is it benign? There are strong benign explanations: instrument precision limits, software discretisation, historical reporting standards, “harmless” rounding and so on.
But regardless we end up with systematic artefacts, that affect meta-analyses, will ultimately influence ML models trained on structural data, and shape our intuition about what “typical” data, thresholds/cutoffs looks like.
The Broader Point
This is not really about Ångströms (well, it is). What I wanted to demonstrate was more about measurement, reporting, and the quiet ways human choices imprint themselves onto data. We like thresholds. We like round numbers. We like categories. We love a decision boundary. Nature does not.
Citations
Barnett AG, Wren JD Examination of CIs in health and medical journals from 1976 to 2019: an observational study BMJ Open 2019;9:e032506. doi: 10.1136/bmjopen-2019-032506
vanZwet EW, Cator EA. The significance filter, the winner’s curse and the need to shrink. Statistica Neerlandica. 2021;75:437–452. https://doi.org/10.1111/stan.12241
