Ivan Calandra , 2024-09-16, 17:01:45
- Introduction
- Scale of observation
- Magnification
- Resolution
- Scale, magnification, resolution - summary
- Processing
- Saving
- Reporting and sharing
- How to contribute
- License
- References
Recommendations on how to publish microscope images
The releases are available and citable on Zenodo
This good-practice document is directed at archaeologists and paleontologists working with microscopes, but anyone working with microscope images might learn a few things here.
Microscopes now deliver digital images. “Digital images are data” (Cromey 2013). However, digital images are different from analog images and from what the observer/analyst sees through the oculars. Therefore, digital images have specific properties that must be reported in publications, so that other researchers have the necessary details to assess the published images and also so that they can repeat/reproduce these images.
This document is by no means exhaustive and is limited in scope on purpose, but I hope it will help many archaeologists and paleontologists.
As a disclaimer, this document only presents my opinion, at least for now. You are free to disagree. Whether you agree or not, I would really appreciate if you could share your opinion and contribute to the development of this document, so that it can be accepted by a larger community.
A microscope is obviously used to observe features of an object that are too small to be visible to the naked eye. To do so, a series of lenses magnifies the object so that the analyst sees more details. But which scale(s) is (are) appropriate and how to reach this (these) scale(s) with a microscope?
So, the first question the analyst should ask is: What is the size of
the features of interest? It is important to consider the range of
sizes: the size of the smallest features is important, but so is the
size of the largest. This is because the larger this range, the less
likely it is that all these sizes can be observed with a single set of
settings. In other words, if features of different sizes are of
interest, several acquisitions at different scales will probably be
needed.
Finding the size(s) of the features is often done directly during
observation: increasing the magnification until the features of interest
are visible. This is a valid approach if the microscope used has the
appropriate hardware (objectives, zoom, camera, etc.) to resolve the
features of interest. But it can also happen that the features of
interest are smaller or larger than what is permitted by the hardware at
hand. Moreover, when dealing with digital images, some other issues
might occur. This is why it is always advisable to start with at least a
rough idea of the size(s) of the features and of the capabilities of the
microscope(s) available to the analyst.
But how do we know what the microscope’s hardware can achieve in terms of scale? Most publications and reports talk about magnification. But what is it really and what does it mean? In the next section, I will argue that magnification is pretty much useless in digital microscope imaging and actually only confuses the analyst. Resolution is much more important, but it is unfortunately still cryptic to many analysts; I will therefore try to explain some concepts afterwards.
Magnification is simply: the size of the magnified area divided by the real size of the imaged sample’s area.
But let us first have a look at the optical magnification.
The optical magnification of a microscope image when observed through
the oculars is straightforward to calculate:
For example, using a 20
Things get more complicated in the digital world. It all comes down to one single question: what is THE size of the magnified area when stored on a digital image? Is it the size of the screen when you acquire the image? Is it the size of the screen when you view the image? Is it the size printed on paper, or projected to a screen? There is no single size for a digital image because it depends on the viewing medium: increase the size of the viewing medium and the magnification will increase as well, even though the image is the same. Does that mean that increasing the size of the viewing medium will improve the resolution? Unfortunately, no, as we will see below.
Let us see how the digital magnification is calculated. The problem with
a digital image (as opposed to an analog image) is that many components
come into play to magnify the sample: all the optical (objective,
optical zoom and camera adaptor) and digital (camera sensor, viewing
medium and digital zoom) components must be considered. Note that the
oculars are not relevant anymore because the light goes to the camera
without passing through the oculars.
This is how to calculate the digital magnification:
For example, using a 20
As you can see, with the same optical components (except oculars) as in
the example calculation of the optical magnification above (see formula
(1)), the resulting digital magnification is about 3.5
An example I like to show is the figure 8a of Pedergnana
(2020) (the paper can be read
here). According to the legend, the original
magnification of the optical light microscopy image on the left is 10
Also, the formula (2) shows that calculating the total digital magnification is not so easy in practice: the camera sensor diagonal and the viewing medium diagonal are not always known.
So what now? Should we all forget about magnification in digital microscopy? Well, pretty much, in my opinion.
If the analyst has observed and analyzed while looking through the ocular, the optical magnification is very relevant and should be reported (see section Reporting and sharing). However, the image published is a digital one. This implies that the analyst and the readers/audience will observe the sample at different magnifications (and resolutions as well, see section Resolution). In some cases, the readers/audience might not be able to see what the analyst is referring to. This might result in misunderstanding and potentially even disbelief in respect to the analyst. Even when the analyst analyzed the digital image directly, the viewing medium diagonal might change for each person anyway (laptop screen vs. desktop screen vs. projected screen vs. printed…).
As a side note, the field of view of a digital image is usually smaller (to avoid vignetting) than the field of view of the image as seen through the oculars, and the contrast and clarity when looking through the oculars is usually much better than any digital image. These are other sources of differences when the analyst observed through the ocular but reports/publishes a digital image.
There is another reason for avoiding the digital magnification. When I
was tasked to compare different digital microscopes (i.e. microscopes
without oculars and where the image is directly and only viewed on a
screen), I noticed that there were very large variations in the
magnification ranges advertised by the different manufacturers. Looking
at the small print, I realized that, for example, one manufacturer
reported the magnification as seen on a 15” screen diagonal, while
another one as seen on a 17.5” screen diagonal. This made sense because
this was the size of the image as seen in the respective acquisition
software packages. But this is very confusing to the buyer/user.
Just for the comparison, let us simply change the screen from 15” to
17.5” (444.5 mm) in the example above:
So this makes a significant difference!
There is another similar example, namely in SEM imaging. When using an
SEM, only the digital magnification is availble, but it is given
relative to a reference. For example, in the case of our Zeiss EVO 25,
the magnification is given relative to the Polaroid 545 format (other
formats can be selected), i.e. as if printed on 4
So, if the digital magnification is useless, how can we know the size of a digital microscope image? The good old scale bar! The field of view, i.e the size of the imaged area of the sample, or rulers or anything similar will do the trick too.
One would expect that everybody knows that a scale bar (or some form of
scaling) must be provided with each image. Yet, “[a]pproximately half
of papers in physiology (49%) and cell biology (55%) and 28% of plant
science papers provided scale bars with dimensions (in the figure or
legend) for all images in the paper […]. Approximately one-third of
papers in each field contained incomplete scale information, such as
reporting magnification or presenting scale information for a subset of
images. Twenty-four percent of physiology papers, 10% of cell biology
papers, and 29% of plant sciences papers contained no scale information
on any image” (Jambor et al. 2021, 3).
In sum, only about half of the papers in these fields publish
microscope images with proper scaling information; this is not much! I
do not expect that archaeology is any different, although I am not aware
of such a survey for archaeology.
In general, resolution refers to the level of details that can be observed. In other words, it defines the size of the smallest observable features.
There is not one single resolution but many different, complementary definitions of resolution. First, it is important, once again, to differentiate between optical and digital resolutions. And second, there are lateral (XY) and vertical (Z) resolutions.
Most terms and definitions here refer to the Fair Data Sheet Initiative, v1.2α (the website is only in German, but the Fair Data Sheet itself has an English version). I will abbreviate it “FDS”.
Due to aberrations from the optics and diffraction of the light, an optical system will be limited in its ability to distinguish points; the image will become blurry at the smallest scales. These effects are the basis for the calculation of the optical lateral resolution as defined by the Rayleigh criterion.
The calculated optical lateral resolution is the “minimum theoretical distance between two adjacent, barely distinguishable features of an object” (FDS §2.2.8). In other words, if two points are closer from each other than the optical lateral resolution, then it is not possible to distinguish them (i.e. they will look like one point only). They will appear as two separate points only if they are further from each other than the optical lateral resolution.
The calculated optical lateral resolution
The value of
From formula (3) and the definition of the optical lateral resolution,
it becomes apparent that the smaller
The NOTE 3 of the FDS §2.2.8 states that “the optical [lateral] resolution can be achieved only under ideal conditions (including complete illumination of the pupil) on planar objects. This value is usually not reached on textured surfaces”. In archaeology and paleontology, we usually deal with textured surfaces!
The digital lateral resolution, or measuring point spacing, is the “sampling interval of measuring points in the measuring volume, both in X and in Y direction” (FDS §2.2.7). This is also referred to as the pixel size (although pixel size or pixel pitch can also refer to the size of the photodiodes on the detector/camera, which is a fixed value independent of the field of view) or spatial sampling.
The measuring point spacing is calculated by dividing the measuring area (FDS §2.2.1, also called field of view) by the maximum number of measuring points in a single measurement (FDS §2.1.2, also called frame size or number of pixels):
So far, so good, and not too complicated. The problem is that we now have two different values for the lateral resolution. Which one should we care about? Both, of course! But what should the relationship between optical and digital lateral resolutions be?
The Nyquist criterion (based on the Nyquist–Shannon sampling
theorem)
states that the value for the digital lateral resolution (measuring
point spacing) should be 2-3 times smaller (approx. 2.8 times smaller
according to Pawley 2006) than the value for optical
lateral resolution (
For example, if
If the digital lateral resolution is more than 3
If the digital lateral resolution is less than 2
There are many considerations to take into account in the analog to digital conversion process, but as Pawley (2006, 70) wrote: “If you don’t want to worry about any of this, stick to Nyquist!”
Note that in the section Optical lateral
resolution, I argued that we should avoid
the terms “larger” and “smaller” when talking about resolutions. But in
this section, it is really about the relationship between the values of
Formula (2) includes the digital zoom in the calculation of the digital magnification, but it is not part of formula (4) for the digital lateral resolution. So what happens when zooming in digitally?
Let us start with an original image 100
The digital zoom acts like a cropping + enlarging tool: the original
image will be first cropped and then enlarged to the same viewing size
as the original image.
A digital zoom factor 2
So what about the digital lateral resolution of the zoomed image? 0.1
µm, just like the original image.
All this to say that the digital zoom obviously changes the digital magnification but does not influence the digital lateral resolution.
Still, it seems that zooming in digitally helps to see more details. This is just because the details are displayed larger on the screen, not because they are better resolved. For a better digital resolution, a smaller field-of-view for the same number of pixels, or more pixels for the same field-of-view, is needed. This can be achieved by using a higher magnification objective and/or an optical zoom, or a camera with more pixels.
The optical vertical, or axial, resolution is equal to the depth of field for a bright field (BF) microscope:
with
For a confocal microscope, the axial resolution is calculated by the optical slice thickness, which also depends on the pinhole diameter (Artigas 2011):
Here again, the smaller
The digital vertical resolution is meaningless in case of a 2D image (even for extended depth of focus - EDF - images) but it is relevant for 2.5 and 3D images.
The digital vertical resolution of a 2.5D image acquired from a confocal
microscope is better than the step size (i.e. the distance between the
individual images, or slices, of a stack) because of the way it is
processed (see Artigas 2011).
In short, the focal distance is constant for a given objective, so when
the microscope’s objective moves up and down, different points of the
sample are in focus. Each point (pixel) on each slice records the
intensity of the signal (called axial response) at that location (X+Y
from stage coordinates and Z from the slice’s height). The intensity
increases when the focal plane gets closer to the point on the sample
and decreases again when the focal plane moves further away from the
point on the sample. The curve of the axial response (intensity)
vs. height of the slice therefore has a maximum that can be
mathematically computed (see fig. 11.6 in Artigas
2011) and that do not necessarily fall on the height
of a given slice. This maximum is the value used for the height of the
given pixel. Repeat the process for all pixels, and you get a 2.5D
height map with a digital vertical resolution better than what the
mecanics (drives or piezo) can achieve.
Unfortunately, I do not know how to calculate the digital vertical
resolution of a z-stack.
If anyone knows more about it, please contribute!
Magnification and resolution are different concepts and are not interchangeable. Resolution is given in a unit of length (e.g. µm), while magnification is without unit (“$\times$”).
Magnification can be understood as an aid to see the resolution: resolution will define what can be seen or not, while increasing the magnification will make the visualization of the observable details easier. This is why, in my opinion, talking about resolution is much more useful and relevant than magnification, especially in the digital world.
Scale is a fundamental concept for any observation. Magnification is often, unfortunately, used as a proxy for scale. However, magnification is quite decoupled from scale. Scale can be properly described by field of view (FOV, i.e. the size of the region observed) and resolution (i.e. the level of details visible on the optical and digital images). Magnification informs only on the FOV, but even for that, it is not very useful when most microscopes now allow for stitching in order to increase the FOV at constant magnification and resolution.
As a side note for use-wear analysts, the two approaches to use-wear analysis are called high-power and low-power, not high-magnification and low-magnification!
In summary, forget about magnification and learn about resolution!
Few images are published without some form of processing. Sometimes,
processing is even necessary to create (e.g. EDF, stitching, topography
reconstruction) or analyze (e.g. contrast/brightness, wavelength
filters) the image. Even the raw data output by every instrument are
processed to some degree.
While processing deserves a whole good-practice document on its own, I
would just like to give some general advices here that will hopefully be
applicable to most processing routines. But feel free to
contribute for more!
Following the definitions of Aaron and Chew (2021, 1), processing “serves to digitally transform an acquired dataset by enhancing or isolating signals of interest and/or suppressing other signals and noise that will otherwise prevent accurate analysis”. Analysis refers to “a vast set of diverse approaches to extract […] meaningful, quantitative measurements from a dataset”. Lastly, pre-processing includes the “steps that are required to be taken before image data can be properly processed […]. Such steps generally do not serve to enhance particular features in an image per se, but rather correct for imperfections commonly encountered in imaging systems.”
Here are some examples in each category when considering surface texture analysis (dental microwear texture analysis or quantitative artifact microwear analysis):
- pre-processing: topography reconstruction from a Z-stack, stitching…
- processing: leveling, form removal, thresholding, wavelength filters…
- analysis: texture analysis following ISO 25178, scale-sensitive fractal analysis, furrow analysis…
In the remaining of this section, I will use the term processing to include both pre-processing and true processing. This makes the sentences easier to read and the discussion applies to both anyway.
Even if we sometimes tend to assume that there is only one way to process an image, there are in fact often several methods or algorithms to do so, and there are always various settings to adjust. Additionally, some processing happens before you get the raw data from the instrument. Lastly, “[a]s images are numerical data, image processing invariably changes these data and thus needs to be transparently documented” (Schmied et al. 2023, 7). This is why it is important to be as precise as possible when reporting about the acquisition and processing (see section Reporting and sharing).
It is also crucial to process all images of a dataset in the same
way. This might seem obvious, but different processing methods might
produce results that might appear identical/similar even if they
actually are different. Processing all images in the same way ensures
that all images are comparable at least within a dataset. Of course,
this is true only if the images were acquired in the same way!
So, while the details are more nuanced, I believe the message is clear
and every analyst should try his/her/their best to acquire and process
images and data in a consistent and comparable manner.
Another general piece of advice is to keep the raw data, in this case the acquired image, and to process a copy of it rather than overwriting the acquired image (see e.g. Cromey 2013; Schmied and Jambor 2021). This way, it is possible to check the result and process differently if necessary. See section Saving for a discussion on raw/derived data and file formats.
Last but not least, it is important to keep a very detailed record of the processing and analysis steps that were performed. Templates and scripts will make it easier to repeat and reproduce the processing/analysis. This is not only important to share data (see section Sharing); it is also important for yourself as it is likely that you will have to repeat the analysis at some point, at least for some images. protocols.io can be a useful tool for this.
Fiji/ImageJ is a great, open-source tool for image analysis. Schmied and Jambor (2021) proposed guidelines for image analysis and a processing workflow in Fiji; there is no reason not to follow and use them! The checklists of Schmied et al. (2023) are also very useful.
I totally agree with Ben Marwick and colleagues that raw data should be kept raw and clearly separated from derived data (e.g. Marwick and Pilaar Birch 2018; Marwick, Boettiger, and Mullen 2018).
Still, it is not always clear to distinguish between the two, nor practicable to keep the raw(est) data. For example, the raw data of an EDF image or a 2.5D topographical model (= height map, e.g. from a confocal microscope) is a Z-stack. Z-stacks can be very large files so keeping all these Z-stacks requires a lot of storage space. While storage is not really limiting anymore, the environmental impacts of data centers and online repositories cannot be ignored (e.g. Samuel and Lucivero 2020).
Also the derived data of an analysis can be the raw data of another. Using the example of surface texture analysis (dental microwear texture analysis or quantitative artifact microwear analysis), the raw data of the analysis are the height maps (which are themselves derived from pre-processing the Z-stacks) that are used as input for MountainsMap (or equivalent) to calculate the surface texture parameters for each height map, which are eventually exported into a CSV file. This CSV file is in-turn the raw data for the subsequent statistical analysis.
All this to say that the decision should be made on a case-by-case basis about what data is raw and what is derived. Nevertheless, there are some general principles that apply in most (all?) cases. Make sure to:
- save the rawest data possible, usually the acquired, unprocessed image;
- save every derived data that is included in the final analysis (i.e. input data of the results), especially if the processing/analyses are not fully automatized;
- save all the metadata associated with both the raw and derived data (see section File formats and software);
- save in appropriate format(s) (see section File formats and software).
For long-term usability, accessibility and sustainability, proprietary formats (i.e. files that can only be opened with a paid software) should be avoided and stable, open, non-proprietary formats should be preferred. The DANS maintains a list of preferred and non-preferred formats. For microscopy images, the OME-TIFF format has become the standard.
Most microscope manufacturers offer software packages to acquire images;
in some cases, these software packages are even necessary to operate the
microscopes. Not all of these software packages can save in an open
format, even less so in OME-TIFF. Unfortunately, metadata sometimes get
lost when the image is not saved in the native, proprietary format.
This is why, if your acquisition software cannot natively save in an
open format, I would recommend to save the image in the
proprietary format (in order to save all the metadata) and to
additionally export to an open format (for long-term accessibility).
Sometimes, it is also possible to export the metadata into an open
format (e.g. JSON, XML, CSV or TXT).
Following on this discussion, it is therefore important to favor
open-source software packages over paid ones. Nevertheless, open-source
projects often have difficulties to keep running and being developed
(see e.g. Coelho and Valente 2017). Also, open-source
software packages are often less user-friendly, and are slower to add
new functionalities and to adapt to OS upgrades. So, open-source is not
always better.
Still, Fiji/ImageJ is a great open-source
solution for image analysis with a 25-year-long history (meaning it is
likely to survive).
Bio-Formats is an
important “software tool for reading and writing image data using
standardized, open formats”, which is incorporated into ImageJ. Check
the how-to ImageJ2-Fiji for details on how
to read Zeiss files and access their metadata with this software.
Some microscope manufacturers also offer a free viewer or a free light
version (e.g. Zeiss’ ZEN
starter
and ZEN
lite).
Following up on the previous section, every one of us has wondered at least once which image format should be preferred and what the differences are between JPG, JPEG, PNG, TIF and TIFF. Here, I will try to give an overview and general recommendations. This section concerns only raster image formats, as microscope images are always raster images.
I recommend reading the documentation about the different raster file formats on the Adobe website, which is, in my opinion, clear and detailed.
- TIFF (.tif or .tiff, same format but different extensions for historical reasons) is the format to choose if you want to save images at full resolution and with best quality, which in turn implies that TIFF files are large. Metadata can be saved together with the image when using this format (e.g. for SEM images).
- JPEG (.jpg or .jpeg, same format but different extensions for historical reasons) is a widely used format for photos. Files are small and of generally good quality, but the lossy compression means that some details are lost. JPEG files can include metadata, especially the camera’s EXIF data, but is not recommended for scientific imaging (Cromey 2013).
- PNG is best knowned for their transparent backgrounds, but it is also a good format for any type of image thanks to its lossless compression. Files are smaller than TIFF but larger than JPEG. While many users tend to prefer JPEG, I think PNG could be better for many applications and should be preferred to JPEG (Cromey 2013; Schmied and Jambor 2021; Schmied et al. 2023). Compare the two formats here.
- I am not aware of any microscope camera that can save in RAW, but DSLR cameras can usually do it. So this format is worth considering for users who have a DSLR camera mounted on the phototube.
“The underlying premise of image publication and ethics guidelines is that a digital image is data and that the data should not be manipulated inappropriately” (Cromey 2013, 5).
All the background information of the previous sections aims at providing the required knowledge to use a microscope to reach the desired scale of observation and to save all the acquired data and metadata. While it is important for the analyst to see what he/she/they want to see on the sample, it is equally important that the reader/audience that will be presented - or even will assess - the work is able to know what can be observed in terms of scale on the reported/published/shared images. This is also important for repeatability and reproducibility (and pre-producibility sensu Stark 2018).
The importance of reporting for microscopy is not unique to archaeology/paleontology (see e.g. Calandra et al. 2019) and other fields also suffer from a lack of high-quality reporting (see e.g. Marqués, Pengo, and Sanders 2020 in biomedical research).
A large community of researchers in life sciences are working on guidelines for reporting microscope data (e.g. Aaron and Chew 2021; Hammer et al. 2021; Heddleston et al. 2021; Jambor et al. 2021; Ryan et al. 2021; Sarkans et al. 2021; Schmied et al. 2023, as well as the Open Microscopy Environment - OME, the European Light microscopy initiative - ELMI, the Society for Microscopy and Image Analysis - GerBI-GMB and the Quality Assessment and Reproducibility for Instruments & Images in Light Microscopy - QUAREP-LiMi communities). The work they do is incredible, but unfortunately, it concerns mostly life sciences and some aspects are not relevant to archaeology, while some others relevant to archaeology are not addressed.
Below is a list of what I think is necessary for pre-producible microscope images. This list should be adapted according to the equipment/software used.
Number | Category | What | Why | How? (examples) |
---|---|---|---|---|
1 | Observation/analysis method | Observation/analysis through oculars or directly on the digital image? | To know how the analyst performed the analysis | “The features were observed and analyzed through the oculars. The images reported are therefore not equivalent to what has been observed during analysis.” |
2 | Microscope | Manufacturer and model of the microscope | To account for all (unknown) settings that are part of the microscope’s design | “AxioScope.A1 (Carl Zeiss Microscopy GmbH)” |
3 | Technique | Imaging technique / contrast | To interpret the image according to the technique(s) used | “Confocal microscopy” or “bright field” or “DIC” |
4 | Light | Illumination | To interpret the image according to the illumination(s) used | “Ringlight illumination” or “Coaxial illumination” |
5 | Light | Type + wavelength of the light source | To calculate the optical resolutions | “Violet LED 405 nm” |
6 | Light | Light source intensity | To be able to repeat the acquisition | “60%” |
7 | Light | Exposure time | To be able to repeat the acquisition | “100 ms” |
8 | Optics | Objective specifications | To calculate the optical magnification (if applicable, see #1) and optical resolutions | “Objective Zeiss C Epiplan-Apochromat Mag. = 50 |
9 | Optics | Oculars magnification, if applicable (see #1) | To calculate the optical magnification | “10 |
10 | Optics | Optical zoom, if applicable (see #1) | To calculate the optical magnification | “2 |
11 | Camera | Measuring point spacing + field of view + number of pixels (at least two of them) | To calculate the digital lateral resolution | “100 |
12 | Camera | Scale bar | To visualize the size of the imaged region | “Scale bar = 100 µm” |
13 | Confocal | Pinhole diameter, if applicable | To calculate the optical resolutions | “PH = 54 µm (1 AU)” |
14 | Confocal | Step size + z-range, if applicable | To be able to repeat the measurement | “40-50 µm with 0.2 µm steps” |
15 | Confocal | Temperature + humidity during topography scanning, if possible | T & H can influence the fine-scale measurements of topography | “Scans were acquired at 20 < T < 22 °C and 50 < rH < 70 %” |
16 | Confocal | Location of the equipment | An environment without vibrations is necessary for fine-scale measurements of topography | “The confocal microscope is located in the basement and placed on a passive anti-vibration table, itself placed on a concrete based decoupled from the rest of the floor” |
17 | Confocal | Any other equipment-specific setting | To be able to repeat the measurement | “Master gain = 220 V” |
18 | Software | Versions of acquisition, processing and analysis software packages | Different versions might produce different results | “ZEN blue 3.5 Hotfix 7 (Carl Zeiss Microscopy GmbH) with module Shuttle-and-Find” or “ConfoMap 9.2.10042 with module Scale-sensitive Analysis (Digital Surf)” |
19 | Pre-processing | Stitching, if applicable | To be able to repeat the processing | “2 |
20 | Pre-processing | Extended depth of focus (EDF), if applicable | To be able to repeat the processing | “Z-stacks were processed to EDF images with the wavelets method and without z-stack alignment” |
21 | Processing/Analysis | Image enhancements / filters / image analysis settings, if applicable | To be able to repeat the processing | “The contour were enhanced with settings A and B; and brightness, contrast and gamma were adjusted to C, D and E respectively” |
22 | Processing/Analysis | Workflow and settings of the surface texture analysis, if applicable | To be able to repeat the processing | “The surface texture analysis was performed with the following template [add analysis details and link to template]” |
From my point of view, all of these pieces of information should be published together with every microscope acquisition.
I would like to stress that most of these metadata are generated and saved automatically together with the image data (at least when you use the appropriate format) so that there is no extra work needed to compile the image metadata beyond sharing (see section Instrument-specific guidelines). I therefore do not see a reason why we should not report them.
Additionally, I recommend following the guidelines of Jambor et al. (2021) to improve the quality of the images themselves, as well as using the checklists of Schmied et al. (2023) to make sure that at least the relevant “minimal” requirements are met (but aim for the “ideal” requirements!).
A lot could be said about data sharing and Open Science. I will keep it brief here and focused on microscope images; more general and extensive discussions can be found in e.g. Marwick and Pilaar Birch (2018), Gomes et al. (2022), and Karoune and Plomp (2022).
In my opinion, publishing only a few, supposedly representative images of the whole dataset is not enough because of the following reasons:
- The sample might be representative in some aspects but not in others.
- To be honest, in most cases, only the “nicest” images are published and the least informative/relevant are not; this implies a strong bias and the true signal might just be lost by this selective sampling/reporting.
- Images published in a journal article or in a book are usually small and of reduced quality (in the PDF at least), so the published images might not even show what can be seen on the full-resolution image.
- Without seeing all images in full resolution, how can anyone assess whether the data support the conclusions?
- It would be a shame to keep all these images for yourself.
- Why should we show only a small part of the extensive work we did? Let us do ourselves a favor and show how hard we work!
Or as Cromey (2013, 3) puts it: “If we only show other people the pictures we want them to see, then the viewers will most likely reach the interpretation that we want them to have. What if our interpretation is wrong? It may be a poorly kept secret that most published images are somewhat less than “representative”, but is this right? What would happen to the quality of science if we only showed our most compelling images to our project leaders, lab group, or collaborators?”
I would therefore recommend:
- to keep showing a sample as representative as possible in the journal article, but
- to additionally upload all images with their metadata to an online repository.
I would like to stress that the data should be uploaded to a trustworthy
online repository (e.g. Zenodo,
OSF or Figshare), rather
than as online supplementary information (SI) to the article. This is
because SI might be behind a paywall, just like the article. Even when
it is not, publishers might have restrictions about file formats and
names. And since everyone does mistakes, it is easier to correct them if
you have access to the repository rather than relying on the publisher
to do it (which can take a long time or might even never happen).
Ideally, data should be uploaded to a discipline-specific repository
rather than a generalist repository like Zenodo or OSF. But I am not
aware of any archaeology-specific repository that is commonly used. I
have heard about for example The Digital Archaeological Record
(tDAR) but it seems to be mostly used for
excavation data. Please contribute!
Here again, the discussion about raw/derived data and file formats is important (see section Saving): both raw and derived data should be shared in formats that include all the metadata (Schmied et al. 2023).
I propose step-by-step instructions to upload files to Zenodo in how-to Zenodo.
I have tried to make it easy for the users of the Imaging Platform at LEIZA (IMPALA): a Shiny App guides the users to fill in the required information and exports the report to an XLSX or ODS file that can be shared together with the images (see section Sharing). This App is still in development and is part of the repository imaging-reports on GitHub.
To complement this, I have prepared guidelines specific to the
instruments we have at the IMPALA in the folder
Guidelines. The guidelines include text snippets that
can be copy/pasted/edited for the method section of a publication and
have been converted from Markdown to DOCX format using the extension
vscode-pandoc.
They also explain in details what type of data should be shared and
how.
Feel free to contribute guidelines for other
instruments!
Neither the reports nor the text snippets are meant to be used for
every acquisition. Instead, they serve to summarize the acquisition and
analysis processes and should be used only once for each
publication/project.
But this works well only in combination with the detailed metadata
shared together with the data themselves, as detailed in the
guidelines (see also section Sharing).
I appreciate any comment from anyone (expert or novice) to improve this document, so do not be shy! There are three possibilities to contribute.
-
Submit an issue: If you notice any problem or have a question, submit an issue. You can do so here.
A GitHub account is necessary. -
Propose changes: This document is written in Quarto. If you know how to write in this format, please propose text edits as a pull request (abbreviated “PR”).
A GitHub account is necessary. -
Send me an email: If you do not have a GitHub account and do not want to sign up, you can still write me an email (Google my name together with my affiliation to find my email address).
By participating in this project, you agree to abide by our code of conduct.
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
See also License file in the repository.
Author: Ivan Calandra
License file, badge and image from Soler (2022).
Aaron, Jesse, and Teng-Leong Chew. 2021. “A Guide to Accurate Reporting in Digital Image Processing Can Anyone Reproduce Your Quantitative Analysis?” Journal of Cell Science 134 (6): jcs254151. https://doi.org/10.1242/jcs.254151.
Artigas, Roger. 2011. “Imaging Confocal Microscopy.” In, edited by R. Leach, 237–86. Berlin, Heidelberg: Springer Berlin Heidelberg. https://doi.org/10.1007/978-3-642-12012-1_11.
Calandra, Ivan, Lisa Schunk, Konstantin Bob, Walter Gneisinger, Antonella Pedergnana, Eduardo Paixao, Andreas Hildebrandt, and Joao Marreiros. 2019. “The Effect of Numerical Aperture on Quantitative Use-Wear Studies and Its Implication on Reproducibility.” Scientific Reports 9 (1): 6313. https://doi.org/10.1038/s41598-019-42713-w.
Coelho, Jailton, and Marco Tulio Valente. 2017. “ESEC/FSE’17: Joint Meeting of the European Software Engineering Conference and the ACM SIGSOFT Symposium on the Foundations of Software Engineering.” In, 186–96. Paderborn Germany. https://doi.org/grmgjz.
Cromey, Douglas W. 2013. “Digital Images Are Data: And Should Be Treated as Such.” In, edited by Douglas J. Taatjes and Jürgen Roth, 1–27. Methods in Molecular Biology. Totowa, NJ: Humana Press. https://doi.org/10.1007/978-1-62703-056-4_1.
Gomes, Dylan G. E., Patrice Pottier, Robert Crystal-Ornelas, Emma J. Hudgins, Vivienne Foroughirad, Luna L. Sánchez-Reyes, Rachel Turba, et al. 2022. “Why Don’t We Share Data and Code? Perceived Barriers and Benefits to Public Archiving Practices.” Proceedings of the Royal Society B: Biological Sciences 289 (1987): 20221113. https://doi.org/grbx5m.
Hammer, Mathias, Maximiliaan Huisman, Alessandro Rigano, Ulrike Boehm, James J. Chambers, Nathalie Gaudreault, Alison J. North, et al. 2021. “Towards Community-Driven Metadata Standards for Light Microscopy: Tiered Specifications Extending the OME Model.” Nature Methods 18 (12): 1427–40. https://doi.org/10.1038/s41592-021-01327-9.
Heddleston, John M., Jesse S. Aaron, Satya Khuon, and Teng-Leong Chew. 2021. “A Guide to Accurate Reporting in Digital Image Acquisition Can Anyone Replicate Your Microscopy Data?” Journal of Cell Science 134 (6): jcs254144. https://doi.org/10.1242/jcs.254144.
Jambor, Helena, Alberto Antonietti, Bradly Alicea, Tracy L. Audisio, Susann Auer, Vivek Bhardwaj, Steven J. Burgess, et al. 2021. “Creating Clear and Informative Image-Based Figures for Scientific Publications.” PLOS Biology 19 (3): e3001161. https://doi.org/10.1371/journal.pbio.3001161.
Karoune, Emma, and Esther Plomp. 2022. “Removing Barriers to Reproducible Research in Archaeology.” Zenodo, ver. 5 peer-reviewed and recommended by Peer Community in Archaeology, November. https://doi.org/10.5281/zenodo.7320029.
Marqués, Guillermo, Thomas Pengo, and Mark A Sanders. 2020. “Imaging Methods Are Vastly Underreported in Biomedical Research.” Edited by Peter Rodgers and Elisabeth M Bik. eLife 9 (August): e55133. https://doi.org/ghmkv4.
Marwick, Ben, Carl Boettiger, and Lincoln Mullen. 2018. “Packaging Data Analytical Work Reproducibly Using R (and Friends).” The American Statistician 72 (1): 80–88. https://doi.org/gdhvm8.
Marwick, Ben, and Suzanne E. Pilaar Birch. 2018. “A Standard for the Scholarly Citation of Archaeological Data as an Incentive to Data Sharing.” Advances in Archaeological Practice 6 (2): 125–43. https://doi.org/gf5vpk.
Pawley, James B. 2006. “Points, Pixels, and Gray Levels: Digitizing Image Data.” In, edited by James B. Pawley, 59–79. Boston, MA: Springer US. https://doi.org/10.1007/978-0-387-45524-2_4.
Pedergnana, Antonella. 2020. ““All That Glitters Is Not Gold”: Evaluating the Nature of the Relationship Between Archeological Residues and Stone Tool Function.” Journal of Paleolithic Archaeology 3 (3): 225–54. https://doi.org/grmgj2.
Ryan, Joel, Thomas Pengo, Alex Rigano, Paula Montero Llopis, Michelle S. Itano, Lisa A. Cameron, Guillermo Marqués, Caterina Strambio-De-Castillia, Mark A. Sanders, and Claire M. Brown. 2021. “MethodsJ2: A Software Tool to Capture Metadata and Generate Comprehensive Microscopy Methods Text.” Nature Methods 18 (12): 1414–16. https://doi.org/10.1038/s41592-021-01290-5.
Samuel, Gabrielle, and Federica Lucivero. 2020. “Responsible Open Science: Moving Towards an Ethics of Environmental Sustainability.” Publications 8 (4): 54. https://doi.org/grmgjx.
Sarkans, Ugis, Wah Chiu, Lucy Collinson, Michele C. Darrow, Jan Ellenberg, David Grunwald, Jean-Karim Hériché, et al. 2021. “REMBI: Recommended Metadata for Biological Imagesenabling Reuse of Microscopy Data in Biology.” Nature Methods 18 (12): 1418–22. https://doi.org/10.1038/s41592-021-01166-8.
Schmied, Christopher, and Helena Klara Jambor. 2021. “Effective Image Visualization for Publications a Workflow Using Open Access Tools and Concepts.” F1000Research 9 (February): 1373. https://doi.org/10.12688/f1000research.27140.2.
Schmied, Christopher, Michael S. Nelson, Sergiy Avilov, Gert-Jan Bakker, Cristina Bertocchi, Johanna Bischof, Ulrike Boehm, et al. 2023. “Community-Developed Checklists for Publishing Images and Image Analyses.” Nature Methods, September, 1–12. https://doi.org/10.1038/s41592-023-01987-9.
Soler, Santiago. 2022. “Cc-Licenses: Creative Commons Licenses for GitHub Projects.” https://github.com/santisoler/cc-licenses.
Stark, Philip B. 2018. “Before Reproducibility Must Come Preproducibility.” Nature 557 (7707): 613–13. https://doi.org/gdh9xc.