Leonhard Möckl1, W E Moerner1. 1. Department of Chemistry, Stanford University, Stanford, California 94305, United States.
Abstract
Single-molecule super-resolution microscopy has developed from a specialized technique into one of the most versatile and powerful imaging methods of the nanoscale over the past two decades. In this perspective, we provide a brief overview of the historical development of the field, the fundamental concepts, the methodology required to obtain maximum quantitative information, and the current state of the art. Then, we will discuss emerging perspectives and areas where innovation and further improvement are needed. Despite the tremendous progress, the full potential of single-molecule super-resolution microscopy is yet to be realized, which will be enabled by the research ahead of us.
Single-molecule super-resolution microscopy has developed from a specialized technique into one of the most versatile and powerful imaging methods of the nanoscale over the past two decades. In this perspective, we provide a brief overview of the historical development of the field, the fundamental concepts, the methodology required to obtain maximum quantitative information, and the current state of the art. Then, we will discuss emerging perspectives and areas where innovation and further improvement are needed. Despite the tremendous progress, the full potential of single-molecule super-resolution microscopy is yet to be realized, which will be enabled by the research ahead of us.
For much of
the early decades of modern chemistry, it was an undisputed dogma
that it is impossible to directly detect and investigate a single
molecule.[1] Indeed, this task poses unique
challenges: To sense a single molecule, it must be efficiently excited,
and the emitted signal must be recorded with exquisite sensitivity.
Both requirements seemed unachievable for a long time, and with the
exception of certain experiments in vacuum, only large assemblies
of putatively identical molecules were investigated and measured parameters
suffered from ensemble averaging.[2−5] However, the advent of lasers, novel camera
types (such as electron-multiplying charge-coupled device, EMCCD,
cameras), and creative experimental techniques brought single molecules
in condensed phases within reach in the 1980s, culminating in the
first optical detection of a single molecule in 1989.[6] For several years, the precise investigation of single
molecules at low temperatures was a huge playground for physicists:
quantum phenomena, molecular properties, energy transfer mechanisms,
and so on could be studied without ensemble averaging and thus at
the ultimate level of detail.[7−11]A second revolution enabled optical or chemical control over
the emissive state of single molecules or fluorescent proteins.[12] Blinking or fluctuation of single emitters was
initially thought to be a problem,[13] but
this capability eventually brought single molecules and microscopy
together: Super-resolution microscopy with single molecules was born,
which relies on the precise localization of single-molecule signals
to obtain structural information about a labeled specimen at unprecedented
spatial resolution (see section for an overview).[14] With
this, single molecules were no longer only relevant for researchers
interested in fundamental molecular properties: They became a powerful
and versatile tool enabling a large class of super-resolution imaging
methods. Super-resolution microscopy was and still is predominantly
used in biology. This is no surprise—the spatial scale of structures
that can be studied with these approaches is on the order of a few
to a few tens of nanometers. This is precisely the length scale that
is vital for cellular biology. For example, mammalian cells may be
microns in size, but the fine structure of subcellular organelles,
protein complexes, membrane domains, and so on ranges from a few nanometers
to a few hundreds of nanometers. Similarly, bacteria typically exhibit
sizes of few hundreds of nanometers, and proteins within them can
be localized on the nanometer scale.[15,16] Due to a crucial
and fundamental limitation of optical microscopy at the relatively
nondestructive visible wavelengths, the finer architecture of cells
and bacteria is not accessible with conventional light microscopy.
The capability of super-resolution approaches to resolve structures
at the level of single proteins was a considerable improvement, which
could be readily realized as previous research had already established
protocols for specific labeling of cellular targets.Today,
single-molecule super-resolution microscopy is close to
becoming a standard technique: Core facilities regularly implement
it, and many commercial solutions are offered. However, compared to
e.g. confocal microscopy, a large vista is still ahead of us. Improvements
on many methodological fronts are required to make single-molecule
microscopy a more robust and generally useable tool. Without careful
thought in experimental design and execution, the exquisite sensitivity
and resolution pose the risk of producing artifactual results.In this article, after describing the fundamental concepts and
basic experimental considerations as well as surveying the state of
the art, we will discuss two main perspectives for the field: First,
we will focus on innovations needed to improve current implementations
of single-molecule localization microscopy. We will organize these
ideas using the three key experimental areas—sample preparation,
optical setups, and data analysis. Then, we will focus on perspectives
for the field that are emerging and not yet in the center of present
research, i.e. bringing single-molecule super-resolution microscopy
to unexplored topics and joining forces with other powerful methods.
Concept
We will first describe the
key ideas behind single-molecule super-resolution microscopy. For
more in-depth discussions, we refer to the literature.[14,17−23]Conventional optical microscopy with visible wavelengths (approximately
400–700 nm) is fundamentally limited by diffraction which causes
two emitters that are spatially separated by less than ∼200
nm to appear as one. Diffraction defines a best-case achievable resolution
which appears as a blurriness in the image that cannot be removed
with a “better” microscope, as first stated by Ernst
Abbe in 1873.[24] In other words, the response
function of the microscope, i.e. the image of an infinitely small
point emitter (or, in good approximation, a single molecule a few
nm in size), appears as a blurred spot of about 200 nm in size. This
image is called the point-spread-function (PSF) of the microscope,
shown in Figure A
(with camera pixelation). However, it is possible to circumvent this
“resolution limit” with related techniques which are
called “super-resolution microscopy” methods. Two fundamental
types exist: Targeted and stochastic schemes. Broadly, targeted schemes
rely on optical control of the emission wavelength with shaped excitation
spots to shrink the effective PSF of the microscope. Two emitters
that are separated less than 200 nm are thus spatially differentiated.
A prominent example is STED (stimulated emission depletion) microscopy,[25] which has been reviewed elsewhere.[18,26] Another approach is structured illumination microscopy, which also
does not require single molecules.[27,28]
Figure 1
Key ideas of
single-molecule active control microscopy (SMACM).
(A) The camera image of a point emitter or, in good approximation,
of a single molecule (point-spread-function, PSF) showing pixelation
from a camera detector. (B) Determination of the emitter location
by fitting with a model function (here, a 2D Gaussian). (C) Result
of the localization procedure. The uncertainty in the position estimate
is much smaller than the original width of the PSF. (D) Switching
of a single molecule between an emissive state (orange) and a dark
state (gray). (E) A simulated structure consisting of concentric rings
with different diameter (gray lines) and fluorescently labeled with
the orange emitters. The widest ring can just barely be resolved whereas
the three smaller rings are not resolved. (F) Active control of emission,
yielding a sparse subset of emitting molecules in each single camera
frame, and sequential acquisition of many emitter subsets. (G) Super-resolved
reconstruction.
Key ideas of
single-molecule active control microscopy (SMACM).
(A) The camera image of a point emitter or, in good approximation,
of a single molecule (point-spread-function, PSF) showing pixelation
from a camera detector. (B) Determination of the emitter location
by fitting with a model function (here, a 2D Gaussian). (C) Result
of the localization procedure. The uncertainty in the position estimate
is much smaller than the original width of the PSF. (D) Switching
of a single molecule between an emissive state (orange) and a dark
state (gray). (E) A simulated structure consisting of concentric rings
with different diameter (gray lines) and fluorescently labeled with
the orange emitters. The widest ring can just barely be resolved whereas
the three smaller rings are not resolved. (F) Active control of emission,
yielding a sparse subset of emitting molecules in each single camera
frame, and sequential acquisition of many emitter subsets. (G) Super-resolved
reconstruction.Single-molecule based methods
are the focus of stochastic schemes
and of this review. Employing them to surpass the diffraction limit
relies on two key principles. The first is the precise localization
of a single emitter by approximating the recorded camera profile of
the PSF with a suitable function. For a standard microscope, a two-dimensional
(2D) Gaussian is a reasonable choice as shown in Figure B, as long as the emitting
dipole is freely rotating.[29] The position
of the center of the 2D Gaussian is an estimate for the position of
the molecule.[30,31] The uncertainty in the position
estimate scales inversely with the square root of the photons detected
in the image, which leads to localization uncertainties much smaller
than the original width of the PSF (Figure C).[32,33] One could compare this
to a mountain with gently sloped mountainsides: Even though the mountain
(Figure A) extends
over a wide area, the location of its summit is well-defined. However,
precise localization only works when a single emitter
is investigated, and the goal here is to show how to deal with the
case when many emitters are close together. The solution is to use
an on–off active control mechanism, typically optical or chemical,
which ensures that only a sparse subset of molecules is emissive in
each frame (often called “blinking” of fluorophores, Figure D), yielding well-separated
PSFs in each frame.[34] Without a scheme
to control the emitting concentration, all fluorophores emit at the
same time, giving rise to the standard diffraction-limited image where
localization of individual PSFs is impossible (Figure E). With active control over the emitter
state, in contrast, the different single molecules labeling the target
structure can be recorded over time with a movie of the emitted fluorescence
(Figure F). The position
of each single molecule is localized, and the final, super-resolved
image is reconstructed from the records of the positions of the single
molecules (Figure G). Thus, two molecules separated less than 200 nm are temporally
differentiated. Note that this approach is in principle the same for
more complex flavors of super-resolution microscopy which allow e.g.
for 3D imaging. However, the model function and the fitting process
are more elaborate in these cases.[35,36]This
method for achieving super-resolution detail with single molecules
first appeared in 2006, termed (f-)PALM[37,38] for “Photo-Activated
Localization Microscopy”, based on photoactivation, or STORM[39] for “Stochastic Optical Reconstruction
Microscopy”, based on forcing molecules to blink by repeatedly
entering and returning from dark states. Very quickly a large number
of additional active control mechanisms appeared, producing a flood
of new acronyms. Since all these methods rely on single-molecule images
and an experimentally chosen active control mechanism, we will use
the term SMACM, short for “Single-Molecule Active Control Microscopy”,
for all of them. Other researchers use the term SMLM (for “Single-Molecule
Localization Microscopy”) for this area, but SMLM is a broader
term which also applies to single-molecule tracking experiments at
extremely low concentrations where emitter overlap is not an issue,[40−49] which have had a much longer history than super-resolution microscopy.
Experimental Considerations
Experimentally,
there are three main “components” to be considered for
SMACM (Figure ):
Figure 2
Three key experimental components required of all SMACM
experiments.
(A) Sample preparation and labeling. (B) Optical setup, from fluorophore
excitation to detection. (C) Data analysis, i.e. identification and
localization of single-molecule signals for image reconstruction.
The sample. This
component prominently
contains the labeling of the target structure with the desired fluorophore
via an appropriate labeling strategy, and the overall sample preparation
procedure. Notably, the sample preparation must be as clean as possible:
Due to the high sensitivity required for the detection of single molecules,
fluorescent contaminants or unspecific binding of the labeling agent
to nontarget structures is much more problematic than for conventional
light microscopy. If blinking of fluorophores in and out of dark states
is induced by chemical means, the sample preparation also contains
the choice of blinking buffer composition.The optical setup. This component covers
all the parts of the microscope used to excite the fluorophores, to
filter out the excitation light from the emitted fluorescence, to
manipulate the fluorescence to gain more insight about the studied
system (e.g., 3D information), and to detect the fluorescence. In
fact, the optical setup can be very simple—a minimal design
requires not more than an excitation source (typically a laser, but
a lamp can work, too), some lenses, mirrors, and filters, an objective,
a sample holder, and a sensitive camera. Of course, elaborate designs
for more demanding tasks can become quite complicated.Data analysis: While sample preparation
and optical setups are required for all flavors of optical microcopy,
data analysis is of specific importance in single-molecule microscopy,
and therefore for SMACM. This is because the final image is not directly
recorded by the camera—it is reconstructed from the information
obtained from the single-molecule signals. Thus, special attention
must be paid to the analysis algorithms. A data set of the highest
quality can lead to inferior results if analyzed carelessly.Three key experimental components required of all SMACM
experiments.
(A) Sample preparation and labeling. (B) Optical setup, from fluorophore
excitation to detection. (C) Data analysis, i.e. identification and
localization of single-molecule signals for image reconstruction.We will now briefly present a few examples of the
current state
of the art in the application of SMACM in cell biology. Then, we will
discuss perspectives related to these major experimental components.
State of the Art in Biological Applications
SMACM is currently mostly applied in biological contexts, especially
eukaryotic and bacterial cell biology. As described above, this is
mainly due to two reasons: First, the length scales of interest match
the achievable resolution of few to few tens of nanometers well, and
second, protocols to specifically attach a fluorescent label to the
target have been investigated for many decades. Already the first
implementations of SMACM at room temperature demonstrated the imaging
modality in cells.[37−39] Thus, we will give a general overview of prominent
uses of the method in cell biology. Figure shows several examples of the fascinating
structures which can now be observed using SMACM methods (in three
dimensions in fact, discussed in Section ).
Figure 3
Selected examples of 3D super-resolution images
using single-molecule
active control microscopy (SMACM) in cells. (A) Super-resolution image
of actin in a COS7 cell revealing the dual-layer organization, where
labeling has been done by an Affimer reagent.[50] A site-specifically attached DNA strand has been added to the protein
via a Cys mutation and then in the fixed cell, reaction with maleimide-DBCO
is performed, and an azide-terminated target DNA strand has been covalently
added. Then fluorescently labeled DNA-capture strands are added, and
the active control mechanism is binding and unbinding of singles (DNA-PAINT).
(B) 3D super-resolution images of synaptonemal complexes in whole
mouse spermatocytes, imaged by immunolabeling synaptonemal complex
protein 3 (SYCP3) with a blinking fluorescent label.[51] This image required 21 optical sections through six cycles,
covering nearly 9 μm in the cell nucleus, and shows the twisting
band of the paired lateral elements; the resulting organization of
the chromosomes is clear. (C) 3D super-resolution images of tiny membrane
tubules on the surface of a cancer cell, where the sialic acids have
been covalently labeled with a blinking dye.[52] Here the cell has been imaged with the TILT3D microscope,[53] and individual hollow tubules are shown in the
insets.
Selected examples of 3D super-resolution images
using single-molecule
active control microscopy (SMACM) in cells. (A) Super-resolution image
of actin in a COS7 cell revealing the dual-layer organization, where
labeling has been done by an Affimer reagent.[50] A site-specifically attached DNA strand has been added to the protein
via a Cys mutation and then in the fixed cell, reaction with maleimide-DBCO
is performed, and an azide-terminated target DNA strand has been covalently
added. Then fluorescently labeled DNA-capture strands are added, and
the active control mechanism is binding and unbinding of singles (DNA-PAINT).
(B) 3D super-resolution images of synaptonemal complexes in whole
mouse spermatocytes, imaged by immunolabeling synaptonemal complex
protein 3 (SYCP3) with a blinking fluorescent label.[51] This image required 21 optical sections through six cycles,
covering nearly 9 μm in the cell nucleus, and shows the twisting
band of the paired lateral elements; the resulting organization of
the chromosomes is clear. (C) 3D super-resolution images of tiny membrane
tubules on the surface of a cancer cell, where the sialic acids have
been covalently labeled with a blinking dye.[52] Here the cell has been imaged with the TILT3D microscope,[53] and individual hollow tubules are shown in the
insets.Probably the most frequently imaged
cellular structures are microtubules
or other cytoskeletal elements such as actin meshwork (Figure A). This is not only because
fascinating insights can be extracted, e.g. about the architecture
of the whole network,[54] but also because
microtubules in particular are a popular target to test the resolution:
The true width of microtubules is known from electron microscopy and
the only increase beyond this arises from the size of the labeling
agents (e.g., primary and secondary antibodies). Moreover, their linear
appearance makes them easy to identify. Thus, the difference between
measured and expected dimensions is a metric for the resolution capability
of the system. It should be noted, however, that one must be careful
not to overestimate the conclusions drawn from such measurements:
Microtubules are one-dimensional; therefore, more complex or nonstructured
cellular components might exhibit artifacts that are not easily detectable
when looking at microtubules alone. Moreover, aberrations at the glass–water
interface are not too prominent when imaging microtubules close to
the coverslip, yet they may become a major factor when moving deeper
into the cell.[55,56]SMACM unfolds its full
potential when structures exhibiting nanoscale
detail within a complex environment that are specifically targetable
are investigated. An excellent example are studies of the nucleus
and of the nuclear membrane. Here, SMACM has contributed significantly
to our understanding of this vital, incredibly dense, and complicated
organelle. It was predicted for a long time that the nucleus, despite
its seemingly chaotic nature, exhibits some organization. Super-resolution
microscopy, together with orthogonal techniques, provided evidence
for this hypothesis, demonstrating order and regulation by directly
visualizing nuclear structures.[57] In recent
work, much has been learned about chromatin and nucleus architecture,[58,59] chromatin decompaction via Myc,[60] the
nuclear lamina,[51,53] reorganization of the nucleus
upon shifts in cellular state,[61,62] and structures of the
DNA-organizing synaptonemal complexes[51] (Figure B) to name
just a few.In fact, in recent years, the nuclear pore complex
has been developed
into something of a standard for super-resolution microscopy.[63] Compared to microtubules, they have the advantage
of having more structural features (such as domain symmetry, precisely
known distances, and 3D architecture). Furthermore, they are typically
located deeper in the cell, i.e., farther away from the coverslip,
which allows for more realistic aberration correction.Another
cellular organelle that features nanoscale details within
an environment of similar complexity as the nucleus is the mammalian
glycocalyx. The glycocalyx is a meshwork of myriads of different sugar
structures surrounding every cell in the body.[64] These sugar structures either are attached to proteins
or lipids or are free. By combining SMACM with highly specific, bioorthogonal
labeling strategies,[65,66] it was possible to gain insights
into nanoscale glycocalyx architecture and reorganization upon changes
of the cellular state in recent years (see Figure C).[52]Single-molecule
super-resolution microscopy plays a critical role
in the investigation of the cell membrane and of events at the cell
membrane.[67,68] Again, many relevant processes occur at
length scales of few nanometers, which is ideally matched by SMACM.
Often, labeling extracellular targets tends to be cleaner, i.e. causing
less unspecific binding, than labeling intracellular targets where
the cell membrane and the cytosol must be passed, which is an important
factor to take into account. Moreover, the combination of localization
with total internal reflection fluorescence (TIRF) microscopy is an
especially effective approach: TIRF microscopy generates only ∼200
nm spatial imaging depth, and more is not required to study the basal
cell membrane, so any background fluorescence from the cell is easily
suppressed, further improving the imaging quality. Consequently, the
power of SMACM has been used to investigate many areas of membrane
biology—from membrane protein distributions[69,70] to the immunosynapse and membrane domains,[71−74] to clathrin- and caveolin-mediated
endocytosis.[75−77]Another organelle at the cell membrane that
has been extensively
investigated by our lab and others is the primary cilium, a densely
populated tube-like protrusion with a diameter of a few hundreds of
nanometers and length of a few microns which acts as a vital signaling
hub. In recent decades, researchers learned that the primary cilium
is a key regulator of cellular and organismic fate.[78] These discoveries have been enhanced and extended by single-molecule
localization microscopy as well as other advanced microscopy approaches.[79−83]Finally, single-cell organisms such as yeast or bacteria have
also
been studied extensively with SMACM.[16,84−86] Their total size is typically at or below the diffraction limit;
therefore, any fine structure is lost when imaged with conventional
optical microscopy. Super-resolution methods have enabled the investigation
of key bacterial proteins,[87] DNA organization
and transcription,[88−90] or bacterial membrane organization.[91,92] Similarly, yeast biology could be investigated at unprecedented
length scales.[93−95] When SMACM is combined with single-molecule tracking,
further insights have been obtained about bacterial microdomains and
their roles in regulation.
Perspectives on Methods and
Applications
In this section, we will discuss innovations
required to overcome
some current limitations of SMACM in order to unleash its full potential.
We will follow the outline presented in section , discussing each of the three major experimental
components separately.
Sample Preparation
Dyes
The quality of the dye used for SMACM chiefly
determines the quality of the final reconstruction. While SMACM has
been demonstrated with a range of different dyes,[96] only a smaller subset of fluorophores exhibit the required
quality (mainly high photostability, good blinking statistics, and
high quantum yield) to generate reconstructions with exceptional resolution.
Very popular are AlexaFluor647, CF568, and the JF dyes;[97,98] additionally, there are a few more dyes in the infrared spectrum.[99] Fluorescent proteins such as eYFP are widely
used as well and have improved considerably over the past years extending
to new classes of photoactivatable fluorescent proteins,[100−104] but they still perform inferior to organic dyes in total photons
emitted before photobleaching.[105] Advances
in this area are highly desirable. Although currently available dyes
and fluorescence proteins allow for a range of experiments, all of
them would immediately benefit from improved properties and an expanded
color spectrum. The goal in this area should be to ensure that the
choice of dyes is not the deciding factor for the obtained resolution.
Induction of Active Control Process
The active control
of emission is vital for all flavors of SMACM. One broad strategy
is photoactivation, where a dye is converted from a dark fluorogen
to an emissive form by a secondary control beam,[106] but photoinduced blinking is more common.[107,108] In a simple physical picture, a molecule’s emission can blink
due to reversible entry into a dark state followed by subsequent return
to the ground state to emit again. While the underlying physical mechanisms
for blinking by reversible photoreduction have been teased out via
cleverly designed experiments for several dyes and mechanisms have
been explored in proteins,[109−112] we often do not have an in-depth understanding
of all processes that cause a fluorescent dye or protein to blink.
Consequently, only a fraction of available dyes can be used for SMACM
at the moment. In principle, it should be possible to induce blinking
for all types of dyes, provided that the underlying molecular properties
and accessible states are precisely known. Therefore, studies aimed
at a deeper understanding of photophysics and photochemistry via e.g.
transient absorption spectroscopy should be expanded. The generated
insights will yield central benefits: First, existing dyes will be
made compatible with SMACM, and second, the synthesis of novel dye
structures will be informed, tailoring the molecular properties according
to the photophysical and environmental sensing requirements.A highly relevant area regarding induction of active control is cryogenic
imaging. At low temperatures, bleaching of fluorophores is significantly
reduced, which greatly increases the number of photons detected per
dye molecule. As the localization precision scales inversely with
the square root of the detected photons, this allows for localization
precisions below 1 nm. However, only a few active control options
exist currently in this regime as most high-quality dyes and fluorescent
proteins stop blinking at cryogenic temperatures.[103,113,114] A deeper understanding of dye
photophysics and photochemistry is certainly needed and will enable
the development of tailored dyes for cryogenic localization microscopy,
which would significantly expand the scope of single-molecule experiments.An orthogonal approach to direct induction of blinking is PAINT
(short for “point accumulation for imaging in nanoscale topography”).[115−120] Here, a diffusing fluorescent molecule binds to the target structure,
concentrating the emitted photons at the binding site, which leads
to a bright burst of fluorescence. This burst can be further increased
by temporary unquenching of the fluorophore.[121] Thus, blinking is not introduced photophysically, but via transient
binding and unbinding events, which uncouples the blinking statistics
from the dye properties. A powerful implementation of this idea uses
DNA strands to create binding sites at the target structures and complementary
sequences tagged with fluorophores.[122] This
leads to a fundamental trade-off between imaging speed and interaction
specificity: Shorter DNA strands yield faster dynamics, but also more
unspecific binding. Certainly, innovation in this area will allow
for faster imaging without losing specificity, enabling a versatile
alternative for direct induction of blinking.Precise control
over blinking is highly desirable as blinking statistics
are one of the key determinants for the achievable resolution, and
some control, especially improved counting has been achieved by tailoring
of the optical pulse.[123] This is because
each precisely localized single molecule acts as a reporter for the
imaged structure. Therefore, the labels must sample the fine details
of the structure densely to make them resolvable. The basis for this
requirement is the famous Nyquist–Shannon Sampling Criterion.[124−126] Originally, this concept states that a signal in time, composed
of many frequencies, must be sampled at least twice as fast as the
highest frequency component in order to be reconstructed from the
samples. This criterion can be directly translated to images and spatial
frequencies. As shown in Figure A, a test pattern with low spatial frequency can be
reconstructed from both low and high numbers of localizations. In
contrast, for the high spatial frequency case, low numbers of localizations
fail. This is also clearly visible from the respective power spectra
(Figure B). Thus,
even if the structure is densely labeled, it will be poorly resolved
if the blinking statistics are suboptimal (cf. also Figure and the related discussion).
Figure 4
Importance
of sampling the imaged structure. (A) Test patterns
consisting of five (left columns) or 20 (right column) stripes, sampled
with low (top row) or high numbers (bottom row) of localizations.
(B) Corresponding 2D power spectra, radially averaged. PSD = Power
spectral density.
Figure 5
High localization precision,
small labeling footprint, and high
labeling efficiency are equally important for high-quality reconstructions.
Shown are simulated patterns of 4 × 4 infinitely small spots
with a spacing of 25 nm. Each spot contains 10 binding sites. For
each combination of localizations precision and labeling footprint,
0.2, 0.5, and 0.8 average labeling efficiencies are simulated (standard
deviation, SD, 0.1 between spots). Each labeled binding site yields
on average 10 localizations (SD: 5 localizations). The localizations
are binned into 2D histograms with 2 nm bin width, and the edge length
for each reconstruction is 125 nm. The color bar ranges from blue
through green to yellow. Blue always corresponds to 0 localizations,
and yellow to 2 localizations for poor reconstructions and up to 12
localizations for good reconstructions, respectively.
Importance
of sampling the imaged structure. (A) Test patterns
consisting of five (left columns) or 20 (right column) stripes, sampled
with low (top row) or high numbers (bottom row) of localizations.
(B) Corresponding 2D power spectra, radially averaged. PSD = Power
spectral density.High localization precision,
small labeling footprint, and high
labeling efficiency are equally important for high-quality reconstructions.
Shown are simulated patterns of 4 × 4 infinitely small spots
with a spacing of 25 nm. Each spot contains 10 binding sites. For
each combination of localizations precision and labeling footprint,
0.2, 0.5, and 0.8 average labeling efficiencies are simulated (standard
deviation, SD, 0.1 between spots). Each labeled binding site yields
on average 10 localizations (SD: 5 localizations). The localizations
are binned into 2D histograms with 2 nm bin width, and the edge length
for each reconstruction is 125 nm. The color bar ranges from blue
through green to yellow. Blue always corresponds to 0 localizations,
and yellow to 2 localizations for poor reconstructions and up to 12
localizations for good reconstructions, respectively.
Labeling Methods
Much effort has been and is currently
devoted to the development of advanced labeling protocols, for example
via incorporation of unnatural biomolecules that carry a bioorthogonal
reporter for subsequent targeting, artificial amino acids, or, most
recently, CRISPR-based approaches.[127−132] However, there is a lot of room for improvement. Many experiments
in the field still rely on standard immunochemistry protocols. As
the achievable localization precision is nowadays routinely around
10 nm or below, this is becoming a significant problem. The size of
a single antibody is approximately 10 nm and thus no longer negligible.
As bound antibodies are rotationally not completely flexible, their
labeling footprint is smaller than their full size, but still at least
several nanometers. Moreover, when target epitopes are spaced apart
just a few nanometers (which is, considering the crowded environment
of the cell, often the case), steric hindrance can lead to low labeling
efficiencies.[133] Both effects render localization
precision of few nanometers essentially useless.Thus, beside
the needs for low unspecific binding and overall robustness, the ideal
labeling protocol must fulfill two additional key requirements: small
labeling footprint and high labeling efficiency. To illustrate this,
we simulated a test pattern and consider different combinations of
localization precision, labeling footprint, and labeling efficiency
(Figure ). Evidently,
high-quality reconstructions mandate optimization of all three parameters
(provided good blinking statistics, as discussed above). Among them,
we think that labeling efficiency is the one currently least in focus.
The simulation highlights that this should change. Obviously, for
low labeling efficiencies and otherwise excellent parameters, the
test pattern is simply incomplete. While this is already problematic,
the situation is more serious for cases where the other parameters
are also not ideal. For example, consider the case of 2 nm localization
precision and 8 nm labeling footprint: For a labeling efficiency of
0.8, the test pattern is quite fuzzy, but visible, and it blurs for
lower labeling efficiencies.These simulations exemplify that
the field should move away from
immunostaining when possible and pay close attention to the details
of the labeling process. One useful approach with good results uses
enzymatic attachment of fluorophore-functionalized tags.[134,135] SMACM is an ideally suited method to disentangle the many overlapping
interactions of biomolecules that take place on small length scales
within the cell. However, to unfold its full potential, proper labeling
with rigorous control experiments is as important as tailored optical
setups and dyes.
Tissue Sections, Organoids, and Whole Organisms
Imaging
cultured cells is itself not always easy, but all experimental steps
become even harder when moving toward investigation of tissue sections,
organoids, or whole organisms: Mere handling and mounting the sample
is more difficult. Also, labeling of the target structure is more
challenging—unspecific binding usually strongly increases,
and new problems arise, such as penetration of the labeling agent
through many layers of cells as well as the presence of elevated autofluorescence.
Finally, on the optical side, imaging deep in tissue gives rise to
significant aberrations, requiring sophisticated correction schemes.
All these factors and many more make detection and precise localization
of single molecules considerably more difficult.One might wonder—why
try to address these difficulties? The answer is simple: While many
insights into biological processes can be drawn from experiments in
cultured cells, the true environment of an individual cell is the
3D tissue matrix of other cells and various extracellular components.
Thus, a significant portion of the picture is lost when cells are
investigated in culture. It is therefore desirable to move away from
cultured cells and toward more realistic settings when possible. It
would be too much to ask to directly switch from SMACM in cultured
cells to e.g. live Zebrafish embryos, so the shift should be gradual.
Fortunately, a wide space exists between artificial and physiological
conditions. For example, instead of culturing cells on glass coverslips,
an alternative is cell culture on less stiff substrates, similar to
the mechanical properties encountered in tissue. Simply the removal
of glass as substrate shifts the cell into a more physiological state.[136] Co-culturing of different cell lines can help
to emulate the complex tissue environment better.[137] Cell spheroids and organoids are more difficult to create
and maintain, but still manageable, and they are considerably closer
to the physiological case.[138] Finally,
of course, the ultimate level of realism lies in tissue sections and
live animals, and certainly they are experimentally most challenging
to handle.[139−141] A moderate increase in complexity of the
studied system can pay off a great deal, and it is desirable that
these considerations become a standard part of experimental planning.
Optical Setup
Setup Designs
All research thrusts discussed in this
section are in one way or another linked to innovation in the design
of the microscopes used, enabling improved localization precision,
investigation of multiple fluorescent species, or recording of other
parameters than fluorescence brightness alone. This is because implementation
of such improvements mandates various enhancements on the instrumental
side: More powerful lasers allow for more efficient shelving and rapid
blinking of single molecules; faster cameras enable higher frame rates
and thus shorter acquisition times; creative designs for excitation
and detection geometries (e.g., lattice light sheet, TILT3D, 4Pi,
etc.)[51,53,139,142−144] minimize unwanted out-of-focus
fluorescence and maximize the number of detected photons; and automated
acquisition makes super-resolution high-throughput imaging possible.
Progress on the instrumental side is likely to continue, thus opening
windows to either currently not implementable excitation and detection
schemes or simplified, more robust layouts.Happily, it does
not seem necessary to desperately enforce innovation in setup design
as the impetus will naturally follow from the research questions addressed.
However, microscopes employed for state-of-the-art detection of single
molecules are almost exclusively noncommercial. Desiderata in this
area are more user-friendly documentations of setups. Today, setups
are usually described briefly in the materials and methods sections
of publications, and schematic drawings are standard. However, thanks
to the broad availability of computer-aided design software, detailed
blueprints and 3D animations are straightforward to create, which
give a much more detailed picture and allow for easier reproduction.
More common utilization of such resources would significantly facilitate
the implementation of successful setup designs.
Multicolor
Imaging
Much can be learned from imaging
a single cellular structure, but eventually, almost all biological
questions can be traced to interactions of all the different cellular
entities like proteins, DNA, organelles, and so on (sometimes called
“interactome”).[145,146] Therefore, two- or
multicolor super-resolution studies do not just add another imaged
species; rather, they are essential for investigation of fundamentally
new types of questions.As always, the added information content
comes at a cost. For example, registration of different color channels
at the required precisions is far from trivial. Resolving two structures
at the length scale of a few nanometers is futile when the registration
error between the channels is tens of nanometers. Thus, registration
typically requires more than just affine transformations—local
variations induced by field-dependent and chromatic aberrations must
be corrected as well.[147] This process is
complex in two dimensions and can become a major challenge in three
dimensions. Another key problem is the choice of dyes: As discussed
above, not too many dyes exist that can be effectively employed in
single-molecule super resolution microscopy, although some progress
has been made based on decomposition of spectra.[148] When used simultaneously, it is highly advantageous when
all dyes require identical experimental conditions, e.g. the same
blinking buffer.[149] Changing parameters
between the acquisitions of different channels is possible, but it
increases the complexity of the registration further. Progress especially
for precise registration and high-quality, compatible dyes is therefore
essential to further improve multicolor SMACM.
3D Imaging
Similar to the transition from one to two
or more imaged species, the transition from knowledge of the 2D to
the 3D position of a localized emitter significantly increases the
scope of the experiment. Ultimately, cells are 3D objects, and thus,
their investigation requires 3D information.[150] Therefore, shortly after the first description of SMACM in 2D, methods
had been developed to include the axial dimension,[151,152] based on the well-known ability of cylindrical lenses to provide
3D localization.[153] Since then, the field
has progressed considerably. Various engineered PSFs, tailored to
the desired axial range, a small camera footprint, and others, have
been developed.[154−158] Thanks to these efforts, axial localization precisions of 10–20
nm can be easily reached. This should motivate the field to push the
limits of 3D imaging toward routinely achieving single-digit nanometer
precisions. Provided efficient labeling with small footprints, such
an improvement can make the difference between resolving individual
proteins in 3D or not.To achieve this goal, effects that were
so far not in the focus of research must be investigated. Most urgent
is a better understanding of the various consequences of index mismatch
between the cellular environment (with an index of refraction around
1.33) and the optical elements (with an index of refraction around
1.52). Index mismatch effects are noticeable in 2D imaging as well,
but in 3D, they are considerably more problematic. A key example is
focal shift, i.e. the disagreement between nominal focus position
and true focus. It had been characterized early on, but so far mostly
for emitter positions dozens of microns away from the interface.[56,159] Our own lab has recently shown that, at the length scales relevant
for the analysis of single cells, the focal shift varies quickly and
needs to be accurately characterized to yield e.g. correct 3D distance
measurements.[55]Generally, 3D imaging
is an excellent example of the interconnected
nature of various requirements encountered in SMACM: Dyes with tailored
photophysical properties and precisely tunable blinking statistics
are needed to maximize the emitted photons from each emitter and the
information content per frame. Clever setup designs make it possible
to reduce aberrations and detect complex PSFs reliably. Finally, a
detailed understanding of the image formation and optimized algorithms
for rapid and accurate localization are necessary for final image
reconstruction.
Aberration Correction
As mentioned
above, the shape
of the PSF in a standard microscope is, in good approximation, a 2D
Gaussian. If engineered PSFs are used for e.g. 3D imaging, the theoretical
shape of the PSF is more complex, but can still be calculated precisely.
However, the theoretical PSF always departs from the experimental
PSF, which is caused by imperfections in the various optical elements
of the microscope. These aberrations can be minimized by careful calibration
and high-quality optics, but not completely ruled out. This is especially
the case for aberrations induced by the sample itself: If the fluorescence
detected from a single molecule passes through a whole cell or even
a layer of cells, the experimental PSF will differ significantly from
the theoretical one due to the many changes in the index of refraction
that the emitted wavefront encounters. If this path length is small,
the induced aberrations are less of a problem, but in this case, the
investigation is limited to areas close to the coverslip. Methods
have been described to retrieve the PSF in situ and
should be considered.[160] In any case, disagreement
between theoretical and experimental PSFs cause poor fitting results
in e.g. PSF localization, which deteriorates the quality of the reconstruction.Luckily, it is possible to correct for the aberrations in an experiment.
The underlying idea is simple: As the various sources of aberrations
act together to deform the wavefront that is detected by the camera,
giving rise to a distorted PSF, one just needs to install a device
that precisely counteracts the induced distortions, such as deformable
mirrors (DMs) or spatial light modulators (SLMs) which alter the phase
of the detected light. As these devices can adjust to various situations,
they are called “adaptive optics”, which is often used
synonymously with the whole process of aberration correction. The
experimental implementation is a little more challenging than the
underlying idea, but not too cumbersome. The key problem to be solved
is the measurement of the aberrated wavefront in order to decide how
to balance the aberrations. Many strategies exist, but probably the
most straightforward method is image-based correction: Here, PSF images
obtained from point-like emitters such as fluorescent beads are directly
analyzed at the start of the experiment, and the adaptive optics of
the system are changed in a feedback loop until the experimental PSF
is as close as possible to the unaberrated, theoretical PSF.[161−165]Using adaptive optics for aberration correction has been demonstrated
to significantly enhance the quality of super-resolution reconstructions
or to even make reconstructions possible in the first place. Still,
they are currently mainly implemented in specialized setups for deep-tissue
imaging or related experimental questions. More general dissemination
of aberration correction approaches would nevertheless be beneficial,
as it is a relatively easy to implement a method to significantly
improve the quality of the recorded data set.On a more fundamental
level, aberration correction highlights the
benefit of learning from fields that face conceptually similar problems
as single-molecule studies. In the case of aberration correction,
much inspiration has been taken from astronomy, and also for efficient
photon detection, cameras from astronomy have driven advances in single-molecule
imaging.[166] Just as for single-molecule
studies, astronomy faces the challenge to optically detect dim light
from a point-like object while background light and aberrations are
present. In astronomy, however, aberration correction is mandatory
for earth-bound observatories as atmospheric dynamics severely distort
the incoming wavefront. For this reason, the problem was studied early,
and many clever strategies have been put forward. Such inspirations
from neighboring fields are extremely advantageous for single-molecule
studies and should be actively sought after.We have now discussed
roughly half of the emerging perspectives
for future improvements of SMACM. Here, we would like to pause for
a moment and put the issues discussed into a broader context. Ultimately,
we think that innovations in the three main experimental components
of SMACM and their respective subcategories serve one key purpose:
To improve and facilitate current applications of SMACM on the one
hand, and to develop novel applications of SMACM in fields that currently
do not prominently employ single-molecule approaches on the other
hand. These innovations cooperate as schematically shown in Figure .
Figure 6
Cooperation between advances
in the three key components of SMACM,
allowing for improvements in current applications and development
of novel applications.
Cooperation between advances
in the three key components of SMACM,
allowing for improvements in current applications and development
of novel applications.
Correlation of Single-Molecule
and Electron Microscopy
Biological imaging is dominated by
two methods: Light and electron
microscopy. While the latter yields higher resolution, it suffers
from poor contrast information (specificity) as the final image is,
independent of the modality, usually a grayscale representation of
electron density alone. Thus, it is a logical approach to combine
light and electron microscopy: Information on the identity of the
imaged species can be obtained via highly specific fluorescent labeling
and light microscopy, whereas exquisite spatial information and cellular
context are achieved via electron microscopy. However, there is a
price to pay in terms of experimental complexity—electron microscopy
and light microscopy differ considerably in sample preparation and
imaging conditions; thus, protocol optimization is crucial. Such approaches
are termed correlative light and electron microscopy (CLEM), and they
have been used to investigate various biological systems with reasonable
success within the constraints of the diffraction limit.[167−172]In recent years, increased attention has been focused toward
the correlation of SMACM with electron microscopy. The main motivation
is that information on the imaged species is diffraction-limited if
standard optical microscopy is employed, which is problematic because
there is a huge mismatch between the resolution of the electron microscopy
and DL optical imaging when dense biological systems are investigated.
Happily, imaging single molecules at cryogenic temperatures improves
localization precision as discussed above. The experimental demands
for correlation of electron and localization microscopy are unfortunately
even higher than those for CLEM; however, if successfully implemented,
the information content of the obtained data is considerably increased
as demonstrated by our group and others. To make the various correlative
approaches more accessible, innovation on all fronts is required:
Improved dyes for better blinking or activation capabilities at low
temperatures, robust sample handling for smooth transition between
microscopes, precise registration schemes for overlaying of the two
image modalities with subnanometer precision, and many more improvements
are needed (for an in-depth discussion, please see an upcoming review
by P. D. Dahlberg and W. E. M. in Annual Review of Physical
Chemistry). In that sense, correlation of electron and localization
microscopy is an incredibly exciting area as its huge potential is
obvious while the realization of the full potential is not straightforward
at the moment. General progress in the field of single-molecule techniques
will directly influence the development and improvement of correlative
approaches.
Active Sample Stabilization
Although
the acquisition
speed for SMACM experiments has considerably increased over the past
years, the total time to record a data set is still rarely below several
minutes, except for certain tour de force experiments.[173] Even a sturdy setup design after full temperature
equilibration cannot prevent sample drift of tens of nanometers during
long time spans, which is much larger than the achievable resolution.
Thus, drift correction must be performed. At the moment, this is often
done in postprocessing, e.g. using a fluorescent bead (called a “fiducial”)
which is recorded in parallel to the single-molecule data. The fiducial’s
apparent motion during the acquisition then allows for drift correction.[174,175] Image correlation is another oft-used strategy.[176]In recent years, increased attention has been focused
toward alternative feedback-based sample stabilization approaches.
In some demonstrations, a fiducial is localized in real time; the
position estimate is used to move the stage during the experiment
to correct for drift.[164,177] Compared to drift correction
in postprocessing, such approaches feature several advantages: For
example, they often do not rely on fluorescence-based detection of
a fiducial, but rather white-light scattering-based imaging of e.g.
polystyrene beads, which yields localization precisions in 3D of few
nanometers without bleaching of the fiducial or spectral crosstalk.
Furthermore, the decoupling of single-molecule data from fiducial
data acquisition enables feedback at subframe rates, i.e. the position
of the stage can be adjusted several times during a single camera
frame for fluorescence acquisition. Such fine correction is not possible
in postprocessing.These promising strategies are enabled by
significant advances
in camera and stage design as well as data processing, which allow
for rapid recording and localization of the fiducial and equally rapid
stage movement. These advances have also enabled much easier implementation
into new or existing setups, so it is realistic to hope that active
sample stabilization will continue to evolve into a standard tool.
Multiparameter Imaging
So far, the fluorescence emitted
from a single molecule was only discussed with respect to localization:
The detected PSF is used to estimate the position of the emitter from
the pattern of detected photons on the camera. However, it is vital
to note that much more information is contained in molecular fluorescence.
Ultimately, the properties of the fluorescence emission, e.g. wavelength
or polarization, are determined by quantum phenomena that underlie
the excitation and emission processes. Moreover, external factors
such as pH or electrostatic interactions can significantly affect
the emission properties. Due to the small size of a single molecule,
it is only sensitive to influences in the immediate neighborhood.
Therefore, each single molecule acts both as a reporter of its position
in space and as a uniquely sensitive reporter of its local environment.Multiparameter imaging thus uses the additional information contained
in the emission not only to image a structure but also to draw conclusions
on local variations within that structure. For example, polarization
of the emitted fluorescence reports on the rotational mobility of
a single molecule and its orientation during emission.[178−180] Similarly, shifts in the emitted wavelength are evidence for a change
in the dye’s microenvironment. Such measurements can notably
increase the scope of a single molecule-based investigation and provide
valuable orthogonal data. Consequently, numerous studies have used
multiparameter imaging to investigate, for example, membrane architecture,
DNA intercalation, polarity, the behavior of amyloid aggregates, and
many more local properties.[181−188]The considerably increased information content, however, comes
at the cost of a more challenging experimental procedure. For example,
polarization microscopy requires additional optimization of the labeling
step as the dye must be rigidly attached to the target; otherwise,
the measurement will not capture the true orientation of the target.
Spectroscopic imaging necessitates photon budget considerations: a
fraction of the detected photons must be used for detection of the
spectrum and are thus not fully available for localization, reducing
the localization precision. Generally, multiparameter imaging requires
more advanced optical systems than pure localization microscopy. In
our opinion, however, the additional complexity is easily outweighed
by the gained information. A broader use of multiparameter imaging
is highly desirable in order to make use of the huge potential of
the approach.
Data Analysis
Neural Networks
In recent years, the advent of neural
networks and deep learning and their success in analyzing images has
revolutionized the field of microscopy in general and specifically
of SMACM. It has been demonstrated by various groups that all data-related
aspects can significantly benefit from deep learning: Previously impossible
image analysis methods can be performed, or existing analysis methods
significantly accelerated. Thus, deep learning has been successfully
implemented for detection and localization of PSFs,[189−191] for phase retrieval and background correction,[192−195] for phase mask design of optimized PSFs,[157,196] and for compressed sensing,[197,198] to just name a few
applications, and work in this area has been recently reviewed.[199]Therefore, convincing anyone that deep
learning offers intriguing potential for the field is no longer necessary.
Instead, careful calibration of deep learning as a powerful tool is
now required. In the wake of the initial general enthusiasm, it should
not be overlooked that deep learning, just as any other analysis method,
is not able to “create information”; more seriously,
it must not be allowed to do so. If, for example, a neural network-based
approach adds fine details to an initial super-resolution reconstruction,
it infers the existence of these details from the data it was presented
during training. The seemingly created information is just the assumption
of a general law of the appearance of the imaged structure. Thus,
without proper controls, deep learning-based analysis can be especially
prone to creation of artifacts: For example, a rare structural feature
that was not present in the training data set will likely be skewed
or missed completely, both of which are problematic.[199]As a result, there are two key tasks for deep learning
in the single-molecule
field: On the one hand, the method itself should be standardized and
made more generally accessible, e.g. by creating user-friendly software
packages.[200,201] This will allow laboratories
that are not actively developing deep learning-based analysis methods
to benefit from the substantial capacity of the method. On the other
hand, however, this process of dissemination should not present deep
learning as a black box. Users should always be informed upfront about
potential problems of the method, and rigorous, orthogonal controls
should be mandatory.
Effective Computing
Data processing
has become a vital
step in SMACM. Several gigabytes of imaging data are typical for less
sophisticated systems, and hundreds of gigabytes or even terabytes
can easily be acquired in more complex scenarios. There are two main
reasons for these exceptional numbers. First, as described above,
the imaged structure must be spatially sampled in detail to achieve
the highest resolution, necessitating extended acquisition times and
large numbers of recorded camera frames. Second, the requirement for
sparse single-molecule signals demands that a considerable part of
each recorded camera frame is empty (which again highlights that the
precise control of blinking is highly desirable as it allows for maximal
information content per camera frame).Currently, these huge
amounts of data are broadly met with imperfect tools. Frequent copying
of data from one location to another is slow. Real-time streaming
of data to a common storage location would be much more efficient.
Programming languages and apps commonly used for evaluation and visualization,
such as Matlab or ImageJ, are not designed to handle dozens of gigabytes
of data—Python or C could often be superior. Using more or
less standard desktop computers for data processing is not ideal either.
Many tasks encountered in data evaluation can be parallelized, making
GPU computing the method of choice.[202−204]Of course, various
strategies have been demonstrated to implement
all these aspects, showing that the time span between data generation
and final result can be considerably reduced.[205] This is highly beneficial, and not only to save time: If
the data can be analyzed in or close to real time, the experimental
parameters can be tuned flexibly, improving the quality of the generated
data and the science learned, and throughput can be significantly
improved.To make progress toward this goal, a different attitude
toward
data analysis is required that exceeds mere problem solving. Evaluation
of single-molecule data should be viewed as a separate data science
and estimation problem and treated accordingly. This would also mean
a change in education. Although GPU or cloud computing is becoming
more and more user-friendly, it is far from being trivial. Thus, education
in single-molecule techniques should include a prominent section on
data science, be it efficient storage, parallelization techniques,
or its various other aspects.
Benchmarking
When
confronted with an experimental result,
any researcher will ask: Can I trust it? SMACM is no exception. Quite
the contrary, accurate benchmarking is vital in the field: First,
the length scales of interest are so small that tiny inaccuracies
can substantially skew the result. Second, the required exquisite
sensitivity bears the risk of artifacts—e.g. due to nonspecific
binding, overcounting (repeated detection of the same molecule emitting
over several camera frames, causing artifactual clusters with sizes
of approximately the localization precision), or the detection of
fluorescent “dirt”. It is always essential to be sure
that the labeling itself is not affecting the visualized structure,
and this can occur more often as resolution improves.[206] Finally, from a more practical point of view,
SMACM is in many cases still a very custom method, where each lab
builds its own optical setup, uses its own sample preparation protocols,
and writes its own evaluation code.Considering this, it is
especially unfortunate that the field has yet to agree on a common
standard to measure the quality of an experiment. Certainly, there
are developments in that direction, for example the “single-molecule
fight club” as a place where analysis algorithms can be compared,
the nuclear pore complex samples, or, leaving the imaging space, the
multilab FRET study.[63,207,208] However, there is no universally accepted number or set of numbers
that need to be reported when publishing SMACM results. Certainly
all researchers should report numbers of detected photons and estimated
localization precisions, but the metrics are only a start. Again,
we could learn from other fields: For example, when publishing crystallography
or cryo-EM protein structures, a set of metrics that accurately state
the quality of the data set must be included. Of course, compared
to finding the structure of a single purified protein, imaging the
microtubule network in a cell is already less defined and more chaotic
(but certainly interesting and useful). Nevertheless, more rigorous
assessments than those currently used can be easily conceived. An
example would be Fourier Ring or Shell Correlation (FRC/FSC).[209,210] One can hope that a simple, “one size fits all” approach,
i.e. either a quick control experiment or an easy to calculate set
of metrics, would be a wise choice. Otherwise, such a standard will
probably be difficult to find general acceptance in practice. If the
field could converge on a standard that needs to be included in any
publication, it would be a key step toward making SMACM a robust tool
with straightforward comparison of results between different laboratories.
Also, it would force every researcher to carefully calibrate the optical
system, preventing the publication of poor-quality data sets.
Conclusion and Outlook
The previous sections
have discussed various topics in single-molecule
super-resolution microscopy (SMACM) where innovation is needed to
further develop the method into a general, robust approach to investigate
all areas of biology. We would like to conclude this review with some
more general and speculative thoughts on applications of the general
principle of localization-based super resolution imaging.Clearly,
images such as those shown in Figure are filled with detail far beyond that obtainable
from diffraction-limited imaging, and it is to be expected that these
techniques can be expected to be continuously applied. But to consider
future developments, it is useful to briefly revisit the fundamentals:
What is really needed to achieve super-resolution via localization
of an individual signal? Essentially, it comes down to a few key requirements—first,
an emitter is necessary. Second, its emission must undergo detectable
changes, ideally from a “bright” to a “dark”
state and back, yielding nonoverlapping signals in the spatial (or
another imaging) dimension. Third, the expected profile of the emission
on the detector must be known reasonably well to allow for fitting
of an appropriate model function to extract the single emitter position
precisely. (The intriguing new MINFLUX method[400] actually uses a zero of the optical pumping pattern to
find the position of the molecule, so the emission profile shape is
not so essential.) Finally, the time scale of the emission on–off
process must be fast enough to facilitate data acquisition in a reasonable
time and to avoid intrinsic changes of the investigated sample during
acquisition. If the sample is not fixed, the imaging should be restricted
to relatively static structures.Considering these prerequisites,
it becomes clear that the principle
of signal localization is universal. In that sense, the approach is
not limited to fluorescence—one could imagine other types of
signals in time and space that could be analyzed in a similar fashion,
for example nuclear or electron spin resonance. Certainly, the details
of such an implementation would differ quite significantly from current
approaches in optical microscopy and would necessitate some out-of-the-box
thinking. The essential first step is to be able to detect individuals
and this could possibly be envisioned, via linking to fluorescence[211−213] or to index changes from thermal perturbations.[214,215] In any case, these approaches would certainly be intriguing applications.SMACM is predominantly used in biology at the moment. This is not
too surprising, considering that optical microscopy has a long tradition
in this area. However, applications in other areas are increasingly
reported, for example in material sciences.[216] The capability to detect and localize individual emitters to investigate
e.g. the distribution of catalysis activity on different areas of
a nanoparticles,[217−219] or to observe the behavior of separation
polymers,[220,221] avoiding ensemble averaging
and revealing surprising heterogeneity, has turned out to be immensely
powerful in this area. Another area involves single molecules interacting
with metallic nanoantennas in order to observe nanoscale properties
such as hot spots involved in surface-enhanced Raman scattering;[222−224] indeed various combinations of plasmonics with super-resolution
have recently been reviewed.[225] Fundamentally,
super-resolution microscopy with single molecules provides a unique
window into complex systems, be it biological or otherwise. There
can be no doubt that this potential will be increasingly employed
in all areas of natural sciences: Single molecules will continue to
shine brightly.
Authors: Kyong-Rim Kieffer-Kwon; Keisuke Nimura; Suhas S P Rao; Jianliang Xu; Seolkyoung Jung; Aleksandra Pekowska; Marei Dose; Evan Stevens; Ewy Mathe; Peng Dong; Su-Chen Huang; Maria Aurelia Ricci; Laura Baranello; Ying Zheng; Francesco Tomassoni Ardori; Wolfgang Resch; Diana Stavreva; Steevenson Nelson; Michael McAndrew; Adriel Casellas; Elizabeth Finn; Charles Gregory; Brian Glenn St Hilaire; Steven M Johnson; Wendy Dubois; Maria Pia Cosma; Eric Batchelor; David Levens; Robert D Phair; Tom Misteli; Lino Tessarollo; Gordon Hager; Melike Lakadamyali; Zhe Liu; Monique Floer; Hari Shroff; Erez Lieberman Aiden; Rafael Casellas Journal: Mol Cell Date: 2017-08-10 Impact factor: 17.970
Authors: Qingfeng Zhang; Taylor Hernandez; Kyle W Smith; Seyyed Ali Hosseini Jebeli; Alan X Dai; Lauren Warning; Rashad Baiyasi; Lauren A McCarthy; Hua Guo; Dong-Hua Chen; Jennifer A Dionne; Christy F Landes; Stephan Link Journal: Science Date: 2019-09-27 Impact factor: 47.728
Authors: Zhe Liu; Wesley R Legant; Bi-Chang Chen; Li Li; Jonathan B Grimm; Luke D Lavis; Eric Betzig; Robert Tjian Journal: Elife Date: 2014-12-24 Impact factor: 8.140
Authors: Lucien E Weiss; Ljiljana Milenkovic; Joshua Yoon; Tim Stearns; W E Moerner Journal: Proc Natl Acad Sci U S A Date: 2019-02-28 Impact factor: 11.205
Authors: Joyce Woodhouse; Gabriela Nass Kovacs; Nicolas Coquelle; Lucas M Uriarte; Virgile Adam; Thomas R M Barends; Martin Byrdin; Eugenio de la Mora; R Bruce Doak; Mikolaj Feliks; Martin Field; Franck Fieschi; Virginia Guillon; Stefan Jakobs; Yasumasa Joti; Pauline Macheboeuf; Koji Motomura; Karol Nass; Shigeki Owada; Christopher M Roome; Cyril Ruckebusch; Giorgio Schirò; Robert L Shoeman; Michel Thepaut; Tadashi Togashi; Kensuke Tono; Makina Yabashi; Marco Cammarata; Lutz Foucar; Dominique Bourgeois; Michel Sliwa; Jacques-Philippe Colletier; Ilme Schlichting; Martin Weik Journal: Nat Commun Date: 2020-02-06 Impact factor: 14.919
Authors: Kristopher B Barr; Naihao Chiang; Andrea L Bertozzi; Jérôme Gilles; Stanley J Osher; Paul S Weiss Journal: J Phys Chem C Nanomater Interfaces Date: 2021-12-23 Impact factor: 4.177
Authors: Dongkwan Lee; Chenxi Qian; Haomin Wang; Lei Li; Kun Miao; Jiajun Du; Daria M Shcherbakova; Vladislav V Verkhusha; Lihong V Wang; Lu Wei Journal: J Chem Phys Date: 2021-04-07 Impact factor: 3.488