Literature DB >> 36067039

Toward Next Generation Lateral Flow Assays: Integration of Nanomaterials.

Amadeo Sena-Torralba1,2, Ruslan Álvarez-Diduk1, Claudio Parolo3, Andrew Piper1, Arben Merkoçi1,4.   

Abstract

Lateral flow assays (LFAs) are currently the most used point-of-care sensors for both diagnostic (e.g., pregnancy test, COVID-19 monitoring) and environmental (e.g., pesticides and bacterial monitoring) applications. Although the core of LFA technology was developed several decades ago, in recent years the integration of novel nanomaterials as signal transducers or receptor immobilization platforms has brought improved analytical capabilities. In this Review, we present how nanomaterial-based LFAs can address the inherent challenges of point-of-care (PoC) diagnostics such as sensitivity enhancement, lowering of detection limits, multiplexing, and quantification of analytes in complex samples. Specifically, we highlight the strategies that can synergistically solve the limitations of current LFAs and that have proven commercial feasibility. Finally, we discuss the barriers toward commercialization and the next generation of LFAs.

Entities:  

Mesh:

Substances:

Year:  2022        PMID: 36067039      PMCID: PMC9523712          DOI: 10.1021/acs.chemrev.1c01012

Source DB:  PubMed          Journal:  Chem Rev        ISSN: 0009-2665            Impact factor:   72.087


Introduction

In 2003, the World Health Organization (WHO) published its ASSURED criteria which outlined the ideal test for universally available diagnostics. These criteria stipulate that point-of-care (PoC) diagnostics should be Affordable, Sensitive, Specific, User-friendly, Rapid and robust, Equipment-free, and Deliverable to end-users. Two additional criteria, R (real-time connectivity) and E (ease of specimen collection and environmental friendliness), were added then to the original ASSURED, creating the new acronym of REASSURED.[1] Various PoC diagnostic platforms have been developed to realize biosensors that fulfill these criteria in the nearly two decades since they were published. Perhaps none more so than lateral flow assays (LFAs), which have risen to prominence thanks to their low cost and ease of use.[2,3] LFAs are currently the most used PoC sensors for diagnostic[4] (e.g., pregnancy test,[5] COVID-19 monitoring[6,7]), environmental (e.g., pesticides,[8] heavy metals,[9] and bacterial monitoring[10]), and food safety applications (e.g., foodborne allergens[11] and pathogens[12]).[13] The main reason that LFAs are so popular is because they are made on paper-based substrates. Paper-based materials enable low-cost and sustainable manufacturing, while their porous matrices provide equipment-free microfluidics (driven by capillary forces), a biocompatible scaffold capable of supporting biointeractions (e.g., between antibodies and antigens), and a flexible substrate on which diverse nano-biosensing designs and strategies can be developed (e.g., origami, barriers, and constrictions).[14] The fundamentals of LFA have been extensively covered in several peer-reviewed manuscripts and guides.[15,16] While LFAs are extremely versatile, they are still limited to the qualitative detection of a single highly concentrated analyte, lacking the sensitivity required for low-concentration analyte detection, especially in complex biological matrices. As such, much of the recent research on nanoparticle-based LFAs has focused on increasing their sensitivity,[17−19] multiplexing,[20,21] and novel methods to allow measurements in complex media,[22,23] as well as for detecting new biomarkers. However, most publications will seek to address each of these issues individually. To the best of our knowledge this is the first review that brings together the most recent, cutting-edge approaches obtained by the scientific community during the last 10 years, emphasizing the strategies that can synergistically solve the inherent challenges of developing PoC LFAs for clinical applications. This is vital in order to achieve the next generation of LFAs. The Review finishes by discussing the barriers toward commercialization and where lie the future perspectives of the field.

Quantification at the PoC

Quantitative analysis is defined as the ability to define the concentration of an analyte in a sample matrix.[24] Its use in LFAs makes the assays much more powerful than the binary “yes/no” answer obtained from qualitative LFAs.[25] For some applications, such as pregnancy tests, a simple binary readout is sufficient; the person taking the test is either pregnant or not. However, assay quantification is vital in applications where the concentration at which the target analyte is present can be used to inform decision making or where changes in concentration need to be monitored. Such examples include tests for analytes that are present under normal conditions but whose concentrations can be elevated or decreased in certain conditions. Perhaps the best example of this is a glucose sensor: a binary test saying that glucose is present in a sample matrix is not desirable; however, a quantifiable assay that determines whether the glucose concentration in a sample is higher or lower than accepted healthy limits is. There are a multitude of more complex clinical scenarios (e.g., cancer), where variations in the levels and ratios of several biomarkers (molecule’s overexpressions and downregulations) give insight on the efficacy of therapies and disease progression/prognosis.[26,27] Similar scenarios exist in food and environmental safety applications, where, for instance, the slight increase of food-borne allergenic content or pollutant concentrations in a river beyond accepted safe limits can have lethal consequences.[11,28] The potential to quantify the signal on an LFA is determined by the nature of the signal generated at the test line (TL) which needs to be in some way related to the concentration of biomarker in the sample. The signal obtained must then be compared to a calibration curve obtained by performing the assay with known concentrations of the analyte across the relevant concentration range in as similar a sample matrix as possible.[29] Quantification in LFAs has been performed by purpose-built readers with simple operation procedures. In all cases these should be designed to be portable, robust, and cheap for on-site/PoC use. Recent advances have included the development of software and apps as well as readers for more complex signal analysis, including using new techniques (e.g., SERS) and novel LFA design architectures (e.g., microarrays). In this section, we will review the recent advances in the development of LFA readers capable of LFA quantification in PoC/on-site scenarios. We compare the operation mode, price, dimensions, and reported assay sensitivity for each reader (Table ) and will give greater emphasis to those that are best suited to commercialization.
Table 1

Comparison of the Readout Systems for LFA Quantification

readersignaldetection ofLoD (pg mL–1)price ($)size (cm3)weight (g)ref
smartphone-basedcolorimetricE. coli O157:H7104 CFU mL–1∼2002.5 × 3.8 × 2400(38)
smartphone-basedcolorimetricHIgG1.00 × 1004∼2501.5 × 0.8 × 0.1140(39)
smartphone-basedcolorimetricEVD-IgG2.00 × 1005∼20013.6 × 6.9 × 0.7130(4)
smartphone-basedcolorimetricHIgG7.50 × 100421414 × 7 × 1130(40)
smartphone-basedluminescencePSA100∼200  (41)
smartphone-basedcolorimetric and luminescenceAFB1590∼2007 × 7 × 7 (42)
UCNP readerluminescencetriamcinolone acetonide980 30 × 30 × 152000(47)
UCNP readerluminescenceST-25000∼70024 × 9.4 × 5.4900(48)
UCNP readerluminescenceochratoxin A3000∼650  (49)
thermal contrast readerTCAhCG180∼50013.3 × 10.8 × 7.3 (58)
SERS readerSERShCG106>300018 × 11 × 4.7>1000(59)
photoacoustic readerPAglucose5.40 × 1006∼13004 × 2 × 2>4000(60)

LFA Semiquantification by Naked Eye

Semiquantification relies on determining an approximate quantity of the target analyte by visual inspection, without the need for external instrumentation. The methods to make an LFA semiquantifiable are to (i) relate a particular signal intensity or color tonality with a target concentration,[30] (ii) design the assay so that the signal appears only when the detected target analyte is above a threshold value,[31] or (iii) represent different target analyte concentrations using different test lines.[32] However, Saiykhan et al. have developed an innovative method that enables the semiquantification of blood plasma fibrinogen, based on the flow rate and distance traveled by the sample. In this work a wax-printed chromatography paper strip was functionalized with bovine thrombin. This causes the soluble fibrinogen to precipitate as the sample flows along the strip. Thus, the higher the concentration of fibrinogen in the sample, the lower the flow rate and distance traveled. The paper strip is placed inside a holder with a built-in scale (with millimeter resolution). This enables the visual monitoring of the distance traveled by the sample. The device is simple, portable, and cost-effective, enabling the PoC semiquantification of fibrinogen in a range of 0.5 to 7.0 mg mL–1.[33] The use of rulers has also been reported by Li et al., albeit in a much more complex system. The working principle of their assay relies on the use of antibody-functionalized platinum nanoparticles (PtNPs) that are captured on the TL as in a conventional LFA. After the assay has run, the TL is cut out of the strip and inserted into a sample reservoir at the base of a microfluidic ruler. The PtNPs catalyze the generation of oxygen that pushes red ink along the microfluidic channel. The distance traveled by the ink within a specific time can be directly related to the concentration of PtNPs captured on the TL and, therefore, to the concentration of target analyte present in the sample. The authors have applied this approach for the detection of prostate specific antigen (PSA) in clinical samples, achieving a limit of detection (LoD) of 0.54 ng mL–1 and a linear range of 0–12 ng mL–1.[34]

Smartphone-Based Readers

Rather than developing new hardware for LFA quantification and analysis, it would be beneficial to use ubiquitously available technologies for assay analysis. Smartphones have revolutionized the way we live; their portability, accessibility, widespread use, and sophistication have motivated their use as LFA readout systems for PoC quantification.[35] Modern smartphones are an interesting alternative to conventional benchtop readers, since they are portable miniaturized computers with large random access memory, high-speed CPUs, sophisticated camera lenses with Wi-Fi/Bluetooth, and IR network connectivity.[36] Thus, they possess all the features required of a PoC/on-site device: capable of assay quantification and data storage and transfer.[37,36] When taking pictures of an LFA for quantification, it is important that the lighting conditions and optical settings of the camera are optimized and kept constant. Variations in these factors can cause variations in the measured biomarker concentration. In this section, we will review recent advances in the use of smartphones for LFA quantification. The use of 3D printers allows the fabrication of mobile phone supports that guarantee the correct lighting and distance between the sample and the camera. This enables the reproducible collection of images for quantification.[15] For example, Jung et al. have developed a smartphone-based reader consisting of a 3D-printed holder, a smartphone, and a preinstalled dedicated android app. The holder includes a smartphone cradle with optical lens that provides a 3× image amplification and an LED light source with a reflector that focuses the light into a diffuser that spreads the light evenly across the strip. Moreover, the android app incorporates data analysis software that calculates the concentration of analyte in the sample based on a prestored calibration curve. The authors report an LoD of 104 CFU/mL for the direct detection of Escherichia coli O157:H7 in ground beef.[38] In another work, Quesada et al. improved the performance of their smartphone-based readout by attaching a cheap (20 USD) and commercially available microscopic lens onto their smartphone camera (Figure Ai). This increases the number of pixels that can be recorded, improving the quality of high-magnification images taken of the test area (Figure Aii).[39]
Figure 1

(A, i) Picture of microscope lenses attached to a smartphone camera that were used to photograph the detection area colorimetric LFA strips. (A, ii) Picture of the LFA strips taken with the microscopic lens and the smartphone camera. Reprinted (adapted) with permission from ref (39). Copyright 2019 Elsevier. (B, i) Schematic representation of a 3D-printed cartridge that contains all the optical apparatus for smartphone-based quantification of colorimetric and fluorescent LFAs. (B, ii) Screenshots from the universal detection app, showing the image analysis features for assay quantification. Reprinted (adapted) with permission from ref (42). Copyright 2020 Elsevier. (C, i) Picture of a portable UCNPs-LFA reader with dimensions of 24 × 9.4 × 5.4 cm and 0.9 kg weight. (C, ii) Schematic representation of the evaluation of the strips, in which the laser light is transmitted by the dichroic mirror at 45° to the detection zone. The smartphone camera captures the luminescence emission of the UCNPs present in the test and control lines, while the infrared filter blocks the laser’s residual light. Reprinted (adapted) with permission from ref (48). Copyright 2019 Elsevier.

(A, i) Picture of microscope lenses attached to a smartphone camera that were used to photograph the detection area colorimetric LFA strips. (A, ii) Picture of the LFA strips taken with the microscopic lens and the smartphone camera. Reprinted (adapted) with permission from ref (39). Copyright 2019 Elsevier. (B, i) Schematic representation of a 3D-printed cartridge that contains all the optical apparatus for smartphone-based quantification of colorimetric and fluorescent LFAs. (B, ii) Screenshots from the universal detection app, showing the image analysis features for assay quantification. Reprinted (adapted) with permission from ref (42). Copyright 2020 Elsevier. (C, i) Picture of a portable UCNPs-LFA reader with dimensions of 24 × 9.4 × 5.4 cm and 0.9 kg weight. (C, ii) Schematic representation of the evaluation of the strips, in which the laser light is transmitted by the dichroic mirror at 45° to the detection zone. The smartphone camera captures the luminescence emission of the UCNPs present in the test and control lines, while the infrared filter blocks the laser’s residual light. Reprinted (adapted) with permission from ref (48). Copyright 2019 Elsevier. Other researchers have exploited other smartphone capabilities. For instance, Brangel et al. have taken advantage of the smartphone’s GPS technology to develop an app that enables the geotagging of the samples to aid. They proved the applicability of this feature for Ebola surveillance screening and management of infected patients in Uganda.[4] Miller et al. have used smartphone video analysis as a cost-effective and simple method to calculate the dissociation constant (KD) of analyte and bioreceptor in an LFA. The authors suggest that this strategy can overcome several barriers to the quantitative analysis of LFAs, such as color inhomogeneity and reproducibility. The achieved KD showed excellent agreement with a reference benchtop interferometer.[40] Smartphone-based readers have also been applied for the quantification of fluorescent LFAs. Danthanarayana et al. have developed a portable imaging instrument for the evaluation of nanophosphor-based LFAs. In this case, the authors used the flash of the smartphone camera to excite the luminescent nanoparticles for 3 s; then, an image is acquired after a 100 ms delay. This time-gated imaging is a commonly used strategy to decrease the background signal since it allows the applied light to decay before the image is acquired. This approach enabled the detection of PSA and hCG with LoDs of 0.1 and 1 ng/mL, respectively.[41] Liu et al. have developed a smartphone-based reader for the dual quantification of colorimetric and fluorescent LFAs. They 3D-printed a 70 mm long cubic box using black photosensitive materials. The cartridge contains both white and UV LEDs as light sources for visible and fluorescent signaler excitation (Figure Bi). The images taken with the smartphone camera were analyzed with an Android Studio app that splits the image into red, green, and blue channels, with each channel being used to detect a different signaler in a multiplexed LFA (Figure Bii). The results obtained showed good consistency when validated against a liquid chromatography mass spectrometer (LC-MS).[42] Upconverting nanoparticles (UCNPs) are luminescent nanomaterials that display anti-Stokes shifts; i.e., the emitted photon has a higher energy than the incident photon. In such circumstances, it is possible to irradiate with light in the NIR range of the spectrum and measure light emitted in the visible or ultraviolet range.[43] UCNPs have several physical properties that make them suited for LFA quantification such as narrow emission spectra, high chemical stability, long luminescence lifetimes, and high resistance to photoquenching and photobleaching.[44] The evolution of the UCNP reading systems has gone from the use of bulky microtiter plate readers[45] to the use of custom-built commercially available fluorescent readers (UCP-Quant)[46,47] and, in recent years, the development of smartphone-based readers. Gong et al. have developed a compact (24 × 9.4 × 5.4 cm) and ultralight (0.9 kg) static image acquisition device (Figure Ci). The device consists of a compact infrared laser (exc. at 980 nm), a dichroic mirror (with 98% light transmission at 980 nm), and an infrared filter with nearly 0% transmittance at 980 nm (Figure Cii). The image is acquired with a smartphone with an image analysis app that enables simple user operation and easy result interpretation. The device has been validated for the detection of nucleic acids, small proteins, heavy metal ions, and bacteria, making this UCNPs-LFA reader ideally suited for commercialization.[48] Similarly, Jin et al. have developed a portable detection reader based on a 980 nm laser that excites the TL at a 45° angle. They have evaluated UCNP-LFA strips with a CCD camera (Nikon) and a smartphone (and using ImageJ software). They found that the smartphone was not as sensitive as the camera but still had a good linear correlation for ochratoxin A, mercury ions, and salmonella.[49] This result is not surprising and highlights the importance of the camera/smartphone in these readers, which is often overlooked in publications.

Portable LFA Readers for Alternative Signal Transduction Methods

Optical detection methods using colorimetric or fluorescent labels to give a visual readout are by far the most popular type of LFAs. However, despite their simplicity, cost-effectiveness, and fast acquisition times, these detection methods typically suffer from poor sensitivities, high noise levels, and low reproducibility.[50] As mentioned previously, this has led to the development of alternative signal transduction methods; however, these alternative methods need portable readers for PoC/on-site assay quantification. In this section, we will review advances in PoC readers for other (nonvisual) transduction methods

Magnetic Readers

Magnetic nanoparticles have been extensively utilized in LFAs for sample pretreatment, particularly for the isolation and preconcentration of the analyte. However, these nanoparticles also offer outstanding analytical capabilities when used as labels in magnetometric detection.[51] There are currently two reported ways of quantifying magnetic nanoparticles, by inductive sensing and by magnetic particle quantification (MPQ). Both require a coil that excites the magnetic nanoparticles and interrogates the magnetic signal generated at the TL of the LFA strip. The inductive sensing measures the magnitude and phase of the coil impedance, which varies depending on the change in magnetic permeability of the magnetic nanoparticles.[52] In MPQ the coil generates a nonlinear magnetic field that alternates between two frequencies, f1 and f2. The signal emitted by the nanoparticles after their magnetization is recorded by a coil attached to a magnetic reader.[53] Magnetic quantification offers several advantages over the traditional colorimetric reading, such as deep signal registration (it measures the signal generated in the TL volume rather than what is observable at the NC surface), extremely wide dynamic linear ranges (up to 7 orders of magnitude), and better sensitivity when evaluating complex samples (the opacity of the biological sample does not hinder the magnetic readout). Based on these analytical features, some researchers have focused on developing portable and cost-effective readers capable of performing MPQ at the PoC. Orlov et al. have designed an MPQ method and portable reader with outstanding analytical properties. Their assay used 200 nm diameter magneto-radioactive 59Fe-based MPs conjugated to monoclonal anti-PSA. The assay quantification is performed by the incorporation of the test strip in the measuring coil of the MPQ reader (Figure Ai). They report an LoD of 25 pg mL–1 of PSA in human serum, a dynamic range spanning 3.5 orders of magnitude, and a high dose–response sensitivity of klog = 3. The authors also suggest that the developed MPQ reader can perform real-time MP mapping during the biorecognition event. This could be a useful tool for the rapid determination of antibody–antigen binding kinetics.[54] The MPQ reader is also suited to multiplexing strategies, by the use of several coils that enable the interrogation of different TLs present on a single strip[55] or by the simultaneous evaluation of several strips in a cartridge (Figure Aii).[56] MPQ readers are still in the early development stage with most prototypes only existing in research laboratories, and none are commercially available at the time of writing. This makes it difficult to assess the commercial viability of such devices.
Figure 2

(A, i) Schematic representation of the MPQ quantification procedure of LFAs using a PoC MPQ reader. Reprinted (adapted) with permission from ref (54). Copyright 2016 Elsevier. (A, ii) Schematic representation of the simultaneous quantification of three LFA strips using a cartridge and the MPQ reader. Reprinted (adapted) with permission from ref (51). Copyright 2016 American Chemical Society. (B) Pictures of a portable LFA reader used for the quantification of the heat generated on the LFA TL due to thermal contrast amplification. The reader includes sensors for the evaluation of both heat conduction and radiation. Reprinted (adapted) with permission from ref (58). Copyright 2016 Springer. (C, i) Schematic representation of a SERS reader. (C, ii) Picture of the SERS reader during the scanning of the LFA strips. Reprinted (adapted) with permission from ref (59). Copyright 2019 Wiley.

(A, i) Schematic representation of the MPQ quantification procedure of LFAs using a PoC MPQ reader. Reprinted (adapted) with permission from ref (54). Copyright 2016 Elsevier. (A, ii) Schematic representation of the simultaneous quantification of three LFA strips using a cartridge and the MPQ reader. Reprinted (adapted) with permission from ref (51). Copyright 2016 American Chemical Society. (B) Pictures of a portable LFA reader used for the quantification of the heat generated on the LFA TL due to thermal contrast amplification. The reader includes sensors for the evaluation of both heat conduction and radiation. Reprinted (adapted) with permission from ref (58). Copyright 2016 Springer. (C, i) Schematic representation of a SERS reader. (C, ii) Picture of the SERS reader during the scanning of the LFA strips. Reprinted (adapted) with permission from ref (59). Copyright 2019 Wiley.

Thermal Contrast Readers

The use of TCA has been widely reported in recent years as a strategy to improve the sensitivity of LFAs. As previously explained in section , the working principle relies on the light-to-heat transduction of metallic nanoparticles upon their irradiation with an NIR laser. The cornerstone of this sensing method is the high resolution of infrared cameras, which can recognize temperature differences of just 0.1 °C, providing high assay sensitivity compared to image analysis of optical readout methods. As was the case with magnetic detection, TCA can overcome the light scattering issues of colorimetric detection systems. Recent research has focused on the development of portable alternatives to the current benchtop TCA readers.[57] Qu et al. have developed the first portable thermal contrast reader. It is small (133 × 108 × 73 mm) and has been fabricated with low-cost components (Figure B). They validate their reader with an LFA that uses anti-hCG functionalized Au nanoprisms. Once the LFA strip is placed in the reader, an NIR light beam is focused on the TL. The gold nanoprisms absorb the light and generate heat, which is detected by either a semiconductor sensor (heat conduction sensing) or infrared thermometer (heat radiation sensing). The semiconductor sensor has a resolution of 16 bit (0.0078 °C), while the infrared thermometer has a slightly higher 17 bit resolution (0.0039 °C). the higher sensitivity of the radiation method did however make it more prone to noise. The researchers recommend the selection of either sensing mode depending on the assay requirements. This approach yielded a 12-fold lower LoD (180 pg mL–1) than the comparable AuNP-based LFA for hCG detection.[58]

Surface-Enhanced Raman Scattering Readers

There have been several attempts to develop PoC SERS instruments. For instance, Tran et al. have developed a portable SERS-LFA reader with a fiber optic probe able to acquire 50 scans at different positions on the TL in 5 s. In this work the strip was scanned orthogonally to the TL, and the motorized stage would move one step every 10 ms (Figure C). This realized a rapid test with data acquisition times several orders of magnitude shorter than conventional Raman microscopes, ideal for PoC applications. Moreover, the method was 15 times more sensitive than the conventional AuNP-based colorimetric detection of hCG (LoD of 0.16 ng mL–1).[59] The main drawback to SERS readers being used in PoC applications is their cost. The most expensive components are the fiber optic Raman probes (>$2000) and the portable spectrometers (>$800). As well as finding ways to reduce the costs of these components, more powerful and user-friendly software will also need to be developed before SERS readout platforms can be used in PoC LFA quantification applications.

Sensitivity Enhancement

As for any diagnostic device, the optimization and design of an LFA should be tailored toward the intended final use; for example, it needs to be sensitive across the clinically relevant biomarker range.[61] Perhaps the most important analytical properties are sensitivity (i.e., the ability to discriminate between different analyte concentrations) and limit of detection (i.e., the lowest analyte concentration that can be discriminated from the blank).[62] The vast majority of clinical applications require the detection of sub-picomolar concentrations of biomarkers. In this section, we present recent advances aimed at improving the LFA sensitivity when using the noncompetitive assay format (Table ); we discuss their pros and cons as well as their commercialization feasibility.
Table 2

Comparison of the Sensitivity Enhancement Strategies Reported for LFA

strategymethod/materialdetection ofenhancementcomp. toref
flow rate decreasehydrophobic PCL nanofiberssynthetic Zika virus10-foldconv. LFA(65)
flow rate decreasecellulose nanofiber aerogelmouse IgG1000-foldconv. LFA(66)
flow rate decrease and reagents concentrationreduce detection area from 5 mm to 1 mmC-reactive protein30-foldconv. LFA(67)
flow rate decrease and pseudoturbulence generationwax pillarsHIgG3-foldconv. LFA(69)
internal incubation on TLdissolvable wax barrierHIgG51.7-foldconv. LFA(70)
flow controlcentrifugal forcesPSA6.2-foldconv. LFA(72)
sample preconcentrationmagnetic focusingvalosin-containing protein4000-foldno preconc.(73)
sample preconcentrationmagnetic focusinginfluenza A (H1N1) antigen26-foldno magnetic focusing(74)
sample preconcentrationfiltrationE. coli10-foldno filtration(10)
sample preconcentrationNafion-based ICPβ-hCG3.9-foldno preconc.(75)
increase antibodies concentrationfixing conj. AuNPs in TLC. sakazakii100-foldconv. LFA(76)
increase signal contrastcarbon nanoparticlesE. Coli3.8-foldconv. LFA(77)
increase signal contrastMWCNTsMeth.10-foldAuNPs-LFA(78)
fluorescence signal transduction and sample pre-concentrationmagnetic CdSe/ZnS QDs nanobeadsInfluenza ALoD of 22 pfu mL−1no comparison(81)
ratiometry with target-dependent and independent reportersred CdSe/ZnS QDs and blue polystyrene nanobeadsH-IgG78-foldAuNPs-LFA(82)
fluorescence signal transduction and increased signal-to-noise ratiofluorescent nanodiamondsbiotin105-foldAuNPs-LFA(84)
signal amplificationsilver staining of AuNPsTroponin I10-foldAuNPs-LFA(88)
signal amplificationAuNPs coated with HRPHIgG10-foldAuNPs-LFA(89)
signal amplificationAuNPs coated with ALPpotato virus X10-foldAuNPs-LFA(90)
signal amplificationAuNPs coated with ultrathin Pt skinsPSA100-foldAuNPs-LFA(91)
signal amplificationporous platinum core−shell nanocatalysts (PtNCs)P24100-foldno catalysis(92)
alternative signal transductionthermal contrast amplificationinfluenza, malaria, and C. difficile8-foldAuNPs-LFA(57)
alternative signal transductionthermal contrast amplificationP24 proteinLoD of 8 pg mL−1no comparison(97)
alternative signal transductionthermal contrast amplificationHCG12-foldAuNPs-LFA(58)
alternative signal transductionphotoacoustic imagingcryptococcal antigen (CrAg) from Cryptococcus100-foldAuNPs-LFA(98)
alternative signal transduction and sample preconcentrationsurface enhanced Raman scattering and magnetic focusingH1N1 and HAdV viruses2000-foldAuNPs-LFA(103)
alternative signal transduction and flow modulationsurface enhanced Raman scattering and sequential flow of Raman reportersthyroid stimulating hormone (TSH)LoD of 0.15 μIU/mLno comparison(104)
alternative signal transduction and bioreceptorssurface enhanced Raman scattering and bacteriophages as bioreceptorsS. Enteritidisno enhancementSERS-LFA based on antibodies(105)

Modulation of the Flow Dynamics

One of the main reasons for the poor sensitivity in LFAs, when compared to other analytical techniques like enzyme-linked immunosorbent assays (ELISA), is that they do not allow as much time for the bioreceptors (probes) to bind to the antigens (targets).[63] This is due to the capillary force driven wicking properties of the paper-based LFA strip, which produces a unidirectional flow of the sample at a constant rate through the substrate. Ideally the sample would be incubated on the test line (TL) to increase antigen capture, and therefore improve the sensitivity of the assay. There are two main methods of reducing the flow rate in LFAs: the first is by chemically modifying the nitrocellulose NC to reduce the flow rate, and the second is to change the geometry of the LFA components. The earliest examples of chemical modification of LFAs to slow the flow rate involved the incorporation of salt barriers to impede the flow. These can be easily drop cast onto the nitrocellulose and will halt the flow until they have been dissolved by the sample solution. Usually these are placed after the test and control lines, but a recent work by Prof. Feng Xu and co-workers stands out because the salt barrier was placed in front of the test and control lines.[64] Besides the reduction of the flow velocity and the consequent 10-fold sensitivity enhancement, the authors also claim that the increase in salt concentration improves the hybridization efficiency of the nucleic acid bioreceptors used. While this is a low-cost and simple approach, the high ionic strength of the sample solution at the target line will not be compatible with other bioreceptors and can denature some protein and aptameric receptors; therefore, its suitability should be evaluated for any given application. Recent advances in the field have focused on using polymeric chemical barriers as opposed to salts. For example, Yew et al. have electrospun a 10% solution of polycaprolactone nanofibers onto a nitrocellulose membrane for 60 s in order to delay the flow time by 17 s (0.29 mm/s). compared to uncoated nitrocellulose (NC) membranes (0.35 mm/s). The incorporation of polycaprolactone increases the hydrophobicity of the coated region, reducing the flow rate and increasing the interaction time between the biorecognition element and the analyte.[65] This strategy resulted in a 10-fold increase in the sensitivity of the assay, compared to the conventional LFA. In such approaches it is important to consider the properties of the materials being incorporated into the strip, as these should in no way compromise the test performance. Tang et al. have shown that the insertion of a nanofibrillated cellulose aerogel (pore size of 100 nm in the wet state) just after the conjugate pad can decrease the sample flow rate in the lateral flow device (Figure A). The aerogel transforms into a hydrogel-like state upon wetting, with an almost 103-fold pore size reduction compared to other reported stacking pads. This results in a 40–60% higher reaction time between the bioreceptors and the target analyte, which enables a 1000-fold sensitivity enhancement compared to the LFA without the aerogel (LoD of 0.72 ng mL–1 for mouse IgG in human serum). It is noteworthy that the aerogel has a shelf life of 6 months.[66] Alternatively, Katis et al. have developed a strategy not based on the incorporation of novel materials, but on the reduction of the Test line width from 5 to 1 mm.[67] They use a laser to polymerize a pattern on the nitrocellulose membrane. By narrowing the flow path, they achieve an increase in both the flow time and a concentration of the reagents. They report a 30-fold sensitivity enhancement for the detection of C-reactive protein compared to the assay with no flow focusing. The authors point out that this method is low-cost, since the polymerization can be performed using conventional lasers which is advantageous compared to other patterning techniques such as photolithography.[67]
Figure 3

(A, i) Schematic representation of a sensitivity enhancement approach based on the introduction of a cellulose nanofiber (CNF) aerogel after the conjugate pad. (A, ii) SEM image of the CNF aerogel, showing an average pore size of 250 μm. (A, iii) Comparison of the flow after 30, 60, and 90 s in strips with and without the CNF aerogel. Reprinted (adapted) with permission from ref (66). Copyright 2021 Elsevier. (B, i) Schematic representation of a sensitivity enhancement approach based on the incorporation of a dissolvable wax barrier after the test line. (B, ii) Optical microscopy picture (40×) of wax barriers with different widths. (B, iii) Picture of the LFA strips with and without the wax barrier, after the detection of HIgG (0–1000 ng mL–1). Reprinted (adapted) with permission from ref (70). Copyright 2020 Elsevier. (C, i) Modulation of the direction of flow of graphene quantum dots by the adjustment of the electrolyte solution pH in a paper-based electrophoretic bioassay. (C, ii) Separation of a mixture of CdSe@ZnS QDs, CdTe QDs, and N,S-doped carbon dots according to the electrophoretic mobility of the nanomaterials. Reprinted (adapted) with permission from ref (71). Copyright 2021 American Chemical Society.

(A, i) Schematic representation of a sensitivity enhancement approach based on the introduction of a cellulose nanofiber (CNF) aerogel after the conjugate pad. (A, ii) SEM image of the CNF aerogel, showing an average pore size of 250 μm. (A, iii) Comparison of the flow after 30, 60, and 90 s in strips with and without the CNF aerogel. Reprinted (adapted) with permission from ref (66). Copyright 2021 Elsevier. (B, i) Schematic representation of a sensitivity enhancement approach based on the incorporation of a dissolvable wax barrier after the test line. (B, ii) Optical microscopy picture (40×) of wax barriers with different widths. (B, iii) Picture of the LFA strips with and without the wax barrier, after the detection of HIgG (0–1000 ng mL–1). Reprinted (adapted) with permission from ref (70). Copyright 2020 Elsevier. (C, i) Modulation of the direction of flow of graphene quantum dots by the adjustment of the electrolyte solution pH in a paper-based electrophoretic bioassay. (C, ii) Separation of a mixture of CdSe@ZnS QDs, CdTe QDs, and N,S-doped carbon dots according to the electrophoretic mobility of the nanomaterials. Reprinted (adapted) with permission from ref (71). Copyright 2021 American Chemical Society. Other works have also investigated changing the flow geometry to improve LFA sensitivity; for example, Parolo et al. have enlarged the width of the sample and conjugate pads 3-fold in order to increase the bed volume of the strip. The authors achieve an 8-fold enhancement in the limit of quantification (LoQ) for HIgG detection; the simplicity of this strategy makes it ideal for incorporation into other LFAs.[68] Complex microfluidics can be cheaply, quickly, and easily introduced into LFAs, using wax printers, to improve the assay performance. For example, Rivas et al. introduced wax pillars at the beginning of the nitrocellulose membrane to produce a flow delay and generate pseudoturbulences in the capillary flow, both of which improve probe–analyte binding. Their method resulted in a 3-fold sensitivity enhancement for HIgG detection, in comparison to the conventional LFA.[69] The position of the printed microfluidics along the nitrocellulose membrane affects the assay’s sensitivity. The incorporation of the wax pillars just after the conjugate pad (as in the case of Rivas et al.) enhances mainly the interaction time of the first biorecognition event, the binding of the target analyte, and the labeled bioreceptors, while the incorporation of such artifacts after the TL provides an increase in the interaction time of both the first and second biorecognition event, the second being the immuno-sandwich formation with the secondary antibody. By optimizing both binding events, Sena-Torralba et al. report a 51.7-fold sensitivity enhancement for HIgG detection by inserting a 0.05 mm wide dissolvable wax barrier 1 mm downstream of the TL (Figure B).[70] A novel way of modulating the flow dynamics in LFAs is the application of alternative mobility forces, such as centrifugal or electrical forces. These provide the end-user with greater control of the flow rate and direction of the sample compared to capillarity, which is governed solely by the bed volume of the strip.[71] To this end, Shen et al. report a centrifugal force-assisted LFA, with a constant rotation speed of 1500 rpm, adjusted by a portable dedicated device.[72] This test enabled a 6.2-fold sensitivity enhancement for prostate specific antigen (PSA) detection in human serum.[72] Likewise, Sena-Torralba et al. have developed an LFA that incorporates electrophoresis. In this way, the flow rate and direction of the sample can be precisely adjusted by the regulation of the voltage applied to, and pH of, the electrolyte solution (Figure Ci). The authors prove that this paper-based electrophoretic bioassay can separate a mixture of quantum dots of different sizes and charges in just 10 min, due to the differences in the electrophoretic mobility of both nanomaterials (Figure Cii).[71] Based on this proof of concept, the researchers are currently working on the application of this approach for the betterment of the LFA sensitivity.

Sample Preconcentration

The sensitivity of LFAs can be vastly improved by the preconcentration/purification of the target analyte present in the sample, which is usually present at low levels and in a sample matrix with other biomolecules capable of interfering with the assay. Sample enrichment can be performed using magnetic separation or filtration. The former has been applied by Ren et al.,[73] who have used antibody-functionalized magnetic particles in order to concentrate valosin containing protein up to 10-fold, yielding a 4000-times greater sensitivity than the conventional LFA.[73] Similarly, Son Le et al.[74] have used superparamagnetic iron oxide nanoparticles (SPIONs) that provide a magnetic enrichment factor (φ) = 40, for the detection of C-reactive protein. This enables a 26-fold lower LoD (0.08 ng mL–1) compared to the conventional AuNP-based LFA (Figure A).[74] Magnetic enrichment strategies have shown great performance in the lab over the past decade; however, their commercialization has been hindered by the fact that they usually require multiple washing steps that make the assay less user-friendly and more prone to error.
Figure 4

(A, i) Schematic representation of a sensitivity enhancement approach based on sample preconcentration by magnetic focusing. (A, ii, a) TEM image, (b) HAADF-STEM image, and (c–f) EDS elemental mapping images of Pt–P2VP@SPION. Reprinted (adapted) with permission from ref (74). Copyright 2021 American Chemical Society. (B, i) Schematic representation of a sensitivity enhancement approach based on sample filtration. (B, ii) AuNP-based LFA approach for the detection of E. coli in tap water (0–109 CFU mL–1). Reprinted (adapted) with permission from ref (10). Copyright 2021 The Royal Society of Chemistry.

(A, i) Schematic representation of a sensitivity enhancement approach based on sample preconcentration by magnetic focusing. (A, ii, a) TEM image, (b) HAADF-STEM image, and (c–f) EDS elemental mapping images of Pt–P2VP@SPION. Reprinted (adapted) with permission from ref (74). Copyright 2021 American Chemical Society. (B, i) Schematic representation of a sensitivity enhancement approach based on sample filtration. (B, ii) AuNP-based LFA approach for the detection of E. coli in tap water (0–109 CFU mL–1). Reprinted (adapted) with permission from ref (10). Copyright 2021 The Royal Society of Chemistry. In the case of sample filtration, its use is limited to applications with large sample volumes, and only a representative amount of the sample is evaluated: for instance, the detection of E. coli in tap water. In this regard, Bergua et al. have developed a filtration system that uses a 0.25 μm pore size filter, a small peristaltic pump, and microfluidic tubes to filter 300 mL of water in 15 min. With this approach, the authors have improved the LoD by 1 order of magnitude to 104 CFU mL–1 (Figure B).[10] While this is an effective strategy, sample filtration relies on the use of external instrumentation and a time-consuming pretreatment of the sample. It is beneficial to integrate a sample preconcentration step into the LFA strip. Lee et al. have developed an LFA with a Nafion-based ion concentration polarization (ICP) preconcentrator integrated in the conjugate pad. ICP is an ion transport phenomenon that occurs when ions selectively pass through an ion exchange membrane. It enables a 15-fold sample preconcentration in just 8 min using a portable 9 V battery.[75] With this approach they achieved a 26% sensitivity enhancement for the detection of β-hCG.

Increasing Bioreceptor Concentration

One of the most straightforward ways to improve the LFA’s sensitivity is to increase the concentration of the capture bioreceptor in the TL, as this will lead to a greater analyte binding efficiency. However, this approach is limited by the loading capacity of the nitrocellulose membranes, which is usually around 80–100 μg cm–2.[16] Researchers are developing novel strategies that use nanomaterials with high surface-to-volume ratios to increase the immobilization density of bioreceptors. For instance, Pan et al. used immobilized AuNPs functionalized with antibodies to the test line. Simultaneously increasing the surface area and thereby antibody concentration at the test line, and the pre-existence of AuNP at the test line, meant the signal from nonimmobilized AuNPs from the conjugate pad was evident at lower target concentrations. With this method they were able to immobilize up to 2.2 mg mL–1 of antibodies at the TL and achieve a 100-fold sensitivity enhancement for the detection of Cronobacter sakazakii compared to the conventional LFA.[76] In this kind of approach, the nanomaterial selection dictates the assay’s performance. The nanomaterial at the TL needs to have the following four properties: (1) It needs to be small enough to penetrate the pores of the nitrocellulose membrane. (2) The suspension should avoid coffee ring effects. (3) The nanomaterial should contain functional groups that enable the oriented conjugation of the antibodies. (4) The optical properties of the nanomaterial should not interfere with the detection signal.

Novel Signal Transducers

Most commercial LFAs rely on the use of low-cost colorimetric AuNPs or dyed beads that quickly generate an optical signal that can be directly inspected by the naked eye. However, these are not suitable for scenarios where the target analyte is present at ultralow concentrations, mainly due to their low molar absorptivity, which means that a large accumulation of nanoparticles is required on the TL to generate a measurable optical signal. An alternative method of enhancing the sensitivity of LFAs is to use signal transducers with higher absorbances that create a stronger contrast with the background signal. Black carbon-based materials have garnered interest for this application. Porras et al. have reported a 3.8-fold lower LoD when using carbon nanoparticles compared to AuNPs, allowing the detection of E. coli in the nanomolar range by naked eye (Figure A).[77] Similarly, Sun et al. have reported that the use of multiwalled carbon nanotubes (MWCNTs) in LFAs can provide a 10-fold sensitivity enhancement for methamphetamine detection compared to conventional AuNP-based LFAs.[78] As well as providing an easily identifiable color change against the white background of the nitrocellulose, this nanomaterial possesses more binding sites than the AuNPs due to its greater surface area.
Figure 5

(A, i) TEM image of carbon nanoparticles that form aggregates of 10–100 nm. (A, ii) Photograph of LFA strips after detecting E. coli (0–7.09 pg μL–1) using carbon nanoparticles (black) and AuNPs (red). Reprinted (adapted) with permission from ref (77). Copyright 2021 MDPI. (B, i) TEM image of the magnetic QD nanobeads. (B, ii) Picture of the nanobeads before and after the magnetic enrichment step. (B, iii) Picture of the LFA strips after the detection of H1N1 (0–106 pfu mL–1), showing an LoD of 500 pfu mL–1 by naked eye. Reprinted (adapted) with permission from ref (81). Copyright 2021 Elsevier. (C, i) Color palette generated in the TL of the LFA upon the ratiometric combination of different concentrations of 650-QDs (target-dependent) and 450-NBs (target-independent) reporters. (C, ii) Calibration curve for HIgG detection using LFA strips with increasing concentrations (0–0.1%) of 450-NBS in TL. Reprinted (adapted) with permission from ref (82). Copyright 2021 Wiley. (D, i) Schematic representation of a nanodiamond-based LFA for the detection of HIV-1 RNA. (D, ii) An omega-shaped stripline resonator is fixed under the LFA strip to selectively separate the fluorescence signal at the TL from the background autofluorescence. Reprinted (adapted) with permission from ref (84). Copyright 2020 Springer Nature.

(A, i) TEM image of carbon nanoparticles that form aggregates of 10–100 nm. (A, ii) Photograph of LFA strips after detecting E. coli (0–7.09 pg μL–1) using carbon nanoparticles (black) and AuNPs (red). Reprinted (adapted) with permission from ref (77). Copyright 2021 MDPI. (B, i) TEM image of the magnetic QD nanobeads. (B, ii) Picture of the nanobeads before and after the magnetic enrichment step. (B, iii) Picture of the LFA strips after the detection of H1N1 (0–106 pfu mL–1), showing an LoD of 500 pfu mL–1 by naked eye. Reprinted (adapted) with permission from ref (81). Copyright 2021 Elsevier. (C, i) Color palette generated in the TL of the LFA upon the ratiometric combination of different concentrations of 650-QDs (target-dependent) and 450-NBs (target-independent) reporters. (C, ii) Calibration curve for HIgG detection using LFA strips with increasing concentrations (0–0.1%) of 450-NBS in TL. Reprinted (adapted) with permission from ref (82). Copyright 2021 Wiley. (D, i) Schematic representation of a nanodiamond-based LFA for the detection of HIV-1 RNA. (D, ii) An omega-shaped stripline resonator is fixed under the LFA strip to selectively separate the fluorescence signal at the TL from the background autofluorescence. Reprinted (adapted) with permission from ref (84). Copyright 2020 Springer Nature. Over the past decade, fluorescence detection has become a common way of enhancing the sensitivity of LFAs. Quantum dots (QDs) are among the most used labels due to their narrow emission peaks and high quantum yield (high photon emission rate).[79] In such systems it is possible to measure either the absorbance or the emission of the TL to calculate the analyte binding concentration; however, the emitted light is typically measured. The reason for this is that fluorescence measurements are more sensitive, mainly due to the way they are measured. While absorbance is measured over a bright background, fluorescence is measured without any reference beam and over a dark background.[80] This enables the detection of low levels of light. Recently, Bai et al. have developed magnetic QD nanobeads by forming MnFe2O4 magnetic beads that they then coated with PEI. These positively charge nanoparticles with magnetic cores were coated with CdSe/ZnS to create 200 nm diameter superparamagnetic MnFe2O4 core nanoparticles with “quantum dot” coated surfaces. The multifunctional composite nanoparticles were shown to be capable of the enrichment and ultrasensitive detection of influenza A virions, capable of detecting as little as 22 pfu mL–1 in nasopharyngeal swabs (Figure B).[81] Sena-Torralba et al. have developed a simple and cost-effective strategy based on the ratiometric combination of target-dependent red CdSe@ZnS QDs (650-QDs) and target-independent blue polystyrene nanobeads (450-NBs) in LFA (Figure Ci,ii). The approach enables the precise modulation of the assay’s dynamic range and sensitivity upon the use of different concentrations of both reporters. The authors prove a 78-fold higher sensitivity than the AuNP-based LFA when detecting HIgG and an 18-fold faster assay time (compared to ELISA).[82] As sensitive as fluorescence-based detection methods are, their performance can be hindered by the intrinsic background fluorescence of nitrocellulose membranes. The choice of the optimal substrate (nitrocellulose, glass, etc.) when developing an LFA for an analyte depends on many of the material, chemical, and optical properties of the substrate, which must be tuned for the optical mechanism being used. For example, a substrate should be chosen that does not autofluoresce at the same wavelength as the desired fluorophore emits. A recent work by Shah and Yager sets out a quantitative framework to aid the selection of the appropriate substrate for LFA development.[83] Other alternative fluorophores have also shown promise recently, most notably the use of fluorescent nanodiamonds by Miller et al. The emission intensity of these highly fluorescent nanoparticles is very high with emissions in the visible spectrum at (675 and 550 nm). The researchers have incorporated an omega-shaped stripline resonator under the LFA strip, which is used to apply a microwave-frequency electromagnetic field that selectively modulates the fluorescence signal of the nanodiamonds at a set frequency (Figure Di,ii). With this approach, the authors achieve sub-attomolar (8 × 10–19 M) limits of detection in a model biotin LFA, 105 times more sensitive than the detection obtained with AuNPs. Additionally, they demonstrate the clinical utility of this strategy by proving the detection of HIV-1 RNA in real plasma samples after 10 min of isothermal amplification.[84] Despite the recent advances in fluorescent LFA sensors, few have been commercialized, perhaps because they require instrumentation that conventional LFAs do not (excitation source, dark illumination, photodetector, etc.). To address this, portable versions of this equipment is being developed: such as phone-based fluorometers[85] or affordable infrared laser diodes for the excitation of upconverting nanoparticles (UCNPs) (section ).[86]

Signal Amplification

Signal amplification is also a well-known strategy for sensitivity enhancement in LFA. There are several ways to amplify the signal at the test line, the most widely reported of which are AuNP enlargement through silver staining, or coating the AuNPs with enzymes or catalytic metals.[87] Silver staining has been applied to the detection of Troponin I by the integration of water-soluble hybrid nanofibers between the conjugate pad and the TL. Once the AuNPs are bound at the TL, the silver enhancement reagents are released from the nanofibers and are reduced to metallic silver around the AuNPs (Figure A). The darkening of the TL enables a 10-fold sensitivity enhancement.[88] The integration of the signal amplification reagents in the LFA strip is the most interesting advancement of this approach, since the speed and simplicity of use of the LFA is not compromised, but the signal is enhanced.
Figure 6

(A) Schematic representation of the signal amplification strategy using water-soluble nanofibers and the silver enhancement reaction. Reprinted (adapted) with permission from ref (88). Copyright 2018 Elsevier. (B) Signal amplification strategy using AuNPs coated with Pt ultrathin layers (Au@Pt4L), for the detection of PSA. Reprinted (adapted) with permission from ref (91). Copyright 2017 American Chemical Society.

(A) Schematic representation of the signal amplification strategy using water-soluble nanofibers and the silver enhancement reaction. Reprinted (adapted) with permission from ref (88). Copyright 2018 Elsevier. (B) Signal amplification strategy using AuNPs coated with Pt ultrathin layers (Au@Pt4L), for the detection of PSA. Reprinted (adapted) with permission from ref (91). Copyright 2017 American Chemical Society. Enzymatic signal enhancement typically uses horseradish peroxidase (HRP) or alkaline phosphatase (ALP). These enzymes are traditionally used in ELISA and are commercially available, stable, and cheap. Parolo et al. achieved a 1 order of magnitude sensitivity enhancement in the detection of HIgG by the conjugation of AuNPs with anti-HIgG and HRP. The HRP catalyzed oxidation of 3,3′,5,5′-tetramethylbenzidine (TMB) was found to induce the best color change when compared to other methods, as would be expected.[89] Unfortunately this assay required multiple washing steps, which hinders its applicability at the PoC. Moreover, it is reported that the presence of HRP inhibitors or competitors in the real sample matrices impedes these LFAs. Panferov et al. functionalized AuNPs with ALP and obtained a 10-fold sensitivity enhancement for the detection of potato virus X in leaf extracts. The authors claim that the ALP catalyzed LFA retains all the advantages of conventional LFAs: 15 min assay time, no additional equipment, no extra washing steps, and all the components able to be stored in the dry state.[90] The signal amplification can also be performed by catalytic metals such as Pt, thin layers of which can be coated on AuNPs. Pt has a peroxidase-like catalytic activity toward the oxidation of TMB, so it can generate a blue TL. This approach enables a 100 times sensitivity enhancement for the detection of PSA, due to the large molar extinction coefficient of the oxidized TMB (Figure B). The authors also investigated the feasibility of the strategy for human plasma sample analysis, demonstrating that the complex sample matrix did not impair the catalytic activity of the Pt shells.[91] Similarly, Loynachan et al. have taken advantage of the high catalytic activity of porous platinum core–shell nanocatalysts (PtNCs) to achieve an ultralow limit of detection of 0.8 pg mL–1 for p24 protein by naked eye.[92]

Alternative Signal Transduction Methods

In recent years, researchers have made efforts to develop alternative signal transduction methods in LFAs, to address the fact that current optical readers and sensing methods are insufficiently sensitive and precise enough for some clinical applications. While these alternative signal transduction methods [thermal contrast amplification (TCA), photoacoustic (PA) analysis, and surface-enhanced Raman scattering (SERS)] have been successfully utilized in other biosensing platforms,[93−95] the goal has been to integrate these sophisticated detection systems into the simple paper-based architecture of LFAs. The development of PoC readers using user-friendly software to simply and quickly obtain and interpret assay results is discussed in section ; in this section, we will simply discuss advances in the methodology of different signal transduction methods. Thermal contrast amplification (TCA) is based on the measurement of the temperature change with an infrared camera, when irradiating the metal nanoparticles present on the test line with a near-infrared laser. The metal nanoparticles are used as light-to-heat transducers by adjusting the laser wavelength within the localized surface plasmon resonance (LSPR) peak of the nanoparticles. With this signal transduction principle and the relatively high resolution of infrared cameras (0.1 °C),[96] the TCA approach enhances the sensitivity while simultaneously allowing precise quantification in LFAs. For instance, Wang and co-workers observed an 8-fold sensitivity enhancement for the detection of influenza, malaria, and Clostridioides difficile, compared to the optical signal provided by AuNPs.[57] Similarly, Zhan et al. achieved an LoD of 8 pg/mL for the detection of p24 protein spiked into human serum,[97] and Qu et al. obtained a 12-fold lower LoD for HCG biomarker compared with the visual detection mode.[58] This detection method requires a careful calibration to compensate for background environmental temperature fluctuations. Photoacoustic (PA) imaging is based on the measurement of acoustic waves generated after the irradiation of metallic nanoparticles with a laser. The PA signal transduction method includes three steps: absorption of the light by the AuNPs, the conversion of the absorbed energy into heat, and the heat-induced thermal expansion of the air that generates pressure oscillations.[22] PA imaging provides outstanding sensitivity compared to optical imaging due to its ability to penetrate deeper into the nitrocellulose membrane. There is also a greater signal-to-noise ratio, due to the input (light) and output (acoustics) being different energy types. To date, there is only one reported use of PA in an LFA. This was reported by Zhao et al., who improved the sensitivity 100-fold when detecting cryptococcal antigen (CrAg), compared to the visible signal provided by the same AuNPs.[98] As suggested by the authors, a barrier to the adoption of this technique might be the high price of the PA detector and oscilloscope. Hence, researchers are working on the development of portable and cost-effective alternatives such as replacing the laser with LEDs. The integration of surface-enhanced Raman scattering (SERS) into LFAs has been extensively reported during the past decade as a sensitivity enhancement strategy.[99−101] It is based on the ability of metallic nanoparticles to enhance Raman signals.[102] More recently, researchers have focused on further enhancing the analytical properties of LFAs by combining the SERS signal transduction with other strategies. For instance, Wang and colleagues have developed a SERS-based LFA using Fe3O4@Ag magnetic nanoparticles loaded with DTNB dyes for dual magnetic sample preconcentration and SERS signal generation. With this strategy, they have achieved a 2000-fold sensitivity enhancement for the detection of H1N1 and HAdV viruses, compared to the conventional AuNP-based LFA.[103] Kim et al. have applied SERS on a dual-flow LFA, which was fabricated using 3-dimensionally stacked layers of wax-patterned nitrocellulose membranes (Figure A). The smaller 25 nm AuNPs used in this test were functionalized with anti-TSH antibodies and backfilled with biotinylated BSA (TSH, thyroid stimulating hormone). The biotin was then able to bind to streptavidin coatings on 45 nm AuNP that were flown over the TL in a secondary flow step. This caused an increase in SERS signal through electromagnetic enhancement effects. The authors note that the enhancement effect increases as the AuNP size increases, and simulations showed that using 45 nm AuNPs in both flow channels would give a greater SERS enhancement. However, the 25 nm AuNPs are more homogeneously distributed on the TL, which reduces the variation between test strips. With this approach, the authors achieve an LoD of 0.15 μIU/mL for the detection of TSH, with 0.5 μIU/mL being the threshold value for the diagnosis of hyperthyroidism.[104] Finally, Ilhan et al. have used SERS signal transduction in combination with bacteriophages as bioreceptors. Bacteriophages are viruses that specifically infect bacterial species (Figure B). The authors propose their use in LFA as an alternative to antibodies, which are time-consuming and costly to produce. The bacteriophage LFA system enabled an LoD of 7 CFU/mL and exhibited similar analytical performance to an antibody-based LFA for the detection of Salmonella enteritidis.[105]
Figure 7

(A, i) Schematic representation of a dual-flow LFA fabricated using 5 wax patterned nitrocellulose membrane layers. (A, ii) 25 and 45 nm AuNPs are introduced into the first and second inlet, respectively. This sequential addition of differently sized AuNPs allows for reproducible SERS signal enhancement on the TL. Reprinted (adapted) with permission from ref (104). Copyright 2021 American Chemical Society. (B, i) A bacteriophage covalently coupled to AuNPs by carbodiimide coupling is used as a highly specific bioreceptor. (B, ii) The Bacteriophage LFA system enables the quantitative detection of S. enteritidis using SERS. Reprinted (adapted) with permission from ref (105). Copyright 2021 Elsevier.

(A, i) Schematic representation of a dual-flow LFA fabricated using 5 wax patterned nitrocellulose membrane layers. (A, ii) 25 and 45 nm AuNPs are introduced into the first and second inlet, respectively. This sequential addition of differently sized AuNPs allows for reproducible SERS signal enhancement on the TL. Reprinted (adapted) with permission from ref (104). Copyright 2021 American Chemical Society. (B, i) A bacteriophage covalently coupled to AuNPs by carbodiimide coupling is used as a highly specific bioreceptor. (B, ii) The Bacteriophage LFA system enables the quantitative detection of S. enteritidis using SERS. Reprinted (adapted) with permission from ref (105). Copyright 2021 Elsevier.

Multiplexing

Given the complex physiology of many pathological states (e.g., cancer), which produce multiple different biomarkers, the ratios and concentrations of which can be indicative of disease progression,[106] diagnostic tests should ideally measure as many clinically relevant biomarkers as possible. Information about different biomarkers can inform clinical decisions based on the needs of individual patients.[107] This need is also extended to food, environmental, and safety applications, where the assay’s utility increases with the number of targets detected (e.g., detecting multiple pollutants at once would speed up the analysis of water quality). Multiplexing is the simultaneous analysis of several analytes in a single sample and has been investigated for decades given its obvious time and sample saving potential.[108−110] Considering that most diagnostic samples are extracted in limited volumes (e.g., blood and nasopharyngeal swab), multiplexing saves sample volume, time, and cost.[111,112] Despite the many potential advantages that multiplexing offers, adding any complexity to the system inevitably leads to practical difficulties that must be overcome. When developing multiplexed LFAs there a several biological and physical issues that need to be resolved.[20] One major challenge is the simultaneous detection of analytes present in the sample at very different concentrations (e.g., fM and μM),[113] as sensors are developed and designed to work across a clinically relevant range. This has led to the development of cost-effective and simple strategies that allow the modulation of the assay’s dynamic range. Likewise, it is not always easy to resolve the signal from the detection of each of the biomarkers under investigation, which is usually solved by using different detection methods (e.g., fluorophores) for each biomarker; however, the optimization of the rest of the LFA may not be ideal for both biomarkers. Data handling has also appeared as a bottleneck when using multiplexed sensors. This has resulted in an increase in papers in recent years focusing on the development of software for the analysis and interpretation of large data sets in multiplexed diagnostics.[114] In this section, we will review recent advances and technologies that have been developed to address the aforementioned challenges in multiplexed LFA development. These involve the design of new LFA architectures/functionalities or the use of novel signal transduction mechanisms (Table ). Attention will be drawn to those methods that enable PoC and high-throughput multiplexing with ultrahigh sensitivity as well as those that are feasible for commercialization.
Table 3

Comparison of the Multiplexing Strategies Reported for LFA

strategysignal transducertarget analytesdetection ofsamplesensitivity enhancementref
multiple TLsSiO2@DQD2SARS-Cov-2-related HIgG and HIgMhuman serum104-fold(117)
multiple TLsAuNPs3miRNA-21, miRNA-155, and miRNA-210human serumno(118)
multiple TLsAuNPs3Giardia, Cryptosporidium, and Entamoebahuman stoolno(119)
multiple TLsAuNPs3HIV, HCV, and HAV antibodieshuman serumno(121)
multiple TLs and 2 labelsFAM and ROX fluorophores13HPV (16, 18, 31, 33, 35, 39, 45, 51, 52, 56, 58, 59, and 68)human cervical swabno(122)
multiple dotsAuNPs7DNA alleles (FY*01, FY*02, FY02N.01, GYPB*03, GYPB*04, JK*01, and JK*02) related to four blood group SNPshuman whole bloodno(125)
microarrayAuNPs4morphine, amphetamine, methamphetamine, and benzoylecgoninehuman urineno(126)
multiple stripsUCP nanoparticles10E. coli O157:H7, S. paratyphi A, S. paratyphi B, S. paratyphi C, S. typhi, S. enteritidis, S. choleraesuis, V. cholera O1, V. cholera O139, and V. parahemolyticusfood samplesno(128)
wax-patterned multichannelcatalytic signal amplification of AuNPs3Clostridioides difficile toxins A and B, and glutamate dehydrogenase (GDH)human stool8-fold(129)
multiple stripsmagnetic nanolabels3BoNT-A, BoNT-B, and BoNT-Emilk, apple, and orange juices1000-fold(56)
single TL and 3 labelsorange, red, and green AgNPs3DENV NS1 protein, YFV NS1 protein, and ZEBOV glycoprotein no(135)
single TL and 3 labelsAgNBA@Au, AgMB@Au, and AgR6G@Au SERS nanotags3CK-MB, cTnI, and Myo cardiac biomarkershuman serum100-fold(136)
multiple TLs and 2 labelsred and blue latex beads4IgG and IgM specific to DENV and CHIKVhuman whole bloodno(130)
multiple TLs and 3 labelsAgNPs, spherical and desert-rose-like AuNPs3casein, ovalbumin, and hazelnut allergenic proteinsbiscuitsno(131)
multiple TLs and 2 labelsZrMOF@CdTe NPs2heart-type fatty acid binding protein (h-FABP) and cardiac troponin (cTnT)human serum10-fold(133)
multiple TLslanthanide-doped nanoparticles (YVO4: Eu 40%)3staphylococcal enterotoxins SEG, SEH, and SEIPBS buffer100-fold(134)
multiple TLsAgNBA@Au SERS NPs3Myoglobin, cTn1 & CK-MBhuman serum1000-fold(123)
multiple TLsAuNPs labelled with malachite green isothiocyanate (MGITC)2Clostridium difficile surface layer protein A (SlpA) and toxin B (ToxB)human stool1000-fold(124)

Multiplexing Enabled by Architecture Modifications of the LFA Strip

Several Detection Lines in Parallel

The most obvious and simple way to perform multiplexing in LFA is by creating multiple different test lines (TLs) on the detection zone, with each TL responsible for the biorecognition of a different analyte. Most multiplexing approaches that adopt this methodology report the detection of up to three different target analytes. The limited length of the detection zone of the strip, usually around 4 cm, limits the number of TLs that can be incorporated into the LFA.[15] The first TL must be fixed at a proper distance from the conjugate pad so as to provide enough time for the first biorecognition event to occur between the conjugated label and the target analyte.[70] This distance should be optimized during the development of each LFA but is rarely less than 2 cm. Additionally, the distance between TLs should be at least 2 mm to allow the simple error-free readout by eye.[115] Coupled with the inclusion of a control line (CL), there is an obvious limit on the number of TLs that can be incorporated into a standard LFA. An obvious solution would be to elongate the LFA strip; however, this causes an exponential increase in assay time; as described by the Washburn equation.[116] Wang et al. have developed a multiplexed and highly sensitive fluorescent LFA for the detection of SARS-CoV-2-related IgG and IgM. They fixed secondary antibodies specific to human IgM and IgG on the TL1 and TL2, respectively (Figure Ai); in the conjugate pad, they had 200 nm diameter silica-core@dual quantum dot (QD)–shell nanotags (SiO2@DQD) (Ems. 615 nm, Exc. 365 nm) functionalized with SARS-CoV-2 spike proteins (Figure Aii). The efficacy of this multiplexed LFA was established by the simultaneous detection of IgG and IgM from SARS-cov-2 positive and negative patient serum samples.[117] Zheng et al. developed an LFA with three TLs for the simultaneous detection of micro-RNA-21, micro-RNA-155, and micro-RNA-210 cancer biomarkers in human serum. In such tests, the specificity of the probe sequences is vital for proper assay function, to prevent false positive results from nonspecific hybridization/binding.[118] Multiplexed nucleic acid biomarker detection in LFAs has also been reported by Crannell et al. for the detection of DNA sequences specific to Giardia, Cryptosporidium, and Entamoeba protozoa in stool samples, all of which are usually present at very low concentrations. To improve assay sensitivity, this work used an isothermal amplification method (recombinase polymerase amplification, RPA) to generate more of the DNA sequence of interest, achieving an ultralow LoD of 444, 6, and 9 parasites/test for Giardia, Cryptosporidium, and Entamoeba, respectively.[119] Many isothermal amplification technologies exist and are being investigated for use in LFAs; however, to date, most publications do not perform the amplification in the test strips but rather ex situ prior to running the assay. Likewise, it is common for papers to exclude the amplification time from the assay time to result, which they often increase substantially.
Figure 8

(A, i) Schematic representation of a multiplexed assay based on two consecutive TLs for the detection of IgM and IgG for SARS-Cov-2 spike proteins. (A, ii) Elemental mapping images of the SiO2@DQD used as fluorescent labels in the publication by Wang et al. (A, iii) Picture of the LFA strips after performing the assay for the detection of SARS-cov-2 antibodies, where the red fluorescence signal generated in the M, G, and C lines indicates the presence of IgM (M) or IgG (G), with a positive control (C). Reprinted (adapted) with permission from ref (117). Copyright 2021 The Royal Society of Chemistry. (B, i) Schematic representation of the mobile phone reader used to quantitatively evaluate multiplexed LFA microarrays. (B, ii) Picture of the algorithmically determined immunoreaction spot layout, where each row corresponds to a different spotting condition (1–7). (B, iii) Heat map of the microarray used to optimize the spot configuration by machine learning. Reprinted (adapted) with permission from ref (127). Copyright 2020 Springer Nature.

(A, i) Schematic representation of a multiplexed assay based on two consecutive TLs for the detection of IgM and IgG for SARS-Cov-2 spike proteins. (A, ii) Elemental mapping images of the SiO2@DQD used as fluorescent labels in the publication by Wang et al. (A, iii) Picture of the LFA strips after performing the assay for the detection of SARS-cov-2 antibodies, where the red fluorescence signal generated in the M, G, and C lines indicates the presence of IgM (M) or IgG (G), with a positive control (C). Reprinted (adapted) with permission from ref (117). Copyright 2021 The Royal Society of Chemistry. (B, i) Schematic representation of the mobile phone reader used to quantitatively evaluate multiplexed LFA microarrays. (B, ii) Picture of the algorithmically determined immunoreaction spot layout, where each row corresponds to a different spotting condition (1–7). (B, iii) Heat map of the microarray used to optimize the spot configuration by machine learning. Reprinted (adapted) with permission from ref (127). Copyright 2020 Springer Nature. Lee et al. have developed a triple-TL LFA with a 100% clinical sensitivity and specificity for the detection of antibodies specific to HIV, hepatitis A, and hepatitis C antigenic peptides in patient sera. They have achieved this by immobilizing “proteinticles” in the TL, which hold the bioreceptors (antigenic peptides) on the nitrocellulose surface in a homogeneous orientation and conformation. “Proteinticles” are nanoscale protein particles that self-assemble inside cells with constant 3D structure and surface topology.[120] In short, they are made by inserting the genetic sequence of the viral antigen peptide of interest to the C- or N-terminus of known proteins; for instance, in the work cited here by Lee et al., peptides from three viruses (HIV, HCV, and HAV) are inserted into the human ferritin heavy chain to create proteinticles.[121] These are then expressed in bacteria to produce large protein complexes with structured, orientated, 3-dimensional scaffolds containing the antigen peptides being used as a bioreceptor in the assay. Proteinticles serve as an alternative to the direct fixation of the peptides on the TL, which normally suffers from limitations such as uncontrolled orientation, clustering, inactivation, or instability of the peptides, which can reduce the accuracy of the assay.[121] The simplest method of multiplexing without extending the LFA strip length is to print test lines with two different bioreceptors that conjugate to analytes bound to different reporters (i.e., fluorophores that emit at different wavelengths). This method can be further multiplexed by printing multiple test lines, as has been shown by Xu et al.[122] who have developed a multiplexed LFA for the simultaneous detection of the nucleic acid sequences specific to 13 different human papillomavirus (HPV) clades (16, 18, 31, 33, 35, 39, 45, 51, 52, 56, 58, 59, and 68). The researchers deposited 7 TLs, with 2 different bioreceptors in each TL, and used two fluorophores (FAM and ROX) as reporters. In this work, the nucleic acids were preamplified by linear-after-the-exponential (LATE)-PCR. As in all assays that preamplify the DNA, the sensitivity is determined by the amplification method; as such, they report an LoD of 10–102 copies plasmid DNA/μL. They also report a high specificity with no cross-reactivity among 31 common HPV types.[122] Multiple TLs need not only be used for optical readouts. Zhang et al. have taken advantage of the ultrasensitive signal and a wide linear dynamic range of SERS, to develop a multiplexed LFA for the detection of myoglobin, cTnI, and CK-MB cardiac biomarkers in human serum. The authors used SERS nanotags (silver core/gold shell NPs) loaded with Nile blue A dye (AgNBA@Au) and a triple TL assay design. The assay displayed a linear response over the clinically relevant concentration ranges of the three biomarkers spanning over 6 orders of magnitude. They were able to obtain LoDs of 3.2, 0.4, and 0.5 pg mL–1 for myoglobin, cTnI, and CK-MB, respectively.[123] Similarly, Hassanain et al. have utilized SERS for both the multiplexed and highly sensitive detection of Clostridium difficile surface layer protein A (SlpA) and toxin B (ToxB). In this work, the authors report the use of a handheld Raman spectrometer that enables the PoC quantification of the biomarkers down 0.01 pg μL–1.[124]

Microarrays

To make better use of the test area afforded by lateral flow strips, rather than printing test lines, it is possible to make microarrays using test spots. Once example of this is the work published by Gomez-Martinez et al., who designed a 4 × 2 microarray for the simultaneous detection of 7 DNA alleles (FY*01, FY*02, FY02N.01, GYPB*03, GYPB*04, JK*01, and JK*02) for rapid blood group genotyping. The authors spotted 8 capture oligonucleotide probes on the test area of a single nitrocellulose membrane, with one of these being the positive control. While this approach successfully enabled a fast blood group genotyping (total processing time of 1 h), it did require a LATE-PCR preamplification step which increased the assay complexity and price, since extra instrumentation and reagents were required.[125] In another work, Taranova et al. managed to deposit up to 32 spots in a microarray format (4 × 8) on the NC membrane. The spots had a diameter of 250 μm and were precisely deposited by dispensing 20 nL of the capture reagents with a steel pin and program-controlled manipulator. Rather than using this approach to detect 32 different analytes, they used the 8 spots in each row as replicates for the detection of 4 drugs (morphine, amphetamine, methamphetamine, and benzoylecgonine) in urine. Making repeat measurements in a single strip increases the robustness of the assay results, allowing for more accurate quantification. The authors claim that with the resolution of their setup, they have the capability to print arrays with up to 150 spots in the detection zone.[126] This would be a large step forward in the development of high-throughput multiplexed LFAs. Ballard et al. have developed a machine-learning-based framework to determine the optimal configuration of the spots in the microarray and to quantify the analyte concentration. A mobile phone-based readout system that uses custom-built image processing software to analyze the microarrays has also been developed (Figure Bi). The algorithm is designed for assay optimization, designed to accurately select the best sensing conditions out of up to seven possibilities being tested in the microarray (Figure Bii). Additionally, the deep-learning framework developed enables the selection of the best spatial configuration of the immobilized spots (Figure Biii).[127]

Combination of LFA Strips and Design of Innovative Paper Configurations

A quick and simple way to develop a multiplexed LFA is by connecting several test strips to a single sample pad. This has been reported by Zhao et al., who have developed a multiplex strategy based on a 10-channel lateral flow assay for the detection of 10 foodborne pathogens (E. coli O157:H7, Salmonella paratyphi A, S. paratyphi B, S. paratyphi C, Salmonella typhi, S. enteritidis, Salmonella choleraesuis, Vibrio cholera O1, V. cholera O139, and Vibrio parahaemolyticus). Each strip was functionalized with monoclonal antibodies specific to one of the pathogenic bacteria, and the conjugate pad was prepared by conjugating the antibodies with phosphor nanoparticles. The results obtained were 100% consistent when validated against the culture-based detection of samples taken from 279 food samples.[128] Despite being fast and reliable, the use of this kind of multiplexing format is limited by the requirement of larger sample volumes. In this case, the sample volume required was 7 times higher (700 μL) than the single-strip LFAs. This prevents the use of this multiplexing method in scenarios where the sample volumes are prohibitively low. Han et al. have developed a similar strategy but, in this work, stack nitrocellulose sheets on top of each other, using wax patterning to create different flow paths vertically through the sheets (Figure i). This methodology was implemented to allow the sequential flow of signal amplification reagents to increase the flow sensitivity (Figure ii). With this approach, the authors achieved the simultaneous detection of Clostridioides difficile toxins A and B, and glutamate dehydrogenase (GDH) (Figure iii). The signal amplification strategy, based on the aggregation of AuNPs, enabled the detection of the target analytes at ultralow concentrations, with LoDs of 0.16, 0.09, and 0.03 ng mL–1 for GDH, C. difficile toxin A, and C. difficile toxin B, respectively.[129]
Figure 9

(i) Schematic representation of the multichannel device created by Kyoung Han et al. in which wax patterned-NC membranes were stacked in multiple layers (ii) to create flow paths that allow for the sequential flow of the sample and signal amplification reagents to the test line. (iii) Picture of the device after the detection of GDH and Clostridioides difficile toxins A and B. Reprinted (adapted) with permission from ref (129). Copyright 2021 Elsevier.

(i) Schematic representation of the multichannel device created by Kyoung Han et al. in which wax patterned-NC membranes were stacked in multiple layers (ii) to create flow paths that allow for the sequential flow of the sample and signal amplification reagents to the test line. (iii) Picture of the device after the detection of GDH and Clostridioides difficile toxins A and B. Reprinted (adapted) with permission from ref (129). Copyright 2021 Elsevier. A multiplexing strategy based on the interrogation of three individual LFA strips using a single processor unit has been reported by Orlov et al. In this work magnetic nanoparticles were conjugated to monoclonal antibodies and used as labels, with each strip responsible for the detection of a particular target analyte. The assays were developed to detect Botulinum neurotoxins (BoNTs) A, B, and E in whole milk and juice samples and were simultaneously analyzed using a magnetic particle quantification (MPQ) reader.[56] The use of this type of signal transduction enabled high assay sensitivity compared to the conventional colorimetric readout, allowing the detection of the target analytes in the low pg mL–1 range. The authors observe a linear range across 7 decades in this work which highlights the suitability of this method for multiplexed LFAs where analytes can span wide concentration ranges. Special attention should be given to the price, portability, and user-friendliness of the MPQ reader, as this might hinder the applicability of this platform in real scenarios.

Using Nanoparticles for Multiplexing

As has been seen throughout this Review so far, nanoparticles are routinely used in LFAs. By designing LFAs with different nanoparticles that generate different signals it is easy to envisage how multiplexed assays can be developed. In such systems it is important to be able to distinguish the signals quickly and easily from the different nanoparticles to allow the fast, efficient, and qualitative interpretation of the assay results. Recent research has also focused on the integration of low-cost, portable, and robust assay readout technologies. This section of the Review will focus on novel nanoparticle LFAs that have been developed for multiplexed sample analysis. As well as AuNPs, latex beads are very popular nanoparticles for LFA development. An example of latex beads being used in a multiplexed LFA is the work of Lee et al., who developed a multiplex LFA that uses red and blue latex beads for the detection of IgG and IgM produced as a response to infections by dengue virus (DENV) and Chikungunya virus (CHIKV). The appearance of a purple color on the TL indicated the presence of both IgG and IgM antibodies in the sample. A quantitative evaluation of the strips was performed with a smartphone camera and image analysis software. In this work the hue (H) value was analyzed instead of the RGB value, as it was found to be more accurate.[130] Similarly, Anfossi et al. used spherical AgNPs, spherical AuNPs, and “desert-rose-like” AuNPs, with SPR peaks at 420, 525, and 620 nm, respectively, to develop a multiplexed LFA for detection of casein, ovalbumin, and hazelnut allergenic proteins in commercial biscuits. The LFA comprises three lines, each responsive to one allergen. In order to obtain equivalent signal intensities from the different nanoparticles, the authors had to increase the concentration of AgNPs in the conjugate pad. This is not surprising since AgNPs have lower extinction coefficients than AuNPs. The results of this particular multiplexed assay are easy to interpret since the colors of each nanoparticle (cyan, yellow, and magenta) can be easily interpreted using conventional image analysis software.[131] The appropriateness of the colorimetric LFA is usually diminished by the low sensitivity and ease of quantification of this type of signal readout. Quantum dots (QDs) have advanced optical properties that make them ideal for the development of highly sensitive, easily quantifiable multiplexed LFAs. However, the fluorescence emission of QDs can be suddenly quenched by biomolecules present in biological samples, compromising the sensitivity and reliability of the assay.[132] To resolve this, recent research has investigated the encapsulation of QDs into nanoparticles that shield them from quenchers. For instance, Zou et al. have reported the encapsulation of CdTe QDs into cubic zirconium metal organic frameworks (ZrMOFs; approximately 120 nm long) (Figure Ai), and their use as fluorescent labels for the simultaneous detection of heart-type fatty acid binding protein (h-FABP) and cardiac troponin T (cTnT). This nanomaterial is synthesized with a simple hydrothermal method, in which the reaction time can be modulated to obtain ZrMOF@CdTe NPs with different fluorescent properties. The authors have taken advantage of this to develop a multiplexed LFA that uses two TLs for ZrMOF@CdTe NPs that emit yellow and green light after being excited by a single wavelength (Figure Aii). This approach yielded visual LoDs of 1 and 250 μg L–1 for h-FABP and cTnT, respectively, below the lowest clinically relevant concentration. In addition, the researchers report a high specificity over other cardiac biomarkers [including D-dimer (D-D), myoglobin, and C-reactive protein (CRP)].[133]
Figure 10

(A, i) SEM and TEM images of the ZrMOF@CdTe NPs, which show a cubic shape and a size around 120 nm. (A, ii) Multiplexed LFA based on the use of two TLs and ZrMOF@CdTe NPs with yellow and green fluorescent emissions. Reprinted (adapted) with permission from ref (133). Copyright 2021 The Royal Society of Chemistry. (B, i) Multiplexed LFAs showing the simultaneous detection of the staphylococcal enterotoxins I, G, and H, using lanthanide-doped nanoparticles. (B, ii) Smartphone-based readout system for the on-site LFA’s quantification. Reprinted (adapted) with permission from ref (134). Copyright 2021 The Royal Society of Chemistry.

(A, i) SEM and TEM images of the ZrMOF@CdTe NPs, which show a cubic shape and a size around 120 nm. (A, ii) Multiplexed LFA based on the use of two TLs and ZrMOF@CdTe NPs with yellow and green fluorescent emissions. Reprinted (adapted) with permission from ref (133). Copyright 2021 The Royal Society of Chemistry. (B, i) Multiplexed LFAs showing the simultaneous detection of the staphylococcal enterotoxins I, G, and H, using lanthanide-doped nanoparticles. (B, ii) Smartphone-based readout system for the on-site LFA’s quantification. Reprinted (adapted) with permission from ref (134). Copyright 2021 The Royal Society of Chemistry. Lanthanide-doped nanoparticles (YVO4/Eu 40%) have been shown by Mosseau et al. to possess remarkable optical properties compared to the QDs, due to an absence of photobleaching and longer luminescence times. One of the advantages of this work is the ease of nanoparticle synthesis which is achieved by the coprecipitation YVO4 and Eu salts. These nanoparticles were used for the simultaneous detection of staphylococcal enterotoxins G, H, and I, by the conjugation of the nanoparticles with monoclonal antibodies and the immobilization on three consecutive TLs in the detection pad (Figure Bi). This multiplexed assay was 100 times more sensitive than the comparative AuNP-based LFA, reaching LoDs in the pg mL–1 range for all the biomarkers investigated. Additionally, the researchers developed a portable and cost-effective smartphone-based readout system (see section ; Figure Bii).[134]

Multiplexing in a Single TL

The detection of different target analytes within a single TL is probably the most powerful method of multiplexing in LFAs. This retains the simplicity of a conventional double-line LFA, in terms of both fabrication and readout of results. Again, multiplexing in one TL can be coupled to printing several TLs to exponentially increase the multiplexing capability of an LFA. To perform multiplexed detection in a single TL, several labels with easily differentiated signals must be used. For instance, Yen et al. utilized the size-dependent optical properties of AgNPs to develop a multiplexed LFA for the detection of the dengue virus (DENV) NS1 protein, the yellow fever virus (YFV) NS1 protein, and an Ebola virus (ZEBOV) glycoprotein. The three monoclonal capture antibodies were deposited in equal concentrations on a single TL, and the conjugated pad was loaded with a mixture of orange, red, and green AgNPs, each functionalized for a separate target analyte. The presence of the three proteins in the sample turned the TL brown. The assay was shown to be quantifiable using RGB analysis with a smartphone camera.[135] Similarly, Zhang et al. have developed a multiplexing strategy for the detection of three cardiac disease biomarkers for the early detection of acute myocardial infarctions [creatine kinase-MB isoenzymes (CK-MB), cardiac troponin I (cTnI), and myoglobin] in a single TL, using SERS signal transduction (Figure ). In this work antibody-functionalized silver core and gold shell NPs loaded with three different Raman dyes [methylene blue (MB), Nile blue A (NBA), and rhodamine 6G (R6 G)] were used (Figure ). The authors reported ultralow detection limits (in the low pg mL–1 range) and wide linear dynamic ranges (3 orders of magnitude) that cover the clinical ranges of the three biomarkers. This multiplexing approach is particularly useful for SERS, since Raman mapping can be time-consuming. In this work the time to result was reduced from 21 min for three consecutive TLs to 7 min for a single multiplexed test line.[136]
Figure 11

Multiplexing strategy based on the detection of CK-MB, cTnI, and myoglobin cardiac biomarkers in a single TL, using three Raman reporters. Reprinted (adapted) with permission from ref (136). Copyright 2018 Elsevier

Multiplexing strategy based on the detection of CK-MB, cTnI, and myoglobin cardiac biomarkers in a single TL, using three Raman reporters. Reprinted (adapted) with permission from ref (136). Copyright 2018 Elsevier

Evaluation of Complex Samples

One of the biggest challenges in LFA development is to achieve high analytical capabilities in complex real-world samples. This is a fundamental issue that must be resolved to upgrade the technology readiness level (TRL) to more than 4 (see section ).[137] Most real-world samples contain a multitude of components that can rapidly and drastically vary in composition; these variations alter the physicochemical properties of the assay.[138] For example, the pH and ionic strength of the sample can affect the stability and binding constant of the bioreceptors to the target analyte.[139] Therefore, any fluctuations of these in the sample can vastly affect the analyte capture and therefore assay quantification. The optimal working conditions for most antibodies are at pH 7.4[15,140,141] and ionic strengths of around 30 mM (although low ionic strength might also contribute to nonspecific protein–protein interactions).[142,143] Nucleic-acid-based probes have an optimal working pH of 7.5–8.5[144] and show better affinity at higher salt concentrations (>100 mM).[145−151] It is important to know the pH and ionic strength of the media in which the LFA will be run as these can vary drastically, not just between media but also in a single media over time. For instance, saliva has a pH of 7.3–7.4 and ionic strength of 0.15 M,[152] whereas sweat has a much lower pH of between 4 and 6 and an ionic strength that varies greatly with sweat rate, location on the body, and evaporation.[153,154] One method to overcome this is to adjust the pH and ionic strength of the samples to the optimal assay conditions before performing the assay. This is done by drying a buffer into the sample pad of the LFA. There are physical as well as chemical properties of the sample that can affect the LFA; for example, the sample viscosity will affect the flow rate and in extreme conditions can even clog the nitrocellulose pores and prevent the sample from reaching the detection zone.[81] Likewise the opacity of certain real samples such as whole blood[16] can mask the optical signal generated at the TL. Common methods to overcome this include the incorporation of sample dilution or washing steps,[156] often including anionic surfactants such as Tween-20[157] and sodium dodecyl sulfate.[158] The analytical sensitivity and specificity of LFA are frequently impaired when evaluating real samples due to the presence of analytes that are similar to the biomarker of interest. It is therefore vital to identify a bioreceptor that is highly specific to the biomarker of interest.[159] Any assay is only going to be as good as the bioreceptor it uses. Assay performance can be improved by the incorporation of sample pretreatment methods such as analyte preconcentration (e.g., magnetic pre-enrichment[160] and nucleic acid amplification[151,161]) or filtering.[162] However, these usually increase the number of steps, reagents, ease-of-use, and time to result of the assay. It is important to optimize all these conditions, especially when trying to improve the sensitivity of LFAs for the analysis of low-concentration biomarkers such as tau proteins (Alzheimer disease),[163] soluble CD25 (hepatocellular carcinoma),[164] and tumor necrosis factor-α (obstructive sleep apnea syndrome)[165] that are present at trace levels (pg mL–1 range). In this section of the Review, we will discuss recent advancements and novel methodologies to better the analytical performance of LFAs in complex media.

Sample Filtration

One of the most straightforward ways to purify the sample and preconcentrate the target analyte is by using external or integrated filtration membranes. At the time of writing, 0.45 μm pore diameter cellulose nitrate filters are routinely used for the preconcentration of bacteria in water samples.[166] Integrated filters are also commonly used in LFAs that analyze whole blood samples, primarily for the removal of red blood cells. Two commercially available plasma separation membranes are Cytosep and Vivid plasma;[162,167] examples for their use in LFAs include the detection of the capsid protein VP72 from the African swine fever virus[168] and the diagnosis of SARS-CoV-2[169] and antihuman immunodeficiency virus (HIV).[170] One of the drawbacks to these types of filters is that the small pore sizes of the membranes (<9 μm) can promote blood clotting (Figure Ai), which limits the plasma/serum separation yield to just 10–30%.[171] To this end, Lu et al. have developed a low-cost (<$0.10) cross-flow filtration pad that increases the separation yield by up to 86%. This method requires three paper-based membranes stacked on top of each other. The top “differentiation pad” (40 μm pore size) is responsible for preventing the formation of the blood fouling layer. The middle “filtration pad” (1.9 μm pore size) is responsible for filtering out the red blood cells, and the final “calibration” membrane (also 1.9 μm pore size) is used to store the sample (liquid holding capacity of 40 μL cm–2; Figure Aii). The authors have applied this approach for the multiplexed detection of CRP, RBP, and ferritin, achieving similar analytical accuracy as standard centrifuge-based serum separation methods.[171]
Figure 12

(A, i) Schematic representation of red blood cells clogging the filtration membrane. (A, ii) The use of the differential pad avoids the formation of the blood fouling layer, while the filtration and calibration pads are responsible for the separation and storage of the serum sample, respectively. Reprinted (adapted) with permission from ref (171). Copyright 2018 The Royal Society of Chemistry. (B, i) Schematic representation of the integrated PCR amplicon purification approach based on the immobilization of functionalized graphene oxide on the sample pad. (B, ii) The three-layered GO sample pad enables the scrubbing of almost all the residual primers and primer-dimers with <40 pb out of the sample matrix. Reprinted (adapted) with permission from ref (172). Copyright 2017 American Chemical Society.

(A, i) Schematic representation of red blood cells clogging the filtration membrane. (A, ii) The use of the differential pad avoids the formation of the blood fouling layer, while the filtration and calibration pads are responsible for the separation and storage of the serum sample, respectively. Reprinted (adapted) with permission from ref (171). Copyright 2018 The Royal Society of Chemistry. (B, i) Schematic representation of the integrated PCR amplicon purification approach based on the immobilization of functionalized graphene oxide on the sample pad. (B, ii) The three-layered GO sample pad enables the scrubbing of almost all the residual primers and primer-dimers with <40 pb out of the sample matrix. Reprinted (adapted) with permission from ref (172). Copyright 2017 American Chemical Society. In certain situations, it is known what molecules will be present that need to be removed in order to improve the assay performance. In such situations, the filters can be functionalized to specifically remove known contaminants. One such example of sample scrubbing has been shown by Li et al., who have reported an innovative strategy to purify PCR amplicons directly in the LFA strip. They integrate high-surface-area graphene oxide (GO) functionalized for the specific capture of their PCR primers into the sample pad of their LFA (Figure Bi). They report the adsorption of almost all the residual primers, and primer-dimers with less than 40 bp, within the sample pad, while their 106 bp amplicons were able to flow to the test line (Figure Bii). With this approach, the authors achieved a 1000-fold sensitivity enhancement for the detection of bacteriophage λ-DNA, obtaining an LoD of 30 copies by naked eye.[172]

Sample Pretreatments Based on Active yet Automatic Manipulation of the Sample

Sample preconcentration can also be performed by the integration of physical separation techniques in LFA. For example, Tang et al. have integrated dialysis into an LFA for the preconcentration of the sample. They incorporate a semipermeable membrane (MWCO 3.5 KD) onto a fiberglass sample pad that has been pretreated with PEG. The highly hygroscopic nature of the PEGylated fiberglass allows it to act as a dialysate in this system (Figure A). With this approach, the authors report that a 10 min preconcentration resulted in a 10-fold and 4-fold signal enhancement in the detection of HIV specific nucleic acid sequences and myoglobin, respectively. This strategy shows great promise for PoC applications given its simplicity, speed, and low cost; however, the approach still needs to be validated with real-world samples.[173]
Figure 13

(A) Schematic representation of a 10 min preconcentration approach based on the incorporation of a dialysis membrane onto the sample pad of an LFA strip. PEG conjugated to a fiberglass membrane is used as the dialysate. Reprinted (adapted) with permission from ref (173). Copyright 2016 Elsevier. (B) Schematic representation of a paper electrophoretic device powered by a smartphone battery. The whole blood sample is driven by electrophoresis toward the detection zone, avoiding the clogging on the nitrocellulose pores. Reprinted (adapted) with permission from ref (71). Copyright 2021 American Chemical Society.

(A) Schematic representation of a 10 min preconcentration approach based on the incorporation of a dialysis membrane onto the sample pad of an LFA strip. PEG conjugated to a fiberglass membrane is used as the dialysate. Reprinted (adapted) with permission from ref (173). Copyright 2016 Elsevier. (B) Schematic representation of a paper electrophoretic device powered by a smartphone battery. The whole blood sample is driven by electrophoresis toward the detection zone, avoiding the clogging on the nitrocellulose pores. Reprinted (adapted) with permission from ref (71). Copyright 2021 American Chemical Society. Moghadam et al. have used isotachophoresis in LFAs for integrated sample extraction and preconcentration. Isotachophoresis is a nonlinear, equipment-free, electrophoretic technique where leading (LE) and trailing (TE) electrolytes are used to focus the analyte in a reduced area. This technique was found to improve the binding affinity between human IgG (HIgG) and anti-HIgG resulting in a 1 order of magnitude increase in the concentration of analyte bound at the TL and a 400-fold improvement in the LoD of the assay, when compared to the conventional LFA. Additionally, the authors show that isotachophoresis improved analyte extraction from the sample matrix; in their case, up to 80% of IgG could be extracted from real human serum samples.[174,175] On a similar note, Sena et al. have developed a portable and low-cost paper-based electrophoretic bioassay (PEB) that enables the detection of HIgG in neat whole blood samples. A potential of 200 V was applied across the nitrocellulose test strip using a smartphone in combination with a joule thief inverter and rectifier, which is compatible with the use of this assay in PoC scenarios where resources are limited. The electrophoresis drives the whole blood sample to the detection zone, avoiding the clogging of the nitrocellulose pores. Any excess blood can then be electrophoretically driven off of the detection zone allowing a clear visual readout that can be easily quantified using a smartphone for image capture and analysis (Figure B). This approach allowed for the assay sensitivity in whole blood to reach the same levels of those obtained in the lab with PBS. These methodologies are new and require further optimization and analysis but clearly have the potential for use in future LFAs in complex media.

Bioreceptor Engineering

A more sophisticated way to improve the analytical capability of LFA when evaluating complex samples is to improve the analytical properties of the bioreceptors. Traditional bioreceptors include DNA fragments and antibodies, but in recent years, novel bioreceptors such as aptamers and antibody fragments have garnered much interest. Aptamers are nucleic acid sequences (normally DNA or RNA) that fold into three-dimensional conformations that allow them to specifically bind to analytes of interest.[176] Aptamer selection is performed through a controlled in vitro process (using the Sequential Evolution of Ligands by Exponential Enrichment, SELEX) that allows not only for nonselective sequences to be discarded but also for those sequences that do bind to the analyte of interest to be ranked and selected based on their binding affinity.[177] While the binding constants tend to be similar to those of antibodies, aptamers generally produce a more specific response. There are three main reasons for this: (i) They have lower molecular weights than antibodies, decreasing the chances of nonspecific interactions. (ii) Being nucleic acids, their interaction with other bioreceptors is lower. (iii) During the SELEX process, it is possible to include a counter-selection step, which actively removes all aptamer sequences that nonspecifically bind to possible contaminants. Huan et al. have recently reviewed the application of aptamers in LFAs, focusing on the combination of these bioreceptors with signal amplification and multiplexing strategies.[178] Similar to aptamers, antibody fragments and nanobodies can also improve the specificity of LFA when employing complex samples. Antibody fragments are generated by the proteolytic enzymatic fragmentation of full size antibodies (mostly IgG). In this way it is possible to remove the Fc region of IgG, which is usually responsible of nonspecific interactions with endogenous Fc receptors. Additionally, the presence of less charged groups reduces the possibility of cross-reactivity.[19] However, to the best of our knowledge there are still no reports proving the aforementioned analytical capabilities of antibody fragments when evaluating complex samples in LFA. On the contrary, nanobodies (or single domain antibodies, sdAb) are the single chain binding portions of camelid antibodies. Thanks to their low molecular weight (<15 kDa), easy fabrication (being a single chain of amino acid, they can be produced by E. coliin vitro), and stability, they seem like an ideal bioreceptor for LFA. In this context, few works have employed them to enhance the LFA capabilities. Between them, the work from Loynachan et al. combined nanobodies with platinum nanocatalysts to achieve the naked-eye detection of p24 spiked into sera in the low femtomolar range, as well as the detection of acute-phase HIV in clinical human plasma samples.[179] The implementation of the CRISPR/Cas recognition systems in LFAs has been reported to provide both high specificity and sensitivity.[180] This method has recently been utilized for the dual detection of the envelope (E) and open reading frame 1ab (Orf1ab) genes of SARS-CoV-2 in nasopharyngeal swab samples by Xiong et al. The use of CRISPR/Cas9 with multiplexed RT-RPA enabled the amplification of the target genes without the need for external instrumentation. Cas9/sgRNA complex formation eliminates interference from the residual primer-dimers, improving the assay specificity. The authors report a high clinical sensitivity (97%) and specificity (100%) after evaluating 64 real patient samples.[181] Nguyen et al. have developed an interesting platform that utilizes isothermal amplification and Cas12a SHERLOCK for the detection of exhaled SARS-COV-2 on a face mask. The aerosol collection and nucleic acid amplification are performed on the interior surface of the face mask, while the detection step is done by the integration of a dedicated LFA (Figure ). This approach showed high selectivity over other human coronavirus strains and had an LoD of 500 RNA copies.[182]
Figure 14

(A) Schematic representation of the face mask-LFA capable of detecting exhaled SARS-CoV-2. Once the water reservoir is pressed, the water flow moves the collected aerosol sample toward the paper μPAD, which contains the freeze-dried assay reagents. (B) First, the RNA is extracted, and then, it is amplified by RT-RPA and detected by the Cas12a SHERLOCK method. The output is visualized in an LFA strip that is built into the face mask. (C) Pictures of the face mask interior and exterior. Reprinted (adapted) with permission from ref (182). Copyright 2021 Springer Nature.

(A) Schematic representation of the face mask-LFA capable of detecting exhaled SARS-CoV-2. Once the water reservoir is pressed, the water flow moves the collected aerosol sample toward the paper μPAD, which contains the freeze-dried assay reagents. (B) First, the RNA is extracted, and then, it is amplified by RT-RPA and detected by the Cas12a SHERLOCK method. The output is visualized in an LFA strip that is built into the face mask. (C) Pictures of the face mask interior and exterior. Reprinted (adapted) with permission from ref (182). Copyright 2021 Springer Nature.

LFA Validation and Barriers toward Commercialization

The LFA market has seen a fascinating growth in recent decades, from USD 2.2 billion in 2005 to USD 8.2 billion in 2020, which has been driven by the high prevalence of infectious diseases and the increasing popularity of home-diagnosis.[183] The outbreak of the COVID-19 pandemic caused a boom in the market by rapidly increasing the demand for at-home (PoC) test kits.[184] Despite their lower accuracy compared to RT-qPCR, LFAs have still proven popular due to their point-of-care properties. Given this, the LFA market size is expected to reach USD 10.2 billion by 2025.[185] As with any technology, one method of estimating how close a particular LFA is to market is by assessing its TRL level. This is a system used to assess the maturity of a particular technology.[137] For the different stages of LFA development, TRL-1 would correspond to the achievement of a measurable signal in the TL upon the detection of a target analyte using a half-stick format; TRL-2 to the identification of the ideal components to build the full LFA; TRL-3 to the determination of the LFA properties (sensitivity, specificity, dynamic range, RSD, etc.) after the evaluation of serial dilutions of the analyte in buffer in the laboratory; TRL-4 to the determination of the LFA accuracy when detecting the analyte in real samples (prevalidated using gold standard methods) in the laboratory; TRL-5 to the validation of the LFA with prevalidated real samples in the hospital with trained users; TRL-6 to the validation of the LFA with clinical samples in the hospital with trained personnel; TRL-7 to the validation of the LFA in the hospital using clinical samples and untrained personnel; TRL-8 to the assessment of the commercial LFA in deployment conditions, proving full compliance of the quality standards and regulations; and TRL-9 to the achievement of a fully operational LFA, ready for commercialization, with the system optimized for full rate production. Most of the technology (with a few exceptions) discussed in sections –5 of this Review are at TRL levels 1–5. Commercially available LFAs are relatively simple devices not utilizing the technological advancements being reported in research laboratories. This begs the question of what barriers are preventing the commercial adoption of these technological advances? One possible reason is that research centers and small companies suffer a funding gap between TRLs 4 and 6, also known as the “valley of death”.[186] This is the point in assay development when tests move away from in-the-lab measurements with ideal conditions and simulated samples (TRL 4) to validation with real-world samples at the intended point of use (TRLs 5 and 6). Most investors only become interested when technologies reach TRLs 7–9.[187] In an attempt to correct this issue, funding organizations (e.g., the European Commission) and governments have started to increase funding to collaborative projects between academia and industry.[188−190] The involvement of industrial partners throughout the entire LFA development process is intended to encourage stricter critical evaluation of LFA cost, performance, manufacturing capability, and scalability, properties that are often overlooked in academic research. In addition to this, the involvement of industrial partners as early as possible is intended to speed up the time to market of LFAs.[191−193] Another important barrier toward commercialization is to obtain all the regulatory approval required for product launch. This is a tedious and expensive process that starts with the analytical and clinical validation of the test. The analytical validation assesses the precision, accuracy, and reliability of the test, while the clinical validation evaluates the parameters such as the diagnostic sensitivity, specificity, positive and negative predictive value, and likelihood ratio of the LFA.[194] There are several companies offering support services to diagnostic companies during this process.[195,196] For goods intended to be sold in the European Union (EU), a CE mark is required. In order to obtain this certification, LFA manufacturers must submit a performance evaluation report to the regulatory agency showing the analytical and clinical validation of their LFA. The in vitro diagnostic medical devices regulation (IVDR) (EU, 2017/746) classifies devices in relation to their risk if inaccurate; in other words, if the test gives a false negative, what risk is posed to the general population. Tests are assigned a letter of the alphabet (A, B, C, or D) based on their score in this test. Diagnostics for the detection of pregnancy, cholesterol level, glucose, or bacteria in urine are classified as moderate health risks (B), while assays for transmissible agents with a high risk of propagation (e.g., SARS-CoV-2) are classified as D. The next step is a conformity assessment, in which an independent certified body evaluates the analytical and clinical performance of the test.[197] In America, the Food and Drug Administration (FDA) has a similar regulation, where the tests are classified as I, II, or III depending on their health risk in case of inaccuracy. Most LFAs are class II and must undergo a stringent premarket review (only class I is exempt), where their analytical and clinical performance is evaluated and compared with equivalent tests or with a standardized reference method.[198] The European and American regulators are the strictest and hardest to obtain. If a test is developed that passes these regulations, then they will normally be accepted worldwide. The main obstacles to commercialization are therefore not necessarily scientific issues, but rather difficulties in raising capital for low TRL level technologies and the high costs (both financial and in terms of time) and complexity of obtaining regulatory approval.

Outlook

In this paper, we have reviewed the advances in LFA from the last ten years with an emphasis on the role of nanomaterials. We have identified four main areas that should simultaneously be improved to achieve the next generation of LFA: sensitivity enhancement, multiplexing, quantification at the PoC, and evaluation of complex samples. These improvements need to retain the key advantageous properties of LFAs, namely, their low cost, rapid time to result, and ease of operation interpretation. Notable advances have been made that increase the sensitivity of LFAs. These are important as assays are being developed for low-concentration analytes, increasing the utility of LFAs. Sensitivity can be enhanced using novel reporters, signal amplifiers, and alternative signal transduction methods. Likewise, advances in multiplexing methodologies have been reviewed and are sure to form a large part of future research and commercial LFA products. Technologies that allow the detection of multiple analytes in a single TL are perhaps the most interesting as they allow the exponential increase in the number of analytes that can be detected in an LFA, since they can be combined with novel device designs that allow multiple target analytes to be investigated in a single test. Likewise, microarrays optimize the available test area for the detection of a multitude of analytes in a single detection zone. While there has been a miniaturization of LFA readers in recent years to analyze and quantify the signal on test strips, the past decade has seen a real trend in the use of smartphones for LFA analysis. This often requires portable boxes or cassettes to control the lighting conditions of image acquisition with a smartphone camera. The advent of portable 3D printers is allowing the fabrication of many of these at the point of use. In the future the increasing power of smartphones and the advent of 5G networking promise to further improve PoC LFAs, in terms of both their analytical performance and their data analysis. The development of point-of-care readers that enable PoC analysis using alternative detection methods (such as SERS, TCA, and MPQ) is still in its infancy but undoubtedly has potential. We anticipate that the next decade will see an increase in research to reduce the cost and improve the performance of such readers for LFA applications. The simplicity of use of LFAs is what makes them so desirable; it is therefore encouraging to see the work by researchers to incorporate sample pretreatment methods into LFAs. Any technological developments aimed at allowing the analysis of untreated sample matrices is of great interest to the entire field of LFA development. The future of LFAs is undoubtedly a “closed box” system for the end user where they add their sample and get a simple, reliable, and quantified readout. Such devices will most likely include combinations of the technologies reviewed herein. The practice of combining technologies is rare, with most publications understandably focusing on a single, novel technological advancement. The combination of several technological advancements in a single LFA has the potential to make a greater societal impact and should be given encouragement. Throughout this Review, we have seen that there are various techniques that have been employed to address the four main challenges in the field: quantification at the PoC, sensitivity enhancement, multiplexing, and the evaluation of complex samples. While all of the techniques investigated require more development, it is worth considering which are currently the most promising. This is a largely subjective exercise, but the authors would suggest that, of all the techniques covered, perhaps the use of SERS in LFAs is the most advanced and currently the best able to solve the four main challenges highlighted in this Review. We have seen that portable SERS readers are being successfully developed for on-site quantification of signals in LFAs;[59] there are improved sensitivities when using SERS compared to colorimetric detection, with LoDs as low as pg/mL being reported.[123] Different SERS nanotags that emit different SERS signals have been employed for multiplexing,[136] and while SERS is naturally more suited to complex media analysis than colorimetric or fluorescent measurements, the ability to couple SERS labels to magnetic nanoparticles has also facilitated complex media analysis.[103] The greatest challenges in SERS-based LFA development concern the readers, which need much more development to be practically usable. Current limitations include high costs, long data acquisition times, and not being very user-friendly. The authors believe work should be conducted to increase the scan speed and software developed to give simple-to-understand readouts. Unlike optical readout methods, the results of SERS are not intuitively understood, so software that can accurately and reliably process, interpret, and report the results in an easy-to-understand fashion is required for real-world applications where the assays will be performed by untrained personnel. Advances in the other technologies discussed in this Review (such as TCA) show that they are very promising and developing well; however, at the time of writing, they are not able to solve all four of the issues discussed (multiplexing with TCA has not been demonstrated yet). Further development of these techniques to address these issues is required, before integrating those developments into a single sensing model, which means they are some years behind SERS. Similarly, the recently published work by Miller et al. also achieves quantification and ultralow sensitivity;[84] however, it would probably require extensive sample treatment, and its ability to detect multiple targets would require the printing of multiple detection zones. There is an obvious trend in the literature to develop new methods of signal transduction. There have been decades of research and development of optical (colorimetric and fluorescent) LFAs, and their limitations seem to have been reached. They lack the sensitivity and reliability to detect the biomarkers of interest for future sensing applications, currently only being able to detect highly concentrated biomarkers. Different types of energy (sound and thermal) and forces (magnetism) are being used to replace light in LFAs; we envisage that in the future more development of these and the use of other signal transduction methods will be prominent in LFA development.
  143 in total

1.  Analysis of quantum dot fluorescence stability in primary blood mononuclear cells.

Authors:  Huw D Summers; Mark D Holton; Paul Rees; Paul M Williams; Catherine A Thornton
Journal:  Cytometry A       Date:  2010-10       Impact factor: 4.355

2.  Validity of a single antibody-based lateral flow immunoassay depending on graphene oxide for highly sensitive determination of E. coli O157:H7 in minced beef and river water.

Authors:  Abdelrahim Hussein Abdelazeem Hassan; José Francisco Bergua; Eden Morales-Narváez; Arben Mekoçi
Journal:  Food Chem       Date:  2019-06-08       Impact factor: 7.514

3.  Impedimetric measurement of DNA-DNA hybridisation using microelectrodes with different radii for detection of methicillin resistant Staphylococcus aureus (MRSA).

Authors:  Poh Quan Li; Andrew Piper; Ilka Schmueser; Andrew R Mount; Damion K Corrigan
Journal:  Analyst       Date:  2017-05-30       Impact factor: 4.616

Review 4.  CRISPR/Cas Systems towards Next-Generation Biosensing.

Authors:  Yi Li; Shiyuan Li; Jin Wang; Guozhen Liu
Journal:  Trends Biotechnol       Date:  2019-01-14       Impact factor: 19.536

5.  Electrospin-coating of nitrocellulose membrane enhances sensitivity in nucleic acid-based lateral flow assay.

Authors:  Chee-Hong Takahiro Yew; Pedram Azari; Jane Ru Choi; Fei Li; Belinda Pingguan-Murphy
Journal:  Anal Chim Acta       Date:  2018-01-24       Impact factor: 6.558

6.  Platinum-Decorated Gold Nanoparticles with Dual Functionalities for Ultrasensitive Colorimetric in Vitro Diagnostics.

Authors:  Zhuangqiang Gao; Haihang Ye; Dianyong Tang; Jing Tao; Sanaz Habibi; Adrienne Minerick; Dianping Tang; Xiaohu Xia
Journal:  Nano Lett       Date:  2017-08-18       Impact factor: 11.189

7.  Enhancing the Sensitivity of Lateral Flow Immunoassay by Magnetic Enrichment Using Multifunctional Nanocomposite Probes.

Authors:  The Son Le; Sizun He; Mari Takahashi; Yasushi Enomoto; Yasufumi Matsumura; Shinya Maenosono
Journal:  Langmuir       Date:  2021-05-19       Impact factor: 3.882

8.  A SERS-based lateral flow assay biosensor for highly sensitive detection of HIV-1 DNA.

Authors:  Xiuli Fu; Ziyi Cheng; Jimin Yu; Priscilla Choo; Lingxin Chen; Jaebum Choo
Journal:  Biosens Bioelectron       Date:  2015-12-02       Impact factor: 10.618

9.  Development of a recombinase polymerase based isothermal amplification combined with lateral flow assay (HLB-RPA-LFA) for rapid detection of "Candidatus Liberibacter asiaticus".

Authors:  Dilip Kumar Ghosh; Sunil B Kokane; Amol D Kokane; Ashish J Warghane; Manali R Motghare; Sumit Bhose; Ashwani Kumar Sharma; M Krishna Reddy
Journal:  PLoS One       Date:  2018-12-12       Impact factor: 3.240

10.  A DNA-Based Biosensor Assay for the Kinetic Characterization of Ion-Dependent Aptamer Folding and Protein Binding.

Authors:  Irene Ponzo; Friederike M Möller; Herwin Daub; Nena Matscheko
Journal:  Molecules       Date:  2019-08-08       Impact factor: 4.411

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.