| Literature DB >> 33244129 |
Alessio Lugnan1,2, Emmanuel Gooskens3,4, Jeremy Vatin5, Joni Dambre6, Peter Bienstman3,4.
Abstract
Machine learning offers promising solutions for high-throughput single-particle analysis in label-free imaging microflow cytomtery. However, the throughput of online operations such as cell sorting is often limited by the large computational cost of the image analysis while offline operations may require the storage of an exceedingly large amount of data. Moreover, the training of machine learning systems can be easily biased by slight drifts of the measurement conditions, giving rise to a significant but difficult to detect degradation of the learned operations. We propose a simple and versatile machine learning approach to perform microparticle classification at an extremely low computational cost, showing good generalization over large variations in particle position. We present proof-of-principle classification of interference patterns projected by flowing transparent PMMA microbeads with diameters of [Formula: see text] and [Formula: see text]. To this end, a simple, cheap and compact label-free microflow cytometer is employed. We also discuss in detail the detection and prevention of machine learning bias in training and testing due to slight drifts of the measurement conditions. Moreover, we investigate the implications of modifying the projected particle pattern by means of a diffraction grating, in the context of optical extreme learning machine implementations.Entities:
Year: 2020 PMID: 33244129 PMCID: PMC7691359 DOI: 10.1038/s41598-020-77765-w
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1(a) Schematic of the employed setup. A PMMA microfluidic channel (cross section ) is illuminated by a laser radiation (HeNe laser, ) focused on a pinhole. The resulting beam passes through a double axis holographic diffraction grating (only in one of the employed configurations) and is captured by a CMOS camera. (b) Schematic of the illuminated microfluidic channel region. The larger the particle distance from the field of view center, the weaker the acquired particle signal (measured by the perturbation quantity P). (c–l) Respectively for the NDG (top row) and DG (bottom row) configurations, examples of background pattern (1st column), background-subtracted particle patterns with increasing intensity (2nd to 4th columns) and class separation colormaps (last column). (e,h) are well below the respective acceptance thresholds, in this case and (for a particle ratio ). (f,i) are just above and (g,j) are well above the respective acceptance thresholds. Grey arrows suggest a qualitative link between these examples and the particle position w.r.t. the FoV shown in (b).
Figure 2Schematic of the machine learning classification pipeline. Intensity patterns are acquired by the image sensor in free-run mode. The difference between consecutive images is calculated (background subtraction), and if the squared sum of its pixels is lower than a chosen acceptance threshold value the image is considered as background and discarted. A linear classifier (trainable weighted sum) is applied to accepted background-subtracted images. If the outcome is positive, the analyzed particle is classified as belonging to class A, to class B otherwise.
Figure 3Box plots of the classification error evaluated by means of cross-validation on images down-sampled to different resolutions (x axis). Each box represents the distribution of the error values, corresponding to different folds, obtained through k-fold cross-validation. Boxes, whiskers, orange lines and green triangles respectively represent the interquartile range, the range, the median and the mean of the error values. The outliers (outer points distant more than (interquartile range) from the interquartile range) are represented by circles. The employed samples were selected among the acquired images considering a particle ratio value of (see subsection “Microbeads classification for different fields of view”). Left: the samples employed for training, validation and test were obtained from a single measurement session per class, providing misleadingly low average errors and high variance due to measurement bias. Middle: test errors are evaluated on samples from dedicated measurement sessions, showing the correct generalization capability of the trained classifier. Right: the proposed intertwined class measurements and validation algorithm were employed to remove the measurement bias influence from classification training, validation and test. The comparison with the middle box plot shows an improved generalization capability of the trained classifier.
Figure 4(a) Box plot of the classification error evaluated by the proposed validation algorithm on sample sets obtained through different choices of R (on the x axis), i.e. applying different acceptance thresholds. provides the best classification performance (low error average and variance) due to a trade-off between the field of view and the number of samples N. (b) Corresponding classification error obtained through the proposed UM test. The training is performed on uniformly mislabelled data and therefore the obtained test error is expected to be (random choice) for our two classes, if the learning is not affected by measurement bias.
Figure 5(a,b) Box plots of the classification error for evaluated on particle images of different resolution (x axis), with and without holographic double axis diffraction grating interposed between the camera and the microfluidic channel. Classification errors lower than were obtained for image resolutions down to just pixels. Generally, the interference patterns processed by the diffraction grating provide particle classification with similar or slightly higher errors. Note that the number of samples used to evaluate the first two points was further reduced by feature selection. (c,d) Particle rate R as a function of the acceptance threshold for different measurement sessions. (c) Comparison between the configuration without interposed diffraction grating (NDG, blue dots) and with diffraction grating (DG, red dots). The diffraction grating changes the relation in a nonlinear way. (d) Comparison between measurement sessions (both in NDG configuration) performed with a time distance of 3 days. The curve do not change significantly from one measurement session to another, indicating stability in our measurements.
Execution time per particle of the proposed classification algorithm for different image resolutions, evaluated on a laptop (Intel Core i5-8250U, 1.60GHz 8) using a Python script (Numpy library). The reported time values are averaged (median) over 10000 iterations of the following steps: computation of the difference between the target and the background image after conversion to float type matrices; application of the acceptance threshold to the sum of the squared elements of the difference matrix; weighted sum of the difference matrix (i.e. machine learning inference).
| Image resolution (pixels) | |||||||
|---|---|---|---|---|---|---|---|
| Classification time ( | 200 | 38 | 19 | 13 | 10 | 9.0 | 8.8 |
Comparison of machine learning-related aspects regarding three other works (reporting online label-free classification via particle imaging) and our work. CNN is the acronym for Convolutional Neural Network, while mAP is the abbreviation of mean Average Precision.
| Classification task | Classifier | Image resolution | Imaging method | Image FoV | Classification performance | Accelerator | execution time / particle | Meas. bias control |
|---|---|---|---|---|---|---|---|---|
| Beads with diameters of 7, 10 and | CNN | Microscope | Centered and cropped | GPU | Unreported | |||
| 3 white blood cell (WBC) types[ | Rand. forest on extracted features | Lens-free - raw hologram | Unreported | GPU | 0.2 ms | Unreported | ||
| 1 WBC type and an epithelial cancer cell[ | Deep CNN | Unreported | Time-stretch microscope | GPU | 3.6 ms | Unreported | ||
| Beads with diameters of 15.2 and | Linear (log. regression) | Lens-free - raw hologram | None | 0.013 ms | Yes |
Correspondence between chosen particle ratio R values (same for both particle classes), the number of images accepted as samples for classification (with strong enough particle signal) and estimated FoV of the classification process. Left and right tables regard respectively the configurations with and without a diffraction grating interposed between the microfluidic channel and the camera (NDG and DG configurations).
| Particle rate | # accepted images | Field of view (mm) | Particle rate | # accepted images | Field of view (mm) | ||||
|---|---|---|---|---|---|---|---|---|---|
| class A | class B | class A | class B | class A | class B | class A | class B | ||
| No diffractive layer | Diffraction grating | ||||||||
| 0.02 | 1427 | 2108 | 0.09 | 0.25 | 0.02 | 1416 | 2288 | 0.08 | 0.27 |
| 0.04 | 4008 | 3067 | 0.27 | 0.37 | 0.04 | 4173 | 3213 | 0.27 | 0.38 |
| 0.06 | 6452 | 4120 | 0.45 | 0.51 | 0.06 | 6826 | 4207 | 0.45 | 0.51 |
| 0.08 | 7954 | 6051 | 0.56 | 0.76 | 0.08 | 8354 | 6199 | 0.57 | 0.76 |