| Literature DB >> 27493680 |
Henry Joutsijoki1, Markus Haponen2, Jyrki Rasku1, Katriina Aalto-Setälä3, Martti Juhola1.
Abstract
The focus of this research is on automated identification of the quality of human induced pluripotent stem cell (iPSC) colony images. iPS cell technology is a contemporary method by which the patient's cells are reprogrammed back to stem cells and are differentiated to any cell type wanted. iPS cell technology will be used in future to patient specific drug screening, disease modeling, and tissue repairing, for instance. However, there are technical challenges before iPS cell technology can be used in practice and one of them is quality control of growing iPSC colonies which is currently done manually but is unfeasible solution in large-scale cultures. The monitoring problem returns to image analysis and classification problem. In this paper, we tackle this problem using machine learning methods such as multiclass Support Vector Machines and several baseline methods together with Scaled Invariant Feature Transformation based features. We perform over 80 test arrangements and do a thorough parameter value search. The best accuracy (62.4%) for classification was obtained by using a k-NN classifier showing improved accuracy compared to earlier studies.Entities:
Mesh:
Year: 2016 PMID: 27493680 PMCID: PMC4963598 DOI: 10.1155/2016/3091039
Source DB: PubMed Journal: Comput Math Methods Med ISSN: 1748-670X Impact factor: 2.238
Figure 1Structure 1 for automated quality identification of human iPSC colony images.
Figure 2Structure 2 for automated quality identification of human iPSC colony images.
Figure 3Structure 3 for automated quality identification of human iPSC colony images.
Figure 4Structure 4 for automated quality identification of human iPSC colony images.
Figure 5Structure 5 for automated quality identification of human iPSC colony images.
Figure 6Structure 6 for automated quality identification of human iPSC colony images.
Figure 7Example images of good, semigood, and bad quality iPSC colonies. The images have been scaled to have width and height of 1.5 in.
Frequencies and percentages of classes in dataset.
| Class | Frequencies | Percentages |
|---|---|---|
| Good | 74 | 42.8% |
| Semigood | 58 | 33.5% |
| Bad | 41 | 23.7% |
|
| ||
| Total | 173 | 100.0% |
Results of one-versus-one and one-versus-all methods when different kernels were used. True positive rates (%) are given in parentheses and accuracy (%) can be found from the last column.
| Kernel/class | Bad | Good | Semigood | ACC |
|---|---|---|---|---|
| One-versus-all | ||||
| Linear | 22 (53.7%) |
| 18 (31.0%) | 52.6% |
| Polynomial | 23 (56.1%) | 50 (67.6%) | 26 (44.8%) | 57.2% |
| Polynomial | 18 (43.9%) | 43 (58.1%) | 22 (37.9%) | 48.0% |
| Polynomial | 23 (56.1%) | 39 (52.7%) | 20 (34.5%) | 47.4% |
| Polynomial |
| 34 (45.9%) | 17 (29.3%) | 43.9% |
| Polynomial | 23 (56.1%) | 33 (44.6%) | 18 (31.0%) | 42.8% |
| RBF |
| 50 (67.6%) |
|
|
|
| ||||
| One-versus-one | ||||
| Linear |
|
|
|
|
| Polynomial | 21 (51.2%) | 46 (62.2%) | 15 (25.9%) | 47.4% |
| Polynomial | 20 (48.8%) | 45 (60.8%) | 24 (41.4%) | 51.4% |
| Polynomial | 14 (34.1%) | 41 (55.4%) | 21 (36.2%) | 43.9% |
| Polynomial | 24 (58.5%) | 32 (43.2%) | 21 (36.2%) | 44.5% |
| Polynomial | 19 (46.3%) | 39 (52.7%) | 17 (29.3%) | 43.4% |
| RBF | 27 (65.9%) | 50 (67.6%) | 24 (41.4%) | 58.4% |
Results of structures 1–3 given in Figures 1 –3 when different kernels were used. True positive rates (%) are given in parentheses and accuracy (%) can be found from the last column.
| Kernel/class | Bad | Good | Semigood | ACC |
|---|---|---|---|---|
| Structure 1 | ||||
| Linear |
| 52 (70.3%) | 24 (41.4%) | 60.7% |
| Polynomial | 21 (51.2%) | 42 (56.8%) | 19 (32.8%) | 47.4% |
| Polynomial | 20 (48.8%) | 40 (54.1%) |
| 50.3% |
| Polynomial | 14 (34.1%) | 38 (51.4%) | 26 (44.8%) | 45.1% |
| Polynomial | 19 (46.3%) | 26 (35.1%) | 26 (44.8%) | 41.0% |
| Polynomial | 15 (36.6%) | 26 (35.1%) | 24 (41.4%) | 37.6% |
| RBF | 28 (68.3%) | 51 (68.9%) | 24 (41.4%) | 59.5% |
|
| ||||
| Structure 2 | ||||
| Linear |
|
| 24 (41.4%) |
|
| Polynomial | 20 (48.8%) | 43 (58.1%) | 16 (27.6%) | 45.7% |
| Polynomial | 20 (48.8%) | 46 (62.2%) | 23 (39.7%) | 51.4% |
| Polynomial | 13 (31.7%) | 44 (59.5%) | 15 (25.9%) | 41.6% |
| Polynomial | 19 (46.3%) | 37 (50.0%) | 18 (31.0%) | 42.8% |
| Polynomial | 15 (36.6%) | 43 (58.1%) | 12 (20.7%) | 40.5% |
| RBF |
| 53 (71.6%) | 24 (41.4%) | 61.3% |
|
| ||||
| Structure 3 | ||||
| Linear | 27 (65.9%) | 52 (70.3%) | 23 (39.7%) | 59.0% |
| Polynomial | 22 (53.7%) | 46 (62.2%) | 16 (27.6%) | 48.6% |
| Polynomial | 20 (48.8%) | 40 (54.1%) | 24 (41.4%) | 48.6% |
| Polynomial | 19 (46.3%) | 38 (51.4%) | 17 (29.3%) | 42.8% |
| Polynomial | 26 (63.4%) | 26 (35.1%) | 18 (31.0%) | 40.5% |
| Polynomial | 22 (53.7%) | 26 (35.1%) | 12 (20.7%) | 34.7% |
| RBF | 25 (61.0%) | 49 (66.2%) | 21 (36.2%) | 54.9% |
Results of structures 4–6 given in Figures 4 –6 when different kernels were used. True positive rates (%) are given in parentheses and accuracy (%) can be found from the last column.
| Kernel/class | Bad | Good | Semigood | ACC |
|---|---|---|---|---|
| Structure 4 | ||||
| Linear |
| 52 (70.3%) | 19 (32.8%) | 57.2% |
| Polynomial | 25 (61.0%) | 45 (60.8%) | 19 (32.8%) | 51.4% |
| Polynomial | 18 (43.9%) | 42 (56.8%) | 19 (32.8%) | 45.7% |
| Polynomial | 23 (56.1%) | 35 (47.3%) | 19 (32.8%) | 44.5% |
| Polynomial | 23 (56.1%) | 28 (37.8%) | 14 (24.1%) | 37.6% |
| Polynomial | 23 (56.1%) | 24 (32.4%) | 15 (25.9%) | 35.8% |
| RBF | 22 (53.7%) | 48 (64.9%) | 23 (39.7%) | 53.8% |
|
| ||||
| Structure 5 | ||||
| Linear | 24 (58.5%) |
| 21 (36.2%) |
|
| Polynomial | 18 (43.9%) | 44 (59.5%) | 24 (41.4%) | 49.7% |
| Polynomial | 20 (48.8%) | 46 (62.2%) |
| 53.2% |
| Polynomial | 16 (39.0%) | 41 (55.4%) | 23 (39.7%) | 46.2% |
| Polynomial | 19 (46.3%) | 37 (50.0%) | 21 (36.2%) | 44.5% |
| Polynomial | 15 (36.6%) | 38 (51.4%) | 20 (34.5%) | 42.2% |
| RBF | 20 (48.8%) | 52 (70.3%) | 21 (36.2%) | 53.8% |
|
| ||||
| Structure 6 | ||||
| Linear | 20 (48.8%) | 38 (51.4%) | 24 (41.4%) | 47.4% |
| Polynomial | 15 (36.6%) | 38 (51.4%) |
| 45.7% |
| Polynomial | 18 (43.9%) | 34 (45.9%) | 20 (34.5%) | 41.6% |
| Polynomial | 17 (41.5%) | 37 (50.0%) | 20 (34.5%) | 42.8% |
| Polynomial | 21 (51.2%) | 28 (37.8%) | 20 (34.5%) | 39.9% |
| Polynomial | 18 (43.9%) | 22 (29.7%) | 23 (39.7%) | 36.4% |
| RBF | 19 (46.3%) | 43 (58.1%) | 22 (37.9%) | 48.6% |
Results of classification tree, linear discriminant analysis, multinomial logistic regression, and naïve Bayes variants. True positive rates (%) are given in parentheses and accuracy (%) can be found from the last column.
| Method/class | Bad | Good | Semigood | ACC |
|---|---|---|---|---|
| Classification tree |
| 50 (67.6%) |
| 51.4% |
| Linear discriminant analysis | 19 (46.3%) | 35 (47.3%) | 16 (27.6%) | 40.5% |
| Multinomial logistic regression | 17 (41.5%) | 32 (43.2%) |
| 39.3% |
| Naïve Bayes | 16 (39.0%) |
| 14 (24.1%) |
|
| Naïve Bayes with kernel smoothing density estimation and normal kernel | 18 (43.9%) | 59 (79.7%) | 14 (24.1%) |
|
| Naïve Bayes with kernel smoothing density estimation and box kernel | 12 (29.3%) | 56 (75.7%) | 11 (19.0%) | 45.7% |
| Naïve Bayes with kernel smoothing density estimation and Epanechnikov kernel | 13 (31.7%) | 57 (77.0%) | 11 (19.0%) | 46.8% |
| Naïve Bayes with kernel smoothing density estimation and triangle kernel | 13 (31.7%) | 56 (75.7%) | 12 (20.7%) | 46.8% |
Results of k-NN with different weighting and measure combinations. True positive rates (%) are given in parentheses and accuracy (%) can be found from the last column.
| Measure and weighting combination/class | Bad | Good | Semigood | ACC |
|---|---|---|---|---|
| Chebyshev/equal weights | 18 (43.9%) | 52 (70.3%) | 26 (44.8%) | 55.5% |
| Chebyshev/inverse weights | 19 (46.3%) | 53 (71.6%) | 27 (46.6%) | 57.2% |
| Chebyshev/squared inverse weights | 17 (41.5%) | 53 (71.6%) | 27 (46.6%) | 56.1% |
| Cityblock/equal weights | 22 (53.7%) | 55 (74.3%) | 24 (41.4%) | 58.4% |
| Cityblock/inverse weights | 23 (56.1%) | 52 (70.3%) | 27 (46.6%) | 59.0% |
| Cityblock/squared inverse weights | 20 (48.8%) | 51 (68.9%) | 22 (37.9%) | 53.8% |
| Correlation/equal weights | 17 (41.5%) | 59 (79.7%) | 16 (27.6%) | 53.2% |
| Correlation/inverse weights | 16 (39.0%) | 54 (73.0%) | 14 (24.1%) | 48.6% |
| Correlation/squared inverse weights | 22 (53.7%) | 53 (71.6%) | 17 (29.3%) | 53.2% |
| Cosine/equal weights | 19 (46.3%) | 57 (77.0%) | 14 (24.1%) | 52.0% |
| Cosine/inverse weights | 19 (46.3%) |
| 15 (25.9%) | 55.5% |
| Cosine/squared inverse weights | 21 (51.2%) | 50 (67.6%) | 14 (24.1%) | 49.1% |
| Euclidean/equal weights | 24 (58.5%) | 55 (74.3%) |
|
|
| Euclidean/inverse weights |
| 53 (71.6%) | 27 (46.6%) | 60.7% |
| Euclidean/squared inverse weights | 20 (48.8%) | 46 (62.2%) | 26 (44.8%) | 53.2% |
| Standardized Euclidean/equal weights | 22 (53.7%) | 54 (73.0%) | 26 (44.8%) | 59.0% |
| Standardized Euclidean/inverse weights |
| 53 (71.6%) | 27 (46.6%) | 60.7% |
| Standardized Euclidean/squared inverse weights | 20 (48.8%) | 46 (62.2%) | 26 (44.8%) | 53.2% |
| Spearman/equal weights | 15 (36.6%) | 59 (79.7%) | 19 (32.8%) | 53.8% |
| Spearman/inverse weights | 17 (41.5%) | 61 (82.4%) | 19 (32.8%) | 56.1% |
| Spearman/squared inverse weights | 17 (41.5%) | 56 (75.7%) | 17 (29.3%) | 52.0% |