Mobile QR Code QR CODE

  1. (Department of Computer Engineering, Kwangwoon University / Seoul, Korea {yulisun, parkcheolsoo}@kw.ac.kr )
  2. (School of Applied Science, Telkom University / Bandung, Indonesia yulisun@telkomuniversity.ac.id )



White blood cells, Nailfold capillary, Event-based data, Classification, Dynamic vision sensor

1. Introduction

White blood cells (WBCs) are a crucial part of the human immune system. A lack of WBCs can lead to a number of illnesses, such as sepsis [1], infectious disorders [2,3], and cancer [4]. Typically, a blood sample must be drawn by experienced medical personnel for WBC monitoring using advanced technology. Numerous noninvasive methods based on optical technologies have been suggested to enhance WBC monitoring. One example is observing nailfold capillaries, where WBCs are identified as visible gaps [5]. Nailfold capillaroscopy is a non-invasive imaging method to assess micro-circulation in tiny blood vessels under the fingernail [6]. The location of nailfold capillaries in fingers and capillary visualization is shown in Fig. 1. By illuminating the capillaries at a specific wavelength, a microscope can reveal WBC flow in order to evaluate WBC status [7,8]. Moreover, absorption gaps or plasma gaps in a nailfold capillary can be used to quantify the number of WBCs.

Fig. 1. Location of nailfold capillaries (left) and visualization of the capillaries (right).
../../Resources/ieie/IEIESPC.2022.11.4.270/fig1.png
Fig. 2. Differences in output of a standard camera versus an event-based camera (adapted from[16]).
../../Resources/ieie/IEIESPC.2022.11.4.270/fig2.png

Recently, measurement of the plasma gap has usually been done with a standard camera [8,9]. However, difficulties with that method are low visibility of the gap, the small size of the capillary, the high speed of WBC movement, and non-contrast between capillaries and their environment/background. To address these problems, this study uses a dynamic vision sensor (DVS) camera, a so-called event-based camera, and bio-inspired vision sensors that work differently from conventional frame cameras to detect the change in luminance on a pixel basis and to produce a stream of asynchronous event output at a microsecond time resolution [10]. Event cameras such as the dynamic vision sensor are bio-inspired devices that work differently from conventional frame cameras. While the conventional camera captures images at a fixed rate, event cameras asynchronously measure per-pixel brightness changes and provide a stream of events that encode the time, location (X, Y), and sign (polarity) of the brightness changes, as seen in Fig. 2. DVS cameras have been utilized in various applications, such as object detection/tracking [11-13] and gesture detection/ classification [14,15]. DVS sensors capture detailed movement phases and fast movements without blurring [15]. With the advantages of DVS cameras, the existence and movement of WBCs in the nailfold can be captured.

This paper proposes event-based white blood cell classification in the nailfold capillary using various machine learning algorithms. To this end, we recorded a new dataset using the DVS camera for nailfold capillary videos, and then conducted feature engineering and classification. Extensive results are provided to evaluate the various machine learning algorithms.

2. The Proposed Method

This section describes our proposed method. In this study, we propose a classification approach using various machine learning algorithms given input from the WBC event-data stream. Fig. 3 shows the overall procedure of the proposed model consisting of event accumulation, feature engineering, and machine learning classification.

Fig. 3. Proposed WBC classification method.
../../Resources/ieie/IEIESPC.2022.11.4.270/fig3.png

2.1 Event-data Stream

An event-based camera (the DVS camera) is a type of vision sensor that can capture changes in intensity levels at each pixel independently to within microseconds (the so-called $\textit{events}$). Different from a standard camera’s output frame, a DVS camera provides a stream of event data consisting of $\textit{(t, x, y, p)}$ information, where $\textit{t}$ is the timestamp generated when an event occurs, $\textit{x}$ is the x-axis of the spatial location of an event, $\textit{y}$ is the y-axis of the spatial location of the event, and $\textit{p}$ is the event polarity.

A DVS sensor identifies an event in a corresponding pixel only when the intensity change at the pixel surpasses threshold ${\theta}$, as shown in Eq. (1):

(1)
$ \log \left(I_{now}^{\left(x,y\right)}\right)-\log \left(I_{\textit{previous}}^{\left(x,y\right)}\right)>\theta $

where $I_{now}^{\left(x,y\right)}$ and $I_{\textit{previous}}^{\left(x,y\right)}$ are the current and previous intensities at the corresponding pixel (x, y) [15]. Table 1 is an example of the event-data stream in our experiment.

The WBC flows through the nailfold capillary. Due to this WBC flow, the pixel intensity around the WBC changes, and hence, the DVS camera captures WBC movement.

Table 1. Event-data streams.

t

X

Y

P

1631582467514400

340

241

0

1631582467514410

135

15

1

1631582467514410

8

182

1

1631582467514410

248

113

1

1631582467514420

136

52

1

1631582467514420

110

94

1

1631582467514440

55

48

1

1631582467514450

129

64

1

1631582467514450

24

119

0

1631582467514460

67

100

1

.

.

.

.

.

.

.

.

.

.

.

.

1631582518334360

326

106

1

1631582518334370

332

162

1

1631582518334380

77

93

0

2.2 Data Acquisition and Event Accumulation

The original data come from pre-recorded WBC video during previous research [17], and we selected two videos from two subjects. To obtain the event data, we re-recorded the video using a DVS camera and display monitor.

The acquisition process of our event-data stream dataset employed a DVS camera by iniVation, the DAVIS346, which captures event-data stream output at an area size of 346x260. All the event-data streams have x- and y-axes within this area. Alongside the event stream, the DAVIS346 outputs a simultaneous intensity frame or an active pixel sensor (APS) image, which is a grayscale image of a global shutter sensor. However, in this study, the grayscale image is only used to assist with the data labeling procedure.

After the event-data streams are acquired, event-data segmentation and accumulation are conducted. Segmented event data are 200ms long, and all event data within this window are accumulated. Fig. 4 visualizes the event accumulation of a single window with or without the existence of WBCs. This visualization image is only used for easy readability. For the experiments, all data are in the form $\textit{t, x, y, p}$.

Fig. 4. Visualization of the accumulated events: (a) without WBC; (b) with WBC. The red circle highlights the WBCs.
../../Resources/ieie/IEIESPC.2022.11.4.270/fig4.png

2.3 Feature Reduction

One characteristic of the event-based camera is that it works asynchronously, resulting in a non-uniform and arbitrary number of events for each segment. Owing to the nature of the classifiers, there is a necessity to create the same $\textit{N}$ events for each segment. In our experiments, we set $\textit{N}$ to 100,000 events and added zero padding for the segments that had fewer than 100,000 events. Then, we selected prominent features of the input data by employing principal component analysis (PCA). PCA is commonly used to transform a large set of data into smaller sets without losing information from projecting the data to a new set of features called principal components. By selecting a subset of principal components with high variance, the feature space dimension will be reduced. In this research, with 95% variance, the data size of each sample was reduced by 99%. Originally each sample had a length of 100,000, which shrank to 346 features after the PCA process.

2.4 Classifiers

2.4.1 k-Nearest Neighbors

k-Nearest neighbors (KNN) is a non-parametric machine learning algorithm usually applied to classification tasks. The KNN algorithm works as follows: define the class number and a distance metric; find the nearest neighbor for every sample; and then, assign a class to the samples based on the distance between the sample and the center of the class.

2.4.2 Decision Tree

A decision tree (DT) is a non-parametric machine learning algorithm for regression and classification tasks. The model predicts the target after learning a simple rule-based decision method based on data features.

2.4.3 Random Forest

Random forest (RF) is an algorithm for both regression and classification tasks considering multiple decision trees that create a forest. Specifically, it combines several weak models to create a better and more robust model. RF employs a bagging strategy to prevent overfitting problems, and thus results in greater generalization.

3. Performance Evaluation

3.1 The Dataset

A dataset was generated from two videos with a total of 490 segments, among which 279 segments contained WBCs, and 211 did not. The event coordinates, $\textit{x}$ and $\textit{y}$, were used as input. The data were divided into training and testing datasets at an 8:2 ratio where 392 segments were used for training and 98 for testing.

3.2 Performance Evaluation

The performance of the experiments was evaluated using accuracy and F1-score. Accuracy is calculated as follows:

(2)
$ \textit{Accuracy}~ =~ \frac{TP+TN}{\textit{Total}~ Data} $

where \textit{TP} and \textit{TN} are true positive and true negative, respectively. Precision, recall, and F1-score are calculated as follows:

(3)
$ \textit{Precision}=\frac{TP}{TP+FP} \\ $
(4)
$ \textit{Recall}=\frac{TP}{TP+FN} \\ $
(5)
$ F1_{\textit{score}}=2\frac{\textit{precision}\cdot \textit{recall}}{\textit{precision}+\textit{recall}}=\frac{TP}{TP+\frac{1}{2}\left(FP+FN\right)} $

The parameters of each algorithm used in the simulations are listed in Table 2. For KNN, the number of neighbors used in the simulation was five. For the decision tree and random forest, the entropy function was used to measure the quality of a split tree. While the maximum tree depth value was 100 for the decision tree, 20 was chosen as the maximum tree depth and the number of estimators for the random forest algorithm.

Table 2. Parameters for each algorithm.

Classifier

Parameter

KNN

Number of Neighbors = 5

Decision Tree

Criterion = Entropy

Max. Depth = 100

Random Forest

Criterion = Entropy

Max. Depth = 20

Number of Estimators = 20

Table 3 shows the performance comparison from KNN, the decision tree, and random forest in terms of accuracy and F1-score, while Table 4 provides the precision and recall value for each classifier. The best accuracy and F1-score were 75.51% and 75.08%, respectively, obtained using the random forest algorithm. As well as the best accuracy, random forest yielded the best precision and recall at 75.71% and 75.51%, respectively. Fig. 5 shows the confusion matrices for all algorithms. Among them, random forest produced the most balanced performance.

Table 3. Performance Evaluation.

Classifier

Accuracy (%)

F1-score (%)

KNN

65.31

61.11

Decision Tree

67.35

67.25

Random Forest

75.51

75.08

Table 4. Precision and Recall.

Classifier

Precision (%)

Recall (%)

KNN

68.89

65.31

Decision Tree

67.20

67.34

Random Forest

75.71

75.51

Fig. 5. Confusion matrices.
../../Resources/ieie/IEIESPC.2022.11.4.270/fig5.png

4. Conclusion

This paper proposed a new way to detect WBC existence using event-based data. Classification was conducted using three different machine learning algorithms: k-nearest neighbors, the decision tree, and random forest. Based on our evaluation, event-based WBC classification is a promising new method to detect WBC existence in nailfold capillaries. This proposed state-of-the-art method can be used for developing a more efficient, non-invasive WBC classification method. The limitation of this paper is that we focus on fixed length numbers of events in every segment. For future directions, the process could address arbitrary lengths of event-data streams, WBC counting, and implementation of recent deep learning methods.

ACKNOWLEDGMENTS

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2020R1F1A1071712).

REFERENCES

1 
Newman T. B., Draper D., Puopolo K. M., Wi S., Escobar G. J., Aug. 2014, Combining Immature and Total Neutrophil Counts to Predict Early Onset Sepsis in Term and Late Preterm Newborns, Pediatric Infectious Disease Journal, Vol. 33, No. 8, pp. 798-802DOI
2 
Honda T., Uehara T., Matsumoto G., Arai S., Sugano M., Jun. 2016, Neutrophil left shift and white blood cell count as markers of bacterial infection, Clinica Chimica Acta, Vol. 457, pp. 46-53DOI
3 
van Wolfswinkel M. E., et al. , Dec. 2013, Predictive value of lymphocytopenia and the neutrophil-lymphocyte count ratio for severe imported malaria, Malaria Journal, Vol. 12, No. 1, pp. 101DOI
4 
Crawford J., Dale D. C., Lyman G. H., Jan. 2004, Chemotherapy-induced neutropenia, Cancer, Vol. 100, No. 2, pp. 228-237DOI
5 
Golan L., Yeheskely-Hayon D., Minai L., Dann E. J., Yelin D., Jun. 2012, Noninvasive imaging of flowing blood cells using label-free spectrally encoded flow cytometry, Biomedical Optics Express, Vol. 3, No. 6, pp. 1455DOI
6 
Shih T.-C., et al. , 2011, Hemodynamic analysis of capillary in finger nail-fold using computational fluid dynamics and image estimation, Microvascular Research, Vol. 81, No. 1, pp. 68-72DOI
7 
Bourquard A., et al. , 2015, Analysis of white blood cell dynamics in nailfold capillaries, in 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 7470-7473DOI
8 
Bourquard A., et al. , 2018, Non-invasive detection of severe neutropenia in chemotherapy patients by optical imaging of nailfold microcirculation, Scientific Reports, Vol. 8, No. 1, pp. 1-12DOI
9 
McKay G. N., et al. , 2020, Visualization of blood cell contrast in nailfold capillaries with high-speed reverse lens mobile phone microscopy, Biomed Opt Express, Vol. 11, No. 4, pp. 2268-2276DOI
10 
Gallego G., et al. , Jan. 2022, Event-Based Vision: A Survey, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 44, No. 1, pp. 154-180DOI
11 
Perot E., de Tournemire P., Nitti D., Masci J., Sironi A., 2020, Learning to Detect Objects with a 1 Megapixel Event Camera, in Advances in Neural Information Processing System, Vol. 33, pp. 16639-16652sURL
12 
Wan J., et al. , Apr. 2021, Event-Based Pedestrian Detection Using Dynamic Vision Sensors, Electronics (Basel), Vol. 10, No. 8, pp. 888DOI
13 
Cannici M., Ciccone M., Romanoni A., Matteucci M., Jun. 2019, Asynchronous Convolutional Networks for Object Detection in Neuromorphic Cameras, in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1656-1665DOI
14 
Belbachir A. N., Schraml S., Nowakowska A., Jun. 2011, Event-driven stereo vision for fall detection, in CVPR 2011 WORKSHOPS, pp. 78-83DOI
15 
Wang Y., et al. , Jun. 2019, EV-Gait: Event-Based Robust Gait Recognition Using Dynamic Vision Sensors, in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6351-6360DOI
16 
Mueggler E., Huber B., Scaramuzza D., Sep. 2014, Event-based, 6-DOF pose tracking for high-speed maneuvers, in 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 2761-2768DOI
17 
Hariyani Y. S., Eom H., Park C., 2020, DA-Capnet: Dual Attention Deep Learning Based on U-Net for Nailfold Capillary Segmentation, IEEE Access, Vol. 8, pp. 10543-10553DOI

Author

Yuli Sun Hariyani
../../Resources/ieie/IEIESPC.2022.11.4.270/au1.png

Yuli Sun Hariyani received a BS in telecommunication engineering and an MS in electrical engineering from Telkom University, Bandung, Indonesia, in 2010 and 2013, respectively. She is currently pursuing a PhD in the Computer Engineering Department at Kwangwoon University, Seoul, South Korea. Since 2014, she has been a Lecturer with Telkom University, Indonesia. Her research interests include pattern recognition, medical image processing, and biomedical signal processing.

Cheolsoo Park
../../Resources/ieie/IEIESPC.2022.11.4.270/au2.png

Cheolsoo Park is an associate professor in the Computer Engineering Department at Kwangwoon University, Seoul, South Korea. He received a B.Eng. in Electrical Engineering from Sogang University, Seoul, and an MSc from the Biomedical Engineering Department at Seoul National University, South Korea. In 2012, he received his PhD in Adaptive Nonlinear Signal Processing from the Imperial College London, U.K., and worked as a postdoctoral researcher in the Bioengineering Department at the University of California, San Diego, U.S.A. His research interests are in the areas of machine learning and adaptive and statistical signal processing, with applications in brain computer interfaces, computational neuroscience, and wearable technology.