MalariaNet: A Computationally Efficient Convolutional Neural Network Architecture for Automated Malaria Detection

— Despite much progress in detection and treatment, malaria remains one of the most prevalent diseases on earth, both in terms of incidence and death rate. Multiple studies have shown that early detection of malaria is paramount to preventing fatal outcomes; however, current testing methods have notable issues involving cost and accessibility. As a result, deep learning algorithms have been developed for malaria detection and have achieved state of the art results in rapid diagnosis; however, it has been noted that the computational expense of running elaborate models makes deep learning based detection methods inaccessible in remote areas of the world. We develop a computationally efficient, relatively shallow neural network architecture that can diagnose malaria from cell images obtained from thin blood smear slides. Specifically, our algorithm, dubbed MalariaNet, is a 7-layer convolutional neural network trained using the Adaptive Moment Estimation algorithm on the open source NIH malaria dataset, containing 27,588 images of parasitized and uninfected cells. We report that MalariaNet achieves an accuracy of 0.968, F1 score of 0.955, precision of 0.946, and recall of 0.974. We hope that our computationally considerate model inspires more research in producing accessible artificial intelligence solutions for disease detection tasks.


INTRODUCTION
Malaria is a mosquito-borne infectious disease that is caused by single-celled, parasitic microorganisms of the Plasmodium genus and is typically spread via the bite of infected female Anopheles mosquitoes, as the parasites from the saliva of the mosquito are transferred to the blood of the affected human [5]. Subsequently, the parasites transfer to the liver where they are able to efficiently develop and reproduce. The disease manifests itself as fever, tiredness, vomiting, and headaches; in severe cases, it can also result in yellow skin, seizures, coma, and even death. These symptoms typically present themselves ten to fifteen days after infection, but recurrence can occur months later. Although medicinal research has made strides in malaria prevention, detection, and alleviation, malaria remains the leading cause of death in the world and at this time, there exists no effective vaccination. Most cases occur in tropical and subtropical regions, particularly Sub-Saharan Africa, Asia, and Latin America. Overall, malaria was estimated to affect 228 million people and cause an estimated 405,000 deaths in 2018, with 94% of these fatalities occurring in Africa [4]. Consequently, the disease is estimated to cost Africa over $12 billion due to healthcare costs, decreased workforce, and negative effects on tourism.  [4] The most widely used methods for malaria diagnosis fall into two categories: direct and indirect [17]. Direct methods of detection rely upon confirming a diagnosis based upon the discovery of parasitic bodies or parts of parasitic bodies. Indirect methods involve the detection of antibodies known to be relevant agents in the diagnosis of malaria. The advantages and disadvantages of the primary direct and indirect malaria diagnostic test methods were recorded by Talapko, et al. and are re-expressed in Table 1 and Table 2. Time-consuming to perform and expensive As seen in both Table 1 and Table 2, the current methods of malaria diagnosis all feature a combination of issues in price, accessibility, evaluation of results, or low levels of sensitivity.
According to the CDC, microscopic analysis of potentially infected cells remains the gold standard in malaria diagnosis [4]. However, they note that the quality and accuracy of the test is highly dependent on the experience of the laboratorian conducting the diagnostic. As a result, malaria detection through microscopic analysis is inaccessible to millions around the world since there are limited numbers of experienced laboratorians in many African and Asian countries. For these reasons, despite being one of the cheapest and quickest methods of malarial diagnosis, microscopic analysis is still unattainable for the majority of at-risk individuals across the world.

A. Previous Work
Numerous studies in the past have explored the use of artificial intelligence based techniques for malaria related tasks. For example, one group utilized feature extractors implemented through pretrained deep convolutional neural network models in order to classify parasitized and non-parasitized malaria cells from blood smears [13]. Another group developed a five step image analysis system that drew upon various machine learning systems for optimized malaria detection from blood smears. Other groups have drawn upon powerful deep learning architectures for malaria detection. Specifically, models such as ResNet-50, DenseNet-121, Xception, VGG-16, and other elaborate architectures have been used for malaria classification, and have achieved state of the art classification accuracies [6,9]. Finally, a recent study utilized a deep convolutional neural network in conjunction with cyclical stochastic gradient descent and automated learning rate optimization for malaria detection [11].
Outside of classification, other work has been done in using deep learning methodologies for segmentation tasks. Anggraini et al. utilized segmentation models to automatically separate images of parasitized malaria cells from their backgrounds. Another group used unsupervised methods to automatically cluster malaria cells using blood smear images [1]. Other researchers utilized spatial uncertainty sampling to reduce domain shift bias in segmentation mask generation for parasitized cells [16].
Overall, it can be concluded that although machine learning techniques have shown a high propensity for malaria related tasks, limitations exist. Specifically, the aforementioned studies almost exclusively employ complex, deep neural network architectures that demand high levels of computational power for training. To that end, we note an inherent lack in algorithms with the ability to provide state of the art diagnostic metrics while maintaining ideal computational efficiency. Since malaria classification from blood smear images is a feature rich task, our research looks to explore the implementation of a concise, relatively shallow, convolutional neural network architecture for accurate detection.

A. Dataset Acquisition
The dataset used in this study was provided by the National Institute of Health (NIH) and contained 27,558 images with equal instances of parasitized and non-parasitized red blood cells. The cells were stained with Giemsa, mimicking the procedure that would be undertaken when using microscopic analysis to diagnose malaria. A sample of the dataset is shown in Fig. 2 and Fig. 3.

B. Image Preprocessing
The first step was to process the images and split them into train and test subsets, thereby making the dataset ready for usage. The images were processed via the nearest-neighbor interpolation technique, in which a group of pixels automatically assume the magnitude of the pixel closest to it. It is the simplest method of multivariate interpolation, which also makes it the most computationally efficient. The images were downscaled to a standard size of 100x100 pixels, which was necessary as the images were previously of variable size and scale. Applying nearest-neighbors interpolation minimized image distortion and retained information better than less sophisticated methods such as resizing or cropping. We then placed 80% of the images in a training dataset, while the remaining 20% were used for testing.

International Journal of Engineering Research & Technology (IJERT)
ISSN: 2278-0181 http://www.ijert.org III. MALARIANET The malaria detection task is a binary classification problem, where the input is an image of a cell and the output is a binary label ∈ {0, 1} indicating the absence or presence of malaria, respectively.
To accomplish this task, a seven-layer CNN was utilized that we dubbed MalariaNet. This model contained two convolution layers, two pooling layers, a flattening layer, and two fully connected dense layers. The dense layers alleviate the vanishing gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters. The network structure can be seen in MalariaNet Layer Map. Within our model, we utilized Adaptive Moment Estimation (Adam), an adaptive learning rate optimization algorithm. Our activation functions for the intermediate layers was the rectifier function, with the final layer using a sigmoid nonlinearity. A batch size of ten was used over 25 epochs, after each of which the model output accuracy metrics and loss, given by the binary cross-entropy loss function: In order to prevent the networks from overfitting, early stopping was performed by saving the network after every epoch and choosing the saved network with the lowest loss on the tuning set. Overall, 2,177,185 parameters were trained and optimized for this task. The architecture of the network can be seen in Table 3. IV. RESULTS The progression of MalariaNet training can be seen in Fig.  4, 5, and 6. We present loss, accuracy, F1 score, precision, and recall by each epoch.  Table 4.

V. CONCLUSION
Malaria afflicts hundreds of thousands of people around the world and is among the leading causes of death in developing countries [5]. The disease is primarily diagnosed through the technique of microscopic analysis, which necessitates trained laboratories and operational facilities that

International Journal of Engineering Research & Technology (IJERT)
ISSN: 2278-0181 http://www.ijert.org are often unavailable. Seeking to alleviate this predicament, we developed a compact and efficient convolutional neural network architecture, MalariaNet. We then trained MalariaNet on a publicly available dataset containing approximately 27,000 parasitized and uninfected red blood cell images. Our model attained high performance metrics across the board, including a 96.8% accuracy and 0.955 F1 score. MalariaNet's intriguing diagnostic capabilities provide the potential for its implementation in remote locales while concurrently promoting subsequent exploration regarding the ability of relatively shallow deep learning architectures in boosting accuracy and efficiency. We hope that our model will be used to further the development of efficient algorithms for disease detection.