banner
Home / News / Machine learning at the edge for AI
News

Machine learning at the edge for AI

Nov 23, 2023Nov 23, 2023

Scientific Reports volume 13, Article number: 4744 (2023) Cite this article

1941 Accesses

1 Citations

111 Altmetric

Metrics details

Multiplexed detection of biomarkers in real-time is crucial for sensitive and accurate diagnosis at the point of use. This scenario poses tremendous challenges for detection and identification of signals of varying shape and quality at the edge of the signal-to-noise limit. Here, we demonstrate a robust target identification scheme that utilizes a Deep Neural Network (DNN) for multiplex detection of single particles and molecular biomarkers. The model combines fast wavelet particle detection with Short-Time Fourier Transform analysis, followed by DNN identification on an AI-specific edge device (Google Coral Dev board). The approach is validated using multi-spot optical excitation of Klebsiella Pneumoniae bacterial nucleic acids flowing through an optofluidic waveguide chip that produces fluorescence signals of varying amplitude, duration, and quality. Amplification-free 3× multiplexing in real-time is demonstrated with excellent specificity, sensitivity, and a classification accuracy of 99.8%. These results show that a minimalistic DNN design optimized for mobile devices provides a robust framework for accurate pathogen detection using compact, low-cost diagnostic devices.

Detection and identification of biomolecules are essential parts of diagnostic devices in the disease control realm. The COVID-19 pandemic has accelerated the use of home testing for early detection and repeated monitoring, and it can be expected that point-of-care analysis will expand in volume and capabilities. Some of the challenges here are sample preparation and fluid handling, sensitivity and specificity, data acquisition and processing, portability, and connectivity1,2. Various micro and nanotechnologies have resulted in sensors applicable to on-chip molecular diagnostics. For example, a complementary metal–oxide–semiconductor (CMOS) based biosensor utilizing a nanomechanical membrane-bridge design3 was introduced to measure the concentration of a phenytoin drug in a liquid sample with a microelectromechanical (MEMS) approach. Paper-based analytical devices (PADs) are another good example of point-of-care diagnostics. They have evolved over the past decade due to properties such as low cost, biocompatibility, simple design and usage, and flexibility in pursuing affordable disposable testing units4,5,6. While sample analysis can be done in a few minutes, these devices have a relatively high limit of detection (LOD > µM), demanding amplification and cultivation time (usually several hours) for highly sensitive nucleic acid analysis. Among amplification-based techniques, quantitative polymerase chain reaction (qPCR) is still the gold standard method in most of laboratory test instruments due to its very high sensitivity (~ 100 copies/mL) and versatility7,8,9. A newer single-tube technique with a relatively inexpensive and simple amplification process called loop-mediated-isothermal-amplification (LAMP) has become more popular in the past decade10,11. Further assay simplification based on elimination of amplification steps and chip-scale integration may be desirable in some applications. In the single-molecule sensitivity regime, nanopore devices as electrical biosensors have shown promise as label-free and amplification-free ultra-high sensitivity diagnostic tools12,13. Fluorescence-based single-molecule detection sensitivity was demonstrated in optofluidic waveguide devices, enabling amplification-free detection of Ebola virus (EBOV) RNAs with low limits of detection (LoD) down to 90 aM14,15. Multiplex analysis was implemented on these single-molecule optofluidic ARROW sensors by creating spectrally and/or spatially dependent multi-spot excitation patterns at the output of multi-mode interference (MMI) waveguides16,17. The spatial multi-spot pattern translates into a temporal multi-peak signal (see Fig. 1c) which can be easily identified by signal processing algorithms that recognize the characteristic time difference ∆t between fluorescence peaks from a single target. Up to 7× multiplexed detection of nucleic acids has been demonstrated using spectrally, spatially, and velocity-dependent translation of the excitation pattern into time-dependent multi-peak fluorescence signals17,18. Using deliberately patterned signals has been used in other contexts as well. For example, telecom-like signal encoding was implemented in cell cytometry using electrical resistive pulse sensing in which the pattern of events generated by flowing cells depends on the arrangement of the electrodes and decodes digital signals from different channels with a very high accuracy19,20,21. A multi-finger electrode electrical biosensor was studied to demonstrate the effect of increased electrode count in the impedance signal-to-noise ratio (SNR)22,23,24,25. Encoding signals in biosensing applications can take advantage of more complicated information coding techniques such as multiplexing, error correction, and identification23,26,27 with a recent trend toward machine learning techniques28. Optimizing the signal analysis method of choice is critically important as real-world device imperfections tend to compromise signal quality and, thus, the reliability of the signal analysis. In the case of an optofluidic MMI waveguide device, these can include fabrication-induced variations in MMI excitation patterns, velocity variations due to fluidic dynamics of flowing targets, and signal-to-noise ratio variations caused by a range of fluorescence signal intensities. These types of nonidealities are added on top of the signal quality limitations faced by point-of-care devices, where components should be produced at low cost, and environmental factors often impact signal quality. These intrinsic challenges can be alleviated by a powerful signal analysis approach that can operate in real-time.

Experimental setup and dataset preparation. (a) Fluorescence signals are collected from labeled Klebsiella Pneumoniae bacteria captured on microbeads by a target-specific sandwich assay. Inset shows the sandwich design of the fluorescence probes, which consists of both green and red dyes. The sample is prepared off-chip and flows into the ARROW chip using negative pressure-driven flow. The multimode-interference (MMI) pattern formed at the MMI WG-analyte channel intersection is wavelength dependent and defines the signatures of different signals. The experiment is done in three steps depending on if beads are excited by the red laser, green laser, or both colors. Combinations of mechanical shutters (MS1 and MS2) are used to create the three mentioned stages. Fluorescence signals are collected off-chip through a penta-bandpass-filter (PBPF) to remove excitation photons. Emission photons are then detected and recorded on a PC by a single-photon-counting-module. (b) The recorded data trace (training signal) is analyzed by the PCWA program to detect events. Events are cropped and saved for dataset preparation. Locations of detected events are shown with blue markers. Event detection and cropping are done automatically. c Example events from three groups of events.

Recently, we have introduced a new parallel cluster wavelet analysis (PCWA) approach that can identify events in a time series, such as fluorescence from single molecules, with high accuracy and speed29. However, the analysis speed and accuracy drop for multiplexed detection due to the need to tune the custom wavelet functions for different targets before analysis, which adds extra overhead to the chip-to-chip variation of fluorescence signal characteristics. Augmenting PCWA with Machine Learning (ML) can solve this problem and lead to robust, multiplexed diagnostics.

Machine learning and artificial intelligence (AI) ideas and theories have been around for many decades, if not a century, with the model of nonlinear neurons forming a network back in the mid-1990s30 and became a functional tool in pattern recognition with better learning theories being developed in late 1990s31,32,33. Even though the ML concept is not new, the field has exploded in the last 2 decades when large-scale high-performance computation became possible to handle ever-larger amounts of data. AI has become a part of our daily life and is used in numerous areas, including agriculture, environmental studies, autonomous vehicles, economics, marketing, drug development, entertainment, and medical diagnosis34,35,36,37,38,39,40,41,42,43. In biomedical diagnosis, artificial intelligence-assisted methods at different levels, from large-scale population data down to sensory data analysis, have been developed, underscoring the advantages of novel machine-learning over classical techniques44,45,46.

Here, we introduce a full-stack AI-assisted framework integrated with a multiplexed detection platform to detect and identify biomolecules. Its capabilities are demonstrated on an optofluidic MMI waveguide device for multiplexed detection of fluorescently tagged biomolecules16. The system is evaluated by analysis of three combinatorial multi-color signals from representative plasmids of drug-resistant Klebsiella Pneumoniae bacteria. Compared to previously used techniques, outstanding recall and accuracy are observed. Running in real-time, the detection framework was able to discover 93.8% of the events, with 99.8% classification accuracy. An affordable portable AI-specific development board (Google Coral Dev board) was then selected to show the compatibility of the optofluidic diagnostic system with the state-of-the-art processing schemes47. The Edge-TPU (tensor processing unit) available in this development board is a coprocessor capable of performing four tera-operations per second (TOPS) with a power consumption of 500 mW/TOPS48. Finally, we evaluated the timing performance and real-time sample analysis capability of the framework. This system will open new avenues for ultra-fast point-of-care devices for diagnostic and other applications.

Our framework can be applied to any time-series signal that features different time signatures for different targets. We introduce it using a fluorescence detection assay for Klebsiella Pneumoniae bacteria nucleic acid biomarkers. A sandwich design assay binds target DNA strands to a pulldown sequence on a streptavidin-coated microbead, after which they are labeled with fluorescent probes in two colors (Fig. 1a). The assay is, therefore, sensitive to both red and green excitation wavelengths, allowing for three possible fluorescence signals (red only, green only, and red + green), depending on the excitation wavelengths used. Targets are flowed through and detected on an ARROW-based optofluidic as shown in Fig. 1a. The sensitivity (limit of detection) of this flow-based optofluidic fluorescence assay is 90 aM15. Two lasers running at 738 nm and 556 nm are coupled into a single-mode optical fiber, and two mechanical shutters, MS1 and MS2, are placed in the light paths to toggle colors on and off independently. The experiment is done in three stages, first with only 738 nm excitation (MS1 = Open, MS2 = Close), second with only 556 nm (MS1 = Close, MS2 = Open), and third with both 738 nm and 556 nm excitations active simultaneously (MS1 = Open, MS2 = Open). The sample containing the fluorescent targets is introduced into the chip by applying negative pressure to one of the reservoirs and then introducing the sample into the other reservoir using a pipette. The recorded fluorescence time trace from the chip is stored on a desktop computer (Fig. 1b). The multi-mode interference excitation waveguide produces different spot patterns for the two wavelengths in the intersecting fluidic channel with the number of spots N determined by Eq. (1)

where λ is the respective input wavelength, w is the effective waveguide width, nC is the refractive index nC, and L is the length of the MMI waveguide49. These spatial excitation patterns are transformed into the time domain as particles flow through the channel at a given velocity and fluorescence. Events from the first stage have six peaks, events from the second stage have eight peaks, and events from the third stage contain both 6 and 8 peaks. We named these events by splitting the word KPneu into two parts (KPn for six peaks and eu for eight peaks) to represent which probe has been excited in the detected event and will use these labels throughout the article. This approach effectively creates a 3× multiplex assay with three distinct signal shapes corresponding to the excitation wavelength selections. Figure 1c shows example events for each stage which illustrate the different fluorescence patterns and the nonidealities present in this configuration. These include finite background signal, peak-to-peak amplitude variations due to fabrication imperfections, and different peak-to-peak spacing Δt due to velocity variations from pressure fluctuations and position in the channel. A coefficient of variation of 44.85% is observed in the velocity distribution of 1544 detected particles. Minimizing these nonidealities would require raising the production cost to increase the precision of the microfabrication processes. Some nonidealities, such as cross-section fluid flow speed variations, are simply due to the nature of the microfluidic channels being used. Point-of-care devices, meant to be low-cost and compact, must always deal with signal imperfections. A natural strategy is, therefore, to adapt to these imperfections by employing machine learning to recognize the signal pattern for a given device and detect and identify fluorescence events with good sensitivity and accuracy.

The first step in our signal-recognition process is to detect events from the recorded fluorescence signal using a fast multi-scale continuous wavelet-based technique called PCWA29 with multi-spot-Gaussian (MSG) basis functions to match the characteristic signals for events in each stage (see Supplementary Fig. S1a). After locating events in the time trace, a collection of cropped windows comprising 1024 data points centered on an individual event’s time location is used to create the dataset for neural network analysis and classification. False identified or overlapping events are removed from the dataset using a multi-factor signal quality check to improve the neural network model’s training (see Supplementary Fig. S2). The entire event detection, cropping, filtering, and annotation is automated and does not require any user involvement.

The main components of the deep neural network are shown in (Fig. 2a, right). Fluorescence events with temporal signatures representing the MMI pattern exhibit unique time–frequency features for each class of events which also depend on the velocity of the flowing particles. Therefore, we take advantage of a short-time-Fourier-transform (STFT), which has been used in speech, EEG, and ECG data to extract time–frequency features50,51,52,53. The input signal is first transformed into a 128 × 128 pixel image expressing the magnitude of the STFT of the event (Fig. 2a, left). A DNN model is designed, built, and tested from scratch for this application with the goal of minimal complexity and high compatibility with the target inferencing edge platform. First, the input image is down-sampled into a 64 × 64 pixel image. Two cascaded convolutional layers are deployed to extract a pool of features. Finally, a dense layer with a softmax activation function is used to classify input images into three output classes corresponding to red, green, and red-green fluorescence. The model contains only 6454 parameters (6436 trainable) which is orders of magnitude lower than popular image classification models54. The model is a small application-specific model adapted from the LeNet CNN model, which was used to recognize handwritten digits55. The DNN model is trained using a supervised learning technique by a dataset of spectrograms calculated for events detected from the training time trace (see Fig. 2c). Since labeling can be easily and automatically done based on the lasers’ ‘on’ and ‘off’ states, supervised learning was the best-suited learning technique here.

Deep Neural Network and dataset. (a) DNN model consists of two conv2d layers followed by a max_pooling2d and a batch normalization in between. A classifier layer (dense) maps extracted features from previous layers to the output classes. The plot on the left shows how the input signal is first converted into a 128 × 128 pixels spectrogram using STFT. (b) The progress of training with accuracy and loss metrics evolves over 100 epochs. Training is stopped at the 100th epoch to prevent overfitting. After 90 epochs, the loss value stays below 0.1. c Examples of events from the training dataset. Spectrograms and the corresponding cropped time events are shown together, but only the spectrograms and labels of the events are fed to the DNN model during training.

Training was done by minimizing the categorical cross-entropy loss function L, defined by Eq. (2). To minimize the loss, a stochastic gradient descent (SGD) method was applied by using Adam optimizer56 with the default learning parameters (learning rate = 0.001, β1 = 0.9, β2 = 0.999, ε = 10−7) set by TensorFlow.

where pi,c is the probability value of class c, for sample i, returned by a softmax layer of the classifier. yi,c is the true (one-hot) label of each class for sample i, and N = 90 is the size of minibatch used during training. The softmax function is defined by Eq. (3):

After 90 epochs, the loss value drops below 0.1, and we stopped at 100 epochs to prevent overfitting (Fig. 2b). We found that a ~ 250 nL volume of sample (approximate concentration 106 beads/mL) contains enough events (~ 300 events/class) to train the model and the entire training step takes less than 15 min on a single desktop machine.

A Google Coral Dev Board57 is selected as the AI accelerator hardware for the system. It incorporates a system-on-module (SoM) tensor processing unit (Edge-TPU) chip alongside an ARM-based microcomputer with Raspberry Pi form factor and essential connectivity modules58. Currently, this device is limited to only inferencing, meaning that the training step needs to be done on other platforms (i.e., desktop, cloud). The model is implemented in TensorFlow59, which has the best compatibility with the Coral board. The trained DNN model (2D-DNN) on a desktop computer (32-floating point) is converted into a quantized 8-bit integer model compatible with Edge-TPU (2D-DNN-EdgeTPU) via a transfer-learning procedure60. A test dataset from a test experiment is created the same way as the training dataset to assure that none of the classifiers had been exposed to test events. Performance evaluation (Fig. 3a) shows the receiver’s operating characteristic curve (ROC) for the 2D-DNN as well as the conventionally used classifier for the multiplexed fluorescence detection, shift-and-multiply (SaM). The ROC curve visualizes the true positive and false-positive rates, i.e., sensitivity and specificity, for a given selection threshold. The 2D-DNN model outperforms SaM with a near-perfect area-under-curve (AUC ≅ 1) and only a few false identified events visualized in the confusion matrices (Fig. 3b). This demonstrates that our approach is able to correctly find and classify fluorescence signals from single particles. This is the first principal result of this article.

Performance comparison. (a) ROC plot for the studied classifiers. The DNN model, on a PC (2D-DNN) or Coral (2D-DNN-EdgeTPU), outperforms the previously used unsupervised classifier, SaM. (b) Confusion matrices for different classifiers highlight only 3 and 4 events misclassified for the 2D-DNN and 2D-DNN-EdgeTPU models, respectively. SaM fails to classify eu and KPneu events due to their low signal-to-noise (SNR) characteristics. (c) Dimensionality reduction applied to input and outputs of different classifiers is made using PCA and t-SNE. Input data fails to form clusters in linear (PCA) and nonlinear (t-SNE) methods. SaM, conventionally used for MMI event classification, weakly classifies events from different classes. The proposed 2D-DNN model shows robust classification performance noticeable in both PCA and t-SNE scatter plots with clearly separated clusters. The transferred 2D-DNN model running on the Edge device with 8-bit integer quantization performs comparably to the 32-bit float model of 2D-DNN.

Dimensionality reduction, as a prevalent tool for data visualization, especially in high dimensional data, helps to look for a relationship or meaningful structural information in an understandable lower-dimensional space61. Principle component analysis (PCA) as the most common linear and t-distributed stochastic neighbor embedding (t-SNE) as the other popular nonlinear dimensionality reduction techniques are used here to compare how each classifier model is able to group events into clusters61,62,63,64,65. Both PCA and t-SNE fail to find structural information and form clusters when applied to the raw input signals (Fig. 3c, left column). At the same time, SaM can separate some of the events when visualized by nonlinear t-SNE embedding. 2D-DNN (both versions) maps high dimensional input signals into a linearly and non-linearly separable three-dimension space (Fig. 3c). There is a negligible drop in 2D-DNN-EdgeTPU classifier performance related to the precision drop from the transfer-learning step. The average inference time on the Coral board using ARMx64 CPU is 4.14 ms, while a 2 × speedup is observed by utilizing the Edge-TPU accelerator resulting in a 2.07 ms average inference time. The inference is done on events with a 10.24 ms window size, indicating real-time classification performance.

To create a complete real-time signal analysis system, we designed an event detection and classification framework (Fig. 4b) that integrates sensory data acquisition (DAQ), event detection, preprocessing, DNN inferencing, database management, and visualization processes on a single low-power portable device. The individual blocks are distributed into independent processes and threads to run in parallel, which is necessary to achieve real-time performance with limited resources. Counting individual biomarker targets in real-time can improve diagnostic accuracy by providing better insight into the test right after the start compared to the conventional waiting period and final test result.

Detection on edge. (a) Real-time performance of the model running on Edge-TPU (Coral Dev board). By utilizing a pipelined architecture shown in (b), there is only a delay for the first batch, and the rest is processed without any additional delay. The 45-degree slope indicates the real-time calculation of results. The input data is processed in chunks cached in a queue; this is visible in the inset plot where multiple events are revealed at the same time. (b) A simplified block diagram showing the essential blocks of the pipelined data processing framework. Some blocks, i.e., DAQ, PCWA, and STFT, run in parallel using Python’s multiprocessing package. Out of 489 detected events on the Edge-TPU, 485 events are confirmed by manual inspection. Of 488 ground-truth events, 485 are correctly located, and 484 are classified, corresponding to 99% accuracy. (c) Developed Dash app running on Coral Dev board. Data from the sensor with identified events are visualized in different ways. The dashboard provides good insight into the analyzed sample in real-time.

The input data trace is first buffered in a queue with fixed-length chunks and then passed to the PCWA event detector process66 with a single generic rectangular wavelet (Supplementary Fig. S1b). Locations of detected events are used to crop and send cropped windows from the raw data trace to the STFT block, which then is sent to another queue to be classified by the DNN model. As depicted in Fig. 4a, the Coral Dev Board is capable of identifying events from the fluorescence data trace in real-time. This is the second principal result of this work.

The first batch of events takes tens of milliseconds to process, and due to the pipelined architecture, the remaining batches are processed without any further lag (slope of 45° for run time vs. event time location). Processed batches are visible in the (inset of Fig. 4a) as a group of events returned simultaneously with the same run time tag. Table 1 summarizes the framework’s detection rate and accuracy measures. The detection rate is calculated based on manual inspection of events, and classification is verified by event location in the fluorescence data-trace stage. The input data trace, and location of detected events with additional information, i.e., type of target, velocity, and the classification score, are displayed on a minimal live dashboard. A bar graph is used to display the count of detected events for each target type as the experiment evolves (Fig. 4c). The dashboard runs on a local network without any need for cloud resources. This scheme is an excellent practice for addressing recent data privacy concerns related to cloud-based data processing services47.

We have introduced a framework for amplification-free multiplexed, AI-assisted pathogen detection on the fly. The use of a neural network based machine learning algorithm allows for identification and classification of fluorescence signals under realistic, nonideal conditions. Compared to the conventionally used multiplexed single-molecule identification technique, shift-and-multiply, SaM, the classification accuracy for our DNN model has over 40% better ROC-AUC metric while running fast enough to perform classification in real-time. We showed that conventionally demanding data processing tasks such as super high resolution multi-scale event detection and classification can now be done on much smaller devices with a very efficient neural network model containing only a few thousand parameters. Endpoint near-sensor data analysis has proven to be beneficial in terms of energy and network load spent on endpoint-cloud communication plus data privacy and security concerns47,67. Our framework solves similar concerns by taking advantage of an AI-specific edge device with an on-demand inferencing strategy. The efficient neural network model runs inference only if an event is detected in the time trace via the PCWA event detection process. Currently, the Coral Dev board is not capable of training the model and requires a desktop computer for training and model transfer steps. We believe this limitation can be solved with future updates of the hardware and libraries or by utilizing a cloud-based server for this specific task. Another current limitation is the upper limit for concentration range; more specifically, the events have to be spaced apart on the time axis. High concentration solutions can be easily diluted with the buffer, with the caveat that dilution will increase sample volume and test times. Future work will tackle this problem in the context of multiplexed detection with sensitivity down to single molecular biomarkers14,68,69,70.

The ARROW-based fluorescence single-molecule detection setup is depicted in (Fig. 1a). Two lasers running at 738 nm (Ti:Sapphire, Del Mar Photonics) and 556 nm (SSD Nd:YAG, Shanghai Dream Laser Technology Co.) are coupled into a single-mode optical fiber (F-SA, Newport) using a 60× microscope objective (Newport). A pair of modified PC cooling fans are used as mechanical shutters (MS1, MS2) to close/open optical paths for each color. The optofluidic chip is mounted on a 3D printed custom stage using double-sided tape, and the two brass cylindrical fluid reservoirs are glued to the liquid channel ends with wax. The vacuum line is connected to the outlet reservoir to provide negative pressure for sample flow inside the ARROW chip. The fluorescence signal from the labeled targets is guided through the collection waveguide and gathered from the side-facet by a 60× objective (Newport). The excitation light is then removed by a penta-bandpass optical filter (FF01-440/521/607/694/809-25, Semrock) before coupling the collected light with a multi-mode fiber optic patch cable with an FC connector. A single-photon counting module (SPCM-AQRH, Excelitas Technologies) converts fluorescence photons into electrical pulses, and a time-correlated single-photon counting (TCSPC) board (TimeHarp 260 nano, PicoQuant) records time-tagged photon events onto the computer storage disk.

10 µL of 3 µM synthetic nucleic acid strands (corresponding to antibiotic-resistant bacterial target) is mixed with 10 µL of 10 µM target-specific fluorescent probe oligomers [IDT DNA Inc.] (see Supplementary Table S1). The final volume for the target-probe mixture is made to 30 µL by adding 1XT50 buffer. The target-probe mixture is heated to 95 °C for 5 min and incubated for 3 h. Streptavidin-coated magnetic beads [DynabeadsTM] with 1 µm diameter are functionalized with target-specific biotinylated capture oligomers. After incubation, the hybridized target-probe structure is mixed with the functionalized magnetic bead in a rotator mixer for 1 h. A magnet kept under the vial is used to pull down the magnetic beads with the captured target-probe complex. All unbound nucleic acid strands are washed off, and the beads are resuspended. Figure 1a (inset) visualizes the final assay structure. This labeling technique has been previously shown to be very specific to the target sequence14.

The training dataset is created from the time trace shown in Fig. 1b. Events are detected by running the PCWA algorithm with three wavelet functions (MSG-6, MSG-8, and MSG-6&8) to pick three possible classes of events from the time trace. PCWA parameters are set as follows: scales from 0.1 to 1.0 ms with 50 logarithmic steps, spreading coefficients, h = 1, w = 1, and selectivity = 0.3. PCWA returns time location, scale value, CWT coefficient, and corresponding best-matched wavelet function in the form of a structured array which is then used to crop a window of 1024 datapoint from the 0.01 ms binned time trace centered at events location to form the training dataset. The best-matched wavelet function index is used as the label value in the dataset. An extra step of event quality check (see Supplementary Fig. S2) eliminates overlapping and poorly detected events from the training set.

The input signal is converted into a time–frequency spectrogram using short-time-Fourier-transform (STFT) with a 128 datapoint long Hanning window and 120 data points overlapping segments to ensure a balanced time–frequency resolution. A deep neural network with two sets of 2D convolutional layers with ReLU activation functions and 2D max-pooling layers are cascaded after a 2D average-pooling layer on the input image (spectrogram). The average-pooling layer with a pool size of 2 down-samples the input image with a size of 128 × 128 pixels into a 64 × 64 pixels image. We compared the results from the original spectrogram with the down-sampled version, and while we observed a speed up and reduced memory usage during the training and inference, there was no noticeable sacrifice in performance. The first convolutional layer consists of 9 filters with a kernel and stride size of 3 × 3 and 1 × 1 pixels, respectively. An additional batch normalization layer is used prior to the max-pooling layer to improve the training speed and robustness71. The second convolutional layer is a stack of 25 5 × 5 pixels filters with the same stride size as the first convolutional layer. Max-pooling layers have pool sizes equal to the kernel sizes used in the corresponding leading convolutional layer. The extracted features with the dimensionality of (3, 3, 25) are then flattened and fed into a dense layer with a softmax activation function which acts as the classifier. The output of the dense layer has three dimensions corresponding to three possible output classes (KPn, eu, and KPneu). A layer of 0.2 dropouts is appended to each convolutional layer as adaptive regularization72. The supplementary material presents a detailed DNN model composition (Supplementary Fig. S3a,b). The DNN model is implemented in Python-3.8.10 and TensorFlow-2.8.0 on a 64-bit desktop computer running Ubuntu 20.04.4 LTS.

The trained DNN model is quantized into an 8-bit integer model and compiled into an Edge-TPU compatible binary model. A fine-training step involves a few epochs (here 20) of re-exposing the quantized model to the training dataset to recover accuracy from parameters quantization error. The edges-compiler can map nine out of eleven operations to the Edge-TPU, meaning that only input and output float-integer conversions run on the CPU, and the rest of the DNN model operations utilize Edge-TPU resources (see Supplementary Fig. S3c). The compiled model is then transferred into the Coral Dev board. A Python script with blocks shown in (Fig. 4b) is designed to fetch, analyze, and visualize the incoming raw sensory data on the fly. The user interface and dashboard server is implemented using Plotly’s Dash Python library73.

All data that support the findings and conclusions are present in the paper and supplementary materials. Additional data are available from the corresponding author upon reasonable requests.

Vashist, S. K., Luppa, P. B., Yeo, L. Y., Ozcan, A. & Luong, J. H. T. Emerging technologies for next-generation point-of-care testing. Trends Biotechnol. 33, 692–705 (2015).

Article CAS PubMed Google Scholar

Wang, C. et al. Point-of-care diagnostics for infectious diseases: From methods to devices. Nano Today 37, 101092 (2021).

Article CAS PubMed PubMed Central Google Scholar

Yen, Y.-K. & Chiu, C.-Y. A CMOS MEMS-based membrane-bridge nanomechanical sensor for small molecule detection. Sci. Rep. 10, 2931 (2020).

Article ADS CAS PubMed PubMed Central Google Scholar

Bamshad, A. & Cho, H. J. Laserjet printed micro/nano sensors and microfluidic systems: A simple and facile digital platform for inexpensive, flexible, and low-volume devices. Adv. Mat. Technol. 6, 2100401 (2021).

Article Google Scholar

Noviana, E. et al. Microfluidic paper-based analytical devices: From design to applications. Chem. Rev. 121, 11835–11885 (2021).

Article CAS PubMed Google Scholar

Kugimiya, A., Wakimoto, S., Kohda, J., Nakano, Y. & Takano, Y. Development of a one-step analysis method for several amino acids using a microfluidic paper-based analytical device. Sci. Rep. 12, 3427 (2022).

Article ADS CAS PubMed PubMed Central Google Scholar

Yang, S. & Rothman, R. E. PCR-based diagnostics for infectious diseases: Uses, limitations, and future applications in acute-care settings. Lancet Infect. Dis. 4, 337–348 (2004).

Article CAS PubMed PubMed Central Google Scholar

Kibirige, C. N. et al. Development of a sensitive, quantitative assay with broad subtype specificity for detection of total HIV-1 nucleic acids in plasma and PBMC. Sci. Rep. 12, 1550 (2022).

Article ADS CAS PubMed PubMed Central Google Scholar

Carvalho-Correia, E. et al. OmniSARS2: A highly sensitive and specific RT-qPCR-based COVID-19 diagnostic method designed to withstand SARS-CoV-2 lineage evolution. Biomedicines 9, 1314 (2021).

Article CAS PubMed PubMed Central Google Scholar

Notomi, T., Mori, Y., Tomita, N. & Kanda, H. Loop-mediated isothermal amplification (LAMP): Principle, features, and future prospects. J. Microbiol. 53, 1–5 (2015).

Article CAS PubMed Google Scholar

Garg, N., Ahmad, F. J. & Kar, S. Recent advances in loop-mediated isothermal amplification (LAMP) for rapid and efficient detection of pathogens. Curr. Res. Microb. Sci. 3, 100120 (2022).

CAS PubMed PubMed Central Google Scholar

Sampad, M. J. N. et al. Optical trapping assisted label-free and amplification-free detection of SARS-CoV-2 RNAs with an optofluidic nanopore sensor. Biosens. Bioelectron. 194, 113588 (2021).

Article CAS PubMed PubMed Central Google Scholar

Chien, C.-C., Shekar, S., Niedzwiecki, D. J., Shepard, K. L. & Drndić, M. Single-stranded DNA translocation recordings through solid-state nanopores on glass chips at 10 MHz measurement bandwidth. ACS Nano 13, 10545–10554 (2019).

Article CAS PubMed PubMed Central Google Scholar

Cai, H. et al. Optofluidic analysis system for amplification-free, direct detection of Ebola infection. Sci. Rep. 5, 14494 (2015).

Article ADS CAS PubMed PubMed Central Google Scholar

Du, K. et al. Multiplexed efficient on-chip sample preparation and sensitive amplification-free detection of Ebola virus. Biosens. Bioelectron. 91, 489–496 (2017).

Article CAS PubMed PubMed Central Google Scholar

Ozcelik, D. et al. Optofluidic wavelength division multiplexing for single-virus detection. PNAS 112, 12933–12937 (2015).

Article ADS CAS PubMed PubMed Central Google Scholar

Meena, G. G. et al. 7X multiplexed, optofluidic detection of nucleic acids for antibiotic-resistance bacterial screening. Opt. Exp 28, 33019–33027 (2020).

Article CAS Google Scholar

Black, J. A., Ganjalizadeh, V., Parks, J. W. & Schmidt, H. Multi-channel velocity multiplexing of single virus detection on an optofluidic chip. Opt. Lett. 43, 4425–4428 (2018).

Article ADS CAS PubMed PubMed Central Google Scholar

Liu, R., Wang, N., Kamili, F. & Sarioglu, A. F. Microfluidic CODES: A scalable multiplexed electronic sensor for orthogonal detection of particles in microfluidic channels. Lab Chip 16, 1350–1357 (2016).

Article CAS PubMed Google Scholar

Liu, R. et al. Design and modeling of electrode networks for code-division multiplexed resistive pulse sensing in microfluidic devices. Lab Chip 17, 2650–2666 (2017).

Article CAS PubMed Google Scholar

Liu, R., Wang, N., Asmare, N. & Sarioglu, A. F. Scaling code-multiplexed electrode networks for distributed Coulter detection in microfluidics. Biosens. Bioelectron. 120, 30–39 (2018).

Article CAS PubMed Google Scholar

Ahuja, K. et al. Toward point-of-care assessment of patient response: A portable tool for rapidly assessing cancer drug efficacy using multifrequency impedance cytometry and supervised machine learning. Microsyst. Nanoeng. 5, 1–11 (2019).

Article CAS Google Scholar

Sui, J., Xie, P., Lin, Z. & Javanmard, M. Electronic classification of barcoded particles for multiplexed detection using supervised machine learning analysis. Talanta 215, 120791 (2020).

Article CAS PubMed Google Scholar

Xie, P., Cao, X., Lin, Z. & Javanmard, M. Top-down fabrication meets bottom-up synthesis for nanoelectronic barcoding of microparticles. Lab Chip 17, 1939–1947 (2017).

Article CAS PubMed Google Scholar

Xie, P. et al. Processing gain and noise in multi-electrode impedance cytometers: Comprehensive electrical design methodology and characterization. Sens. Actuators, B Chem. 241, 672–680 (2017).

Article CAS Google Scholar

Wang, N., Liu, R., Asmare, N., Chu, C.-H. & Sarioglu, A. F. Integrated sensor networks with error correction for multiplexed particle tracking in microfluidic chips. Biosens. Bioelectron. 174, 112818 (2021).

Article CAS PubMed Google Scholar

Li, Y. et al. Deep cytometry: Deep learning with real-time inference in cell sorting and flow cytometry. Sci. Rep. 9, 11088 (2019).

Article ADS CAS PubMed PubMed Central Google Scholar

Raji, H., Tayyab, M., Sui, J., Mahmoodi, S. R. & Javanmard, M. Biosensors and machine learning for enhanced detection, stratification, and classification of cells: a review. Biomed. Microdev. 24, 26 (2022).

Article Google Scholar

Ganjalizadeh, V. et al. Fast custom wavelet analysis technique for single molecule detection and identification. Nat. Commun. 13, 1–9 (2022).

Article Google Scholar

McCulloch, W. S. & Pitts, W. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 5, 115–133 (1943).

Article MathSciNet MATH Google Scholar

Valiant, L. G. A theory of the learnable. Commun. ACM 27, 1134–1142 (1984).

Article MATH Google Scholar

Duda, R. O. & Hart, P. E. Pattern Classification and Scene Analysis Vol. 3 (Wiley, 1973).

MATH Google Scholar

Fukushima, K. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol. Cybern. 36, 193–202 (1980).

Article CAS PubMed MATH Google Scholar

Talaviya, T., Shah, D., Patel, N., Yagnik, H. & Shah, M. Implementation of artificial intelligence in agriculture for optimisation of irrigation and application of pesticides and herbicides. Artif. Intell. Agric. 4, 58–73 (2020).

Google Scholar

Xiang, X., Li, Q., Khan, S. & Khalaf, O. I. Urban water resource management for sustainable environment planning using artificial intelligence techniques. Environ. Impact Assess. Rev. 86, 106515 (2021).

Article Google Scholar

Ma, Y., Wang, Z., Yang, H. & Yang, L. Artificial intelligence applications in the development of autonomous vehicles: A survey. IEEE/CAA J. Autom. Sin. 7, 315–329 (2020).

Article Google Scholar

Sterne, J. Artificial Intelligence for Marketing: Practical Applications (Wiley, 2017).

Book Google Scholar

Mak, K.-K. & Pichika, M. R. Artificial intelligence in drug development: Present status and future prospects. Drug Discov. Today 24, 773–780 (2019).

Article PubMed Google Scholar

Riedl, M. O. & Young, R. M. Narrative planning: Balancing plot and character. jair 39, 217–268 (2010).

Article MATH Google Scholar

Moloi, T. & Marwala, T. Artificial Intelligence in Economics and Finance Theories (Springer, 2020).

Book Google Scholar

Amato, F. et al. Artificial neural networks in medical diagnosis. J. Appl. Biomed. 11, 47–58 (2013).

Article CAS Google Scholar

Sun, J., Peng, Y., Guo, Y. & Li, D. Segmentation of the multimodal brain tumor image used the multi-pathway architecture method based on 3D FCN. Neurocomputing 423, 34–45 (2021).

Article Google Scholar

van der Lubbe, M. F. J. A. et al. A non-invasive, automated diagnosis of Menière’s disease using radiomics and machine learning on conventional magnetic resonance imaging: A multicentric, case-controlled feasibility study. Radiol. Med. 127, 72–82 (2022).

Article PubMed Google Scholar

Ström, P. et al. Artificial intelligence for diagnosis and grading of prostate cancer in biopsies: A population-based, diagnostic study. Lancet Oncol. 21, 222–232 (2020).

Article PubMed Google Scholar

Tanveer, M. et al. Machine learning techniques for the diagnosis of Alzheimer’s disease: A review. ACM Trans. Multimed. Comput. Commun. Appl. 16, 30:1-30:35 (2020).

Article Google Scholar

Cui, F., Yue, Y., Zhang, Y., Zhang, Z. & Zhou, H. S. Advancing biosensors with machine learning. ACS Sens. 5, 3346–3364 (2020).

Article CAS PubMed Google Scholar

Cao, K., Liu, Y., Meng, G. & Sun, Q. An overview on edge computing research. IEEE Access 8, 85714–85728 (2020).

Article Google Scholar

Edge TPU performance benchmarks. Coral https://coral.ai/docs/edgetpu/benchmarks/.

Soldano, L. B. & Pennings, E. C. M. Optical multi-mode interference devices based on self-imaging: principles and applications. J. Lightwave Technol. 13, 615–627 (1995).

Article ADS Google Scholar

Xu, Y., Du, J., Dai, L.-R. & Lee, C.-H. A regression approach to speech enhancement based on deep neural networks. IEEE/ACM Trans. Audio Speech Lang Process 23, 7–19 (2014).

Article CAS Google Scholar

Yousefi, M. & Hansen, J. H. L. Block-based high performance CNN architectures for frame-level overlapping speech detection. IEEE/ACM Trans. Audio Speech Lang Process. 29, 28–40 (2021).

Article Google Scholar

Keerthi Krishnan, K. & Soman, K. P. CNN based classification of motor imaginary using variational mode decomposed EEG-spectrum image. Biomed. Eng. Lett. 11, 235–247 (2021).

Article CAS PubMed PubMed Central Google Scholar

Huang, J., Chen, B., Yao, B. & He, W. ECG arrhythmia classification using STFT-based spectrogram and convolutional neural network. IEEE Access 7, 92871–92880 (2019).

Article Google Scholar

Sultana, F., Sufian, A. & Dutta, P. Advancements in image classification using convolutional neural network. in 2018 Fourth International Conference on Research in Computational Intelligence and Communication Networks (ICRCICN), pp. 122–129 (2018). https://doi.org/10.1109/ICRCICN.2018.8718718.

Lecun, Y., Bottou, L., Bengio, Y. & Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 86, 2278–2324 (1998).

Article Google Scholar

Kingma, D. P. & Ba, J. Adam: A method for stochastic optimization. Preprint at http://arxiv.org/abs/1412.6980 (2017).

Technology. Coral https://coral.ai/technology/.

Dev Board datasheet. Coral https://coral.ai/docs/dev-board/datasheet/#certifications.

Abadi, M. et al. TensorFlow: Large-scale machine learning on heterogeneous distributed systems. https://arxiv.org/abs/1603.04467 (2016).

Sarkar, D., Bali, R. & Ghosh, T. Hands-On Transfer Learning with Python: Implement Advanced Deep Learning and Neural Network Models Using TensorFlow and Keras (Packt Publishing Ltd, 2018).

Google Scholar

Gisbrecht, A. & Hammer, B. Data visualization by nonlinear dimensionality reduction. WIREs Data Min. Knowl. Discov. 5, 51–73 (2015).

Article Google Scholar

Wold, S., Esbensen, K. & Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 2, 37–52 (1987).

Article CAS Google Scholar

Ringnér, M. What is principal component analysis?. Nat. Biotechnol. 26, 303–304 (2008).

Article PubMed Google Scholar

Van der Maaten, L. & Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579–2605 (2008).

MATH Google Scholar

Abdi, H. & Williams, L. J. Principal component analysis. Wiley Interdiscip. Rev.: Comput. Stat. 2, 433–459 (2010).

Article Google Scholar

Ganjalizadeh, V. & Schmidt, H. P. C. W. A. A fast custom wavelet analysis technique for single molecule detection and identification. Nat. Commun. https://doi.org/10.5281/zenodo.5794624 (2021).

Chen, J. & Ran, X. Deep learning with edge computing: A review. Proc. IEEE 107, 1655–1674 (2019).

Article Google Scholar

Stambaugh, A. et al. Optofluidic multiplex detection of single SARS-CoV-2 and influenza A antigens using a novel bright fluorescent probe assay. PNAS 118, e2103480118 (2021).

Article CAS PubMed PubMed Central Google Scholar

Meena, G. G. et al. 3× multiplexed detection of antibiotic resistant plasmids with single molecule sensitivity. Lab Chip 20, 3763–3771 (2020).

Article CAS PubMed PubMed Central Google Scholar

Meena, G. et al. Ultrasensitive detection of SARS-CoV-2 RNA and antigen using single-molecule optofluidic chip. Apl Photon 6, 066101 (2021).

Article ADS CAS Google Scholar

Ioffe, S. & Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. http://arxiv.org/abs/1502.03167 (2015).

Wager, S., Wang, S. & Liang, P. S. Dropout training as adaptive regularization. in Advances in Neural Information Processing Systems vol. 26 (Curran Associates, Inc., 2013).

Dabbas, E. Interactive Dashboards and Data Apps with Plotly and Dash: Harness the power of a fully fledged frontend web framework in Python–no JavaScript required (Packt Publishing Ltd., 2021).

Google Scholar

Download references

We acknowledge Thomas A. Wall for his help with fabricating the ARROW device. This work was supported by a gift from the Cisco University Research Program Fund #2020-224954 (H.S.) as well as the National Institutes of Health under grants 1R01AI116989-01 (A.R.H., H.S.), 1R01EB028608 (H.S., A.R.H.), and National Science Foundation under grant CBET-1703058 (H.S.), respectively.

School of Engineering, University of California, Santa Cruz, 1156 High Street, Santa Cruz, CA, 95064, USA

Vahid Ganjalizadeh, Gopikrishnan G. Meena & Holger Schmidt

Electrical and Computer Engineering Department, Brigham Young University, Provo, UT, 84602, USA

Matthew A. Stott & Aaron R. Hawkins

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

V.G. and H.S. designed the research; M.A.S. fabricated the optofluidic chips; G.G.M. performed the experiments; V.G. developed the machine-learning framework; V.G. and H.S. performed the data analysis; V.G. visualized the data and results; V.G., A.R.H., H.S. wrote the paper; H.S. supervised the research.

Correspondence to Holger Schmidt.

A.R.H. and H.S. had a competing interest in Fluxus Inc which commercializes optofluidic technology. All other authors declare they have no competing interests.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

Ganjalizadeh, V., Meena, G.G., Stott, M.A. et al. Machine learning at the edge for AI-enabled multiplexed pathogen detection. Sci Rep 13, 4744 (2023). https://doi.org/10.1038/s41598-023-31694-6

Download citation

Received: 29 June 2022

Accepted: 15 March 2023

Published: 23 March 2023

DOI: https://doi.org/10.1038/s41598-023-31694-6

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.