Repository landing page

We are not able to resolve this OAI Identifier to the repository landing page. If you are the repository manager for this record, please head to the Dashboard and adjust the settings.

Pragmatic Evaluation of Health Monitoring & Analysis Models from an Empirical Perspective

Abstract

Implementing and deploying several linked modules that can conduct real-time analysis and recommendation of patient datasets is necessary for designing health monitoring and analysis models. These databases include, but are not limited to, blood test results, computer tomography (CT) scans, MRI scans, PET scans, and other imaging tests. A combination of signal processing and image processing methods are used to process them. These methods include data collection, pre-processing, feature extraction and selection, classification, and context-specific post-processing. Researchers have put forward a variety of machine learning (ML) and deep learning (DL) techniques to carry out these tasks, which help with the high-accuracy categorization of these datasets. However, the internal operational features and the quantitative and qualitative performance indicators of each of these models differ. These models also demonstrate various functional subtleties, contextual benefits, application-specific constraints, and deployment-specific future research directions. It is difficult for researchers to pinpoint models that perform well for their application-specific use cases because of the vast range of performance. In order to reduce this uncertainty, this paper discusses a review of several Health Monitoring & Analysis Models in terms of their internal operational features & performance measurements. Readers will be able to recognise models that are appropriate for their application-specific use cases based on this discussion. When compared to other models, it was shown that Convolutional Neural Networks (CNNs), Masked Region CNN (MRCNN), Recurrent NN (RNN), Q-Learning, and Reinforcement learning models had greater analytical performance. They are hence suitable for clinical use cases. These models' worse scaling performance is a result of their increased complexity and higher implementation costs. This paper compares evaluated models in terms of accuracy, computational latency, deployment complexity, scalability, and deployment cost metrics to analyse such scenarios. This comparison will help users choose the best models for their performance-specific use cases. In this article, a new Health Monitoring Metric (HMM), which integrates many performance indicators to identify the best-performing models under various real-time patient settings, is reviewed to make the process of model selection even easier for real-time scenarios

Similar works

Full text

thumbnail-image

International Journal on Recent and Innovation Trends in Computing and Communication

redirect
Last time updated on 25/10/2023

Having an issue?

Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.