Next Article in Journal
The Italian Validation of the Level of Exposure-Dental Experiences Questionnaire
Next Article in Special Issue
A Novel Digital Modulation Recognition Algorithm Based on Deep Convolutional Neural Network
Previous Article in Journal
Supervised Learning of Natural-Terrain Traversability with Synthetic 3D Laser Scans
Previous Article in Special Issue
Heterogeneous Defect Prediction Based on Transfer Learning to Handle Extreme Imbalance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tele-Treatment Application Design for Disable Patients with Wireless Sensors

by
Alberto Arteta Albert
1,†,
Luis Fernando de Mingo López
2,*,† and
Nuria Gómez Blas
2,†
1
Department of Computer Science, College of Arts and Sciences, Troy University, Troy, AL 36082, USA
2
Departamento de Sistemas Informáticos, Escuela Técnica Superior de Ingeniería de Sistemas Informáticos, Universidad Politécnica de Madrid, 28031 Madrid, Spain
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2020, 10(3), 1142; https://doi.org/10.3390/app10031142
Submission received: 10 December 2019 / Revised: 24 January 2020 / Accepted: 4 February 2020 / Published: 8 February 2020
(This article belongs to the Special Issue Wireless Sensor Networks in Smart Environments)

Abstract

:
This paper consists of the development of a system to help patients with different disabilities, affected by rare or chronic diseases or any kind of dependence through tele assistance, virtual interaction and intelligent monitoring. The main goal is to increase the quality of life of the minorities who cannot take full advantage of the healthcare system by providing an alternative way of monitoring them with the technology embedded in this paper. The result of the paper is not intended to be a single solution, but a modular system that allows the construction of an application that is able to measure the needs of a health administration and the patients. The paper also pursues an educational training to the facultative trainees in a new way to approach patient treatments. It can improve the quality of life of the patients by saving them time and other resources in moving to the Health center and the professionals can also save time as they can take advantage of the online treatments by using the proposed system.

1. Introduction

Telehealth care has revealed itself as a revolutionary high-quality approach to help to treat patients [1]. There are currently several implementations of systems that help to cope with patients that do not enjoy the same mobility as the majority of the population [2]. This seems to evolve and to grow in the last decade [3]. For example, Konstantas et al. [4] introduced a proposal of mobile monitoring establishing the fact that technology can offer promising results to remote patients as a support for conventional treatment. The theoretical study is to complete although the practical application of such system still shows several gaps in the communication channels of the system. Other authors, van Halteren et al., proposed a system that introduce the specifics of a module of telemonitoring by the use of sensors and other smart devices that collect signals from the patients [5]. This proposal, however, lacks of interactive, integrative and communication modules and it does not provide intelligent and automatic processing, such as image processing and computer vision alternatives. Besides, there is barely any use of social media.
As Smart devices and mobile apps are considered one of the first solid model applied to telehealth care, this paper complies with the following services:.
  • Integrate into the audiovisual media system (mainly television, and smart phones because of its ease of use and its easy penetration in the domestic environment); although other elements or devices can be utilized as long as there is a physical channel of bidirectional communication between users, professionals and/or relatives.
  • Provide mechanisms for assistance to the user by professionals, relatives or friends in different fields, such as the distribution of multimedia contents, videos, pictures, articles, etc.
  • The user takes advantage of the TV or any other output device. The time the user spends on watching TV can be utilized for receiving alerts and gives the user the possibility of watching recommended videos that can supplement the therapies or treatments. Below there are some examples.
  • Create alarms and warnings in the follow-up of the medication intake.
  • Support the rehabilitation with the programming of multimedia contents to the user.
  • Provide tele-accompaniment. The end user can be added a social network of patients with the same symptoms, so they can share experiences and tips to improve their results.
  • Personalize programs with multimedia contents to the user. These consist of a composition of social and professional oriented materials to be utilized by the user’s knowledge and supervised by the professional.
  • Develop monitoring systems mostly with the gyroscopes and wireless sensors connected to Smartphones and Smartwatches. This detects Situations and potential Conflicts that deserve special attention.
  • Develop alarm systems and primary alert. The user has the option of notifying the need of assistance by using his watch or any other device.
  • Construct a user notification system that sends reminders to take the right medicines, to review the materials the professional gathered and to be in contact with the supervisor.
  • Develop of a modular architecture, so that the system is adaptable to customer needs.
  • Define and to implement the access levels of Professionals in order to adapt the system to the current policies of the professionals and client institutions.

2. Related Work

Undoubtedly, the use of technology in Healthcare system is growing rapidly. Monitoring tools, software to support diagnosis, markers counters are used more and more by the professionals in hospitals and Health centers.
In particular, remote assistance has been studied in the last years as a support for traditional treatments [6]. A mobile healthcare system is described with a fair number of details. This solution allows patients to be cared for in their preferred location, typically at home. it provides patient education and support for preventive care; for example, to those trying to lose weight by using videophone or web based clinical consultations, replacing routine visits such as face to face annual reviews can be a good option. It also Improves the efficiency of other treatments more conventional. Proactive education and support, such as via web forums, may facilitate self-management and help with coping (for example, in people with chronic back pain) or prevent exacerbations of conditions such as asthma or any other problem triggered by stress. Use of monitoring techniques can enable earlier detection of disease exacerbations, thereby facilitating timely management and support. It allows greater opportunities for continuity of care with the same clinician and more frequent assessments. Furthermore, mobile healthcare is known to reduces costs for patients by obviating the need for time off school or work and due to the time travel. As complete as [6] is, still misses a lot of features that can be very helpful to the patients; for instance, the monitoring component only deals with basic bio signals and has no intelligent processing nodule. Moreover, the system does not have an option for a full integration with the TV which is one of the preferred environments for reduce mobility patients. The social schedule also is very poor. A complete social schedule with a hub of contacts with family and other patients, can make the whole process more enjoyable as the patient can share thoughts and accomplishments with other patients and friends.
Regarding automatic signal and image processing no major telehealth care systems have been fully develop beyond experimental trials. Arteta et al. [7] designed a new approach to improve experimental CPU’s that can speed up the pattern recognition processing applicable to image and computer vision in smart devices. The biggest problem this model faces is finding the Hardware/Bioware to implement such model [8].
The evolution of telehealth care that are described in Wac [3] is the basis. The requirement and possibilities are shown in this work. It aims for a more personalized and decentralized approach in helping patients. In particular, Wac [3] shows the trend of taking advantage of the technology to reduce the mobilization of patients reducing costs in terms of time, money and resources usage.
Smartphone apps have been played an important role in telehealth care too [9]. Several cases of use are shown such as information and time management, health records maintenance, communication and consulting patient monitoring, among others [10,11,12]. However, the approach shown by Dyer et al. [9] is very simple in terms of patient monitoring. Besides, the communication channels are roughly described. The system does not provide a logging server either, neither a communication server nor a centralized repository. In other solutions, image processing [13,14] and computer vision [14,15] have been tested for monitoring patients, but the design has not been implemented to be used for end users. In these cases, the main goals are not providing customized and personalized assistance to patients with disabilities or similar needs.
There are several attempts of recognizing automatic suspicious behaviors in patients by using image processing and computer vision [16]. This work has been very specialized and still needs improvements. Especially for movements that are supposed to be normal but follow after other ones that can mean a potential problem for the patient. Experimental trials with deep learning solutions [17] appear to be promising. The tradeoff is the computational costs and the amount of data required to create a proper learning model that classify correctly the rough movements of a patience.
Streaming multimedia content in healthcare systems over Internet/Intranet [18] has also been a key component in the last decade. A theoretical framework proposed by [19] to have distributor system that delivers the multimedia contents in the healthcare system to doctors and professionals involves a number of different features; however, social schedule and a reliable communication system are not fully functional. There is no monitoring module either, and still lack of many concepts to build a monitoring system as the one presented here. Furthermore, the use of social media apps was not been fully exploited in any of the related projects. The use of a social network integrated in the social agenda is not present in the telehealth care systems.
According to the World Health Organization, falls are the second most common cause of accidental or unintentional injury deaths worldwide. And a staggering 37.3 million falls are severe enough to require medical attention each year. A wide range of factors contributes to fall risks, such as environmental hazards, illness, neurological, mental conditions, muscle weakness, poor vision, and medication [10,20,21,22,23]. While there is no such thing as a 100 % effective fall detection or prevention solution, most experts agree that a combination of available sensor technology can greatly increase system reliability. And there is a growing acceptance by seniors as monitoring is a means to extend their independent living opportunities [21,23].
There is a system created and funded by European Agency: Ministry of industry, tourism and commerce, Spain. This project gets different users connected via smartphones and shares contents of TV programs that are currently on [24]. This is a way of sharing TV contents that are on and also of filtering the commercials based on the profile of the user. This system was not intended to be applied to telehealth care, although the technology and the idea can be used in the communication between the professional and the user to build a social agenda. Another application in telehealth care granted by the Ministry of industry, Energy and Tourism is introduced in Castellanos [25]. The system still requires a lot of improvements such as computer vision, multiple user interaction, and a complete social schedule. These three ideas can contribute with a major change in traditional tele treatment approaches. These three are in the top of the emerging and promising technologies and can improve dramatically the monitoring and the living quality of these minorities during the rehab process. Moreover, it is a small project, used as a demo in hospitals in different regions intended to monitor a small number of patients. This paper targets a bigger number of monitored patients by updating the resources of the communication servers. This is essential not only to increase the number of end users but also to keep technologies such as machine learning algorithms and computer vision running efficiently.
A novel algorithm for biomedical image segmentations is introduced together with an algorithm for biomedical image feature selection in Zhao and Belkasim [26]. Biomedical image segmentations and feature selection are foundations to detect and predict diseases. Discovering biomedical knowledge from medical text documents is a critical means for advancing medical practice and clinical outcomes in patients. A novel domain specific methodology is built to identify the word meanings in medical documents characterized by different domains [27]. This methodology uses a domain fusion algorithm to narrow domain concepts from different domain knowledge and avoids unknown domain problem; it can be effectively used for both biomedical information retrieval and text mining. An approach based on image compression and deposing scheme is shown in  Zhao and Belkasim [26], Zhao et al. [28]. Other authors [29,30] present another proposal done for digging the knowledge that hidden in a big amount of text data. This proposed project needs a further research on social network [31] and image retrieval in Internet [32] in such a way that they can contribute to it, along with image processing methods [33,34]. In addition, the biomedical data usually has uncertain information [35] that are indeed very helpful for this project [7,36,37,38,39,40,41,42].

3. Methodology

Several challenges have been arisen, key issues based upon economic, technological and social sides are briefly described.
System Development Cost: Apart from the required medical issues, enabling the population to benefit from the information and communication technologies generates additional costs. Both technologies are core infrastructure and a large amount of budget is required to invest in necessary computer equipment and developing tools. Hardware and software installation and maintenance together with qualified manpower and knowledge transfer demonstrate one-time and periodic investment.
System Implementation: Telemedicine affects traditional medical care processes and several social concerns may arise. Key concerns including legal, regulatory, security and human resources issues together with successful implementation of the telemedicine services are addressed. Building a telemedicine service is therefore not only about completing the system implementation but also about the impacts of system utilization.
Diagnostic Accuracy: Accuracy of the diagnosis is one of the key concerns especially when the physicians and the patients are at different locations. Face-to-face medical care usually brings conffidence to patients, especially the elderly. However, a short period of medical interrogation and vital sign measurement performed at a hospital may not reflect current symptoms due to the white coat syndrome. Physicians may obtain more accurate results by applying continuous health monitoring based upon electronic healthcare approach.
Telehealth care can be used to deliver medical services regardless of distance and time via information and communication technologies. Feasible opportunities of telehealth care include, among others, cost reduction, preventive medicine promotion, medical education, healthcare equality, service diversity, etc.

3.1. Architecture Principles

Architecture principles define the underlying general rules and guidelines for the use and deployment of all Information Technology (IT) resources and assets across the enterprise. They reflect a level of consensus among the various elements of the enterprise, and form the basis for making future IT decisions, see [43]. Each architecture principle should be clearly related back to the business objectives and key architecture drivers.
Principles are general rules and guidelines, intended to be enduring and seldom amended, that inform and support the way in which an organization sets about fulfilling its mission. In their turn, principles may be just one element in a structured set of ideas that collectively define and guide the organization, from values through to actions and results. Depending on the organization, principles may be established at any or all of three levels:
  • Enterprise principles provide a basis for decision-making throughout an enterprise, and inform how the organization sets about fulfilling its mission. Such enterprise-level principles are commonly found in governmental and not-for-profit organizations, but are encountered in commercial organizations also, as a means of harmonizing decision-making across a distributed organization. In particular, they are a key element in a successful architecture governance strategy (see Architecture Governance).
  • Information Technology (IT) principles provide guidance on the use and deployment of all IT resources and assets across the enterprise. They are developed in order to make the information environment as productive and cost-effective as possible.
  • Architecture principles are a subset of IT principles that relate to architecture work. They reflect a level of consensus across the enterprise, and embody the spirit and thinking of the enterprise architecture. Architecture principles can be further divided into:
    Principles that govern the architecture process, affecting the development, maintenance, and use of the enterprise architecture
    Principles that govern the implementation of the architecture, establishing the first tenets and related guidance for designing and developing information systems
These sets of principles form a hierarchy, in that IT principles will be informed by, and elaborate on, the principles at the enterprise level; and architecture principles will likewise be informed by the principles at the two higher levels.
It is useful to have a standard way of defining principles. In addition to a definition statement, each principle should have associated rationale and implications statements, both to promote understanding and acceptance of the principles themselves, and to support the use of the principles in explaining and justifying why specific decisions are made [43].
In the context of the paper, see Figure 1, three major profiles are defined:
  • Final user: This is the main agent of the product. These are those minorities affected by chronic diseases, people with disabilities and special needs, or with any type of dependency.
  • Client: It is any entity, association, Pharmaceutical laboratory, Hospital or Health center that Implants the system to provide assistance to end users.
  • Professional. Any person providing the service (volunteers, social workers, health workers, etc.).

3.2. Cases of Study

Following are the different scenarios the system considers:
  • Scenario 1: The professional receives an alert on their device with a motion alert that a rough falling movement has occurred. It activates the communication between the professional and the patient and check the status of the user.
  • Scenario 2: The professional schedules the alerts that reminds the user to take their medication: The user visualizes and listens to this alerts on their TV. The alerts can include the visualization of recommended videos explaining more about the benefits of the medicines and more recommendations.
  • Scenario 3: The user need to contact with other users. The professional can contact him and provide networking tools with other users in the same situation and can share experiences through the use of multimedia content, if that is what they decide. This is supervised by the professional.
  • Scenario 4: The device installed at the user’s home has relevant information from them, such as date, body temperature, and next appointment with the professional. They can check the device or that can be always present in the corner of the TV.
  • Scenario 5: The user’s family can work together with the professional to recommend audiovisual contents in the system for the user.
  • Scenario 6: The professional contact the final user (patient) to check on his daily exercises. He supervises online the movements of the user during the rehabilitation process.

3.3. System Components

The system is a combination of different assembled parts (some of them, can be disabled depending on the needs of the user):
  • Recommendation module. A customized menu of options of recommended action to their users based on their preferences.
  • Monitoring module: User is monitored directly (with a device in the smartphone) and indirectly (with cameras and external devices integrated in TV). There are movement detection and monitoring of the user’s movement that detects rough motion, such as falls. (Computer vision module).
  • Tele Rehab: The rehabilitation to recover the mobility with the help of supervisor/professional
  • Tele dosage: The Assistance can be focused on the medicine intake and how to control the dosage by the professional through the system.
  • Social scheduling: A customized program of contents for the user by the system agents (family, volunteers, social agents, professionals and so on) The Architecture of the system is intended to be client /server. The professionals will use a webapp and the users AppAndroid. Both will connect to the backend part with a specific Application Programming interface. (API REST)

3.4. Technical Components (Used Technology)

A brief technical description of the system is found below, see Figure 2.
  • Backend + API REST: The system backend will have an Application Programming Interface available, that centralizes and coordinates all systems actions from the server side. Such API is able to provides an interface for the communication. For the frontend initially, the project use Python, Django and Tastypie.
  • Webapp: The web portal for the professional management is implemented by using HTML5, CSS3, JavaScript, jQuery and possibly Twitter Bootstrap. The communication with this part is managed by the output objects from the API.
  • AppAndroid: AppAndroid for the user is installed in the device connected to the TV (set top-box) and/or smartphone, they are linked to the API with communication objects (possibly JSON) same way as the Webapps does.
  • Database Server (Backend): The database is a relational one, initially MySQL, although the structure might change to a BI model depending on the number of signals to be stored for any user. This is located in a dedicated stand-alone server and accepts Secure-Shell connections (ssh command line utility). In the future, we might consider the option to migrate to a SQL Server (stand-alone) server depending on the growth of data and the needs for storing it.
  • Communication Server: The communication server will keep track of the communication flow and logs the activity between the different components. Elastix technology is a good candidate as it allows VoiP, PBX emails and twits and it does not rely on external sources to process and modulate the calls. Elastix also uses SIP as standard interactive multimedia calls such as video, voice o messaging.
The Client is sending a Session Initiation Protocol (SIP) package to the User Agent Client (UAC) that will initiate the User Server Agent (UAS).
The Proxy server SIP Proxys will reroute the SIP to the UAS and returns the SIP to the UAC. Any request can go through different proxy servers to the UAS for allocating and redirecting server to locate the user and to notify to the rest of the components. This process is the ideal way of communication, although we might consider different alternatives. The system also provides a logging server that logs the physical address. The physical and logical address is associated in a routing table located in the logging server. Internet calls will use SIP, Android client ImsDroid (Open source software), Client Html5 Sip and Server WebRTC2SIP. For signal processing in the communication server we use an experimental simulation of a massive parallel bioinspired system, which will allow to work on new approach for mobile computing. Initially will try P-system [8] architecture in conventional hardware.
The methods can be applicable in an integrated platform of visual and communication devices. The hardware components for the system agents are:
  • Dedicated server for training the Computer vision mode: Desktop Server 2 GPU units for training a Convolutional Neural Network (CNN).
  • Standby Server part: Remote server for logging communication messages and to synchronize the system
  • Professional module: Video camera, Smart device
  • User module: Video camera, Boxes, TV or any other domestic smart device, smartphone
The target system requires an internet connection and as many user modules as patients who will to use the system, see Figure 3.

4. Details of Interaction

Users are engaged in a passive (1) and active way (2).
  • While watching TV and being monitored by the professional on the camera inserted on the TV device. The professional contacts the patient to check on progress of the rehabilitation process.
  • The user actively contacts either the professional or the other users who are part of the social schedule module of the application integrated on the TV box.
The monitoring will be done by using computer vision techniques. A Convolutional Neural Network (CNN) is a deep learning algorithm that can recognize and classify features in images for computer vision. It is a multi-layer neural network designed to analyze visual inputs and perform tasks such as image classification, segmentation and object detection. The software used is Keras which is used as a friendly user interface for using tensor flow library. This contains numerous algorithms for building a convolutional network that are designed for computer visions. The idea is training the CNN with images of a user failing. Initially 15,000 pictures of possible falls are part of the training process. The camera feeds pictures to the computer vision’s algorithm.

4.1. Monitoring with Computer Vision Details

The multi model training system to dynamically train the captured images. People suffered from different diseases/illness may have different ways to fall and stay; different training models have different advantages. Our multi-model training system will use several popular machine learning methods, such as support vector machine, k-means clusters and deep neural network to train the captured images, then combine the results to get the final decision. This process is dynamical, it focuses on the people and the environment. Deep neural network has been widely applied. Although deep neural network has good performance, the training process is still black box. Support Vector Machines (SVM) and K-means clustering are traditional methods whose mathematics background have been well studied. Through comparing the results from SVM and K-means clustering with the ones from deep neural network [44,45], we may get more insights of the mathematics principles hidden in the deep neural network, which can contribute to the improvement of deep neural network as well as the machine learning. In our case, there are only one scenario only to consider, making the computer vision approach fairly fairly simple. Training is done by capturing a series of images from the users that simulate a fall and stay in the floor for different users and taking serially generated images of falling from internet too.
The elective system is using a convolutional network with the following stages:

4.1.1. Training Phase

There will be both simulations and images scanned by different users falling and reuse of other images and videos of falling individuals. During that phase we will train the model by initially using 15,000 images of a person falling. During the training the system will take into account the position of the person falling and the speed of movement, configuring the system for alerting when the speed of the motion drastically changes and the sequential positions generated by the individual fit one of the predefined falling sequences positions defined in the filter of the convolutional layer of the CNN.
The CNN architecture, see Figure 4 will follow a standard that can be adjusted to optimize fitness values.
The input layer will take 3 depth images 280 × 280 × 3. The image is classified as falling/not falling (2 classes, or labels). This classification is done done manually for 15,000 pictures (3-dimensional vectors).

4.1.2. Testing Phase

For testing initially, we considered to use a cross validation approach with a data set split parameter ν = 0.2 . (corresponding to a split into 5 datasets of 3000 pictures each and use 12,000 pictures for training and 3000 for testing in each round, 5 rounds total). This process returned an acceptable rate of hits, detecting successful falls with accuracy (true positives) of 72.2%. False negatives are only 8% of the distribution.

5. Evaluation and Validation of the Developed System

The performance of proposed fall detection system was tested in a real home environment. An external camera is used for recording the real video sequence with the image size of 280 × 280. The recorded video sequence is processed by using OpenCv and Keras for the CNN training.
For evaluating this fall detection system, each volunteer (student) is asked to simulate different falling and non-fall activities in different directions. Fall is detected when the lying posture is classified as falling by the CNN, see Table 1. The proposed CNN based fall detection systems could accurately detect falls (75%) with a low false alarm rate (25%). The misclassifications are attributed to the reason that the bend posture in some directions is viewed similar to the lie posture and we can adopt multiple cameras to capture different posture views for ameliorating it.
The accuracy of the CNN is bellow 75%, some preprocessing effort could be made to optimize CNN structure and improve the results.Yu et al. [46] obtains a better performance but the images need to be preprocess in order to generate a silhouette. From a recorded video stream, background subtraction is applied to extract the human body silhouette. Extracted silhouettes corresponding to daily activities are applied to construct a convolutional neural network, which is applied for classification of different classes of human postures and detection of a fall event.
Rates (72.2%–75%) have been the maximum accuracy rate reached considering the standard CNN topology and a variable data set split factor (0.15–0.28). Results suggest that with these numbers of images the accuracy might not be able to improve considerably. Thus, more training images would be needed to ensure a more reliable system. Nevertheless, considering the number of training observations, and the outcome of using the cross-validated test, this module clearly adds value to the entire system.
In reality, only part of falling human body is included in some images, this technique detect the falling events provided a 80% of body is shown in images. A future research includes information from multi-modality sensors together with deep learning techniques for developing a more robust fall detection system in case the whole human body is not present in the images.

6. Conclusions

This paper has introduced a system designed for helping in remote medical treatments by offering tele assistance and tele support through different modules of the system. In particular, wireless sensors and capturing devices are integrated by collecting biosignals and images from the users at the same time. The biosignals are able to help the professional in being warned about the biometrics of patients and the images are able to catch a possible fall of the patient right in the moment. Apart from that, systems like this offer a new possibility for remotely treat patients with technological solutions that integrate wireless sensors, Desktop Servers, smartphones, Machine learning algorithms, and traditional databases. Wireless devices together with Machine Learning algorithms will play an essential role in the next generation of telehealth care applications. Future work includes exploitation information from multi-modality sensors (such as audio and video sensors) together with deep learning techniques for developing a more robust fall detection system.

Author Contributions

Conceptualization, A.A.A. and N.G.B.; Formal analysis, A.A.A. and N.G.B.; Investigation, L.F.d.M.L. and N.G.B.; Methodology, A.A.A. and L.F.d.M.L.; Validation, L.F.d.M.L.; Writing—original draft, A.A.A. and N.G.B.; Writing—review & editing, L.F.d.M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding, but it is a considerable upgrade of the system developed in project TSI-020605-2012-40 (Can@l de Todos) funded by the Ministry of Industry, Energy and Tourism, Spain.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dinesen, B.; Nonnecke, B.; Lindeman, D.; Toft, E.; Kidholm, K.; Jethwani, K.; Young, H.M.; Spindler, H.; Oestergaard, C.U.; Southard, J.A.; et al. Personalized Telehealth in the Future: A Global Research Agenda. J. Med. Intern. Res. 2016, 18, e53. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Kwon, M.S.; Jang, J.H. Qualitative Analysis of Tele-healthcare Systems based on the Diffusion of Innovation Model. J. Korean Acad. Community Health Nurs. 2017, 28, 129–143. [Google Scholar] [CrossRef] [Green Version]
  3. Wac, K. From Quantified Self to Quality of Life. In Digital Health; Health Informatics Series; Springer: Berlin/Heidelberg, Germany, 2018; pp. 83–108. [Google Scholar] [CrossRef]
  4. Konstantas, D.; Bults, R.; Van Halteren, A.; Wac, K.; Jones, V.; Widya, I.; Herzog, R.; Streimelweger, B. Mobile health care: Towards a commercialization of research results. In Proceedings of the European Conference on eHealth 2006, Fribourg, Switzerland, 12–13 October 2006; pp. 155–166. [Google Scholar]
  5. van Halteren, A.; Bults, R.; Wac, K.; Konstantas, D.; Widya, I.; Dokovski, N.; Koprinkov, G.; Jones, V.; Herzog, R. Mobile Patient Monitoring: The Mobihealth System. J. Inf. Technol. Healthc. 2004, 2, 365–373. [Google Scholar]
  6. McLean, S.; Protti, D.; Sheikh, A. Telehealthcare for long term conditions. BMJ 2011, 342. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Arteta, A.; Blas, N.G.; de Mingo López, L.F. Solving complex problems with a bioinspired model. Eng. Appl. AI 2011, 24, 919–927. [Google Scholar] [CrossRef] [Green Version]
  8. Paun, G. Computing with Membranes; Technical Report 208; Turku Centre for Computer Science: Turku, Finland, 1998. [Google Scholar]
  9. Dyer, E.A.; Kansagara, D.; McInnes, D.K.; Freeman, M.; Woods, S. Mobile Applications and Internet-Based Approaches for Supporting Non-Professional Caregivers: A Systematic Review; VA Evidence-Based Synthesis Program Reports; Department of Veterans Affairs: Washington, DC, USA, 2012.
  10. Dozza, M.; Chiari, L.; Chan, B.; Rocchi, L.; Horak, F.B.; Cappello, A. Influence of a portable audio-biofeedback device on structural properties of postural sway. J. Neuroeng. Rehabil. 2005, 2, 13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Franco, C.; Fleury, A.; Gumery, P.Y.; Diot, B.; Demongeot, J.; Vuillerme, N. iBalance-ABF: A Smartphone-Based Audio-Biofeedback Balance System. IEEE Trans. Biomed. Eng. 2013, 60, 211–215. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Demongeot, J.; Elena, A.; Jelassi, M.; Miled, S.B.; Saoud, N.B.B.; Taramasco, C. Smart Homes and Sensors for Surveillance and Preventive Education at Home: Example of Obesity. Information 2016, 7, 50. [Google Scholar] [CrossRef] [Green Version]
  13. Morik, K.; Brockhausen, P.; Joachims, T. Combining statistical learning with a knowledge-based approach—A case study in intensive care monitoring. In Proceedings of the 16th International Conference on Machine Learning, San Francisco, CA, USA, 27–30 June 1999; pp. 268–277. [Google Scholar]
  14. Castleman, K.R. Digital Image Processing; Prentice Hall Press: Upper Saddle River, NJ, USA, 1996. [Google Scholar]
  15. Kulpa, Z. Area and perimeter measurement of blobs in discrete binary pictures. Comput. Graph. Image Process. 1977, 6, 434–451. [Google Scholar] [CrossRef]
  16. Ikeuchi, K. (Ed.) Computer Vision: A Reference Guide; Springer Reference; Springer: New York, NY, USA, 2014. [Google Scholar] [CrossRef]
  17. Kittipanya-Ngam, P.; Ong, S.G.; Eng, H.L. Computer vision applications for patients monitoring system. In Proceedings of the 15th International Conference on Information Fusion, FUSION 2012, Singapore, 9–12 July 2012; pp. 2201–2208. [Google Scholar]
  18. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Huang, H.K. Multimedia Applications In Health Care. IEEE Multimed. 1997, 4, 23. [Google Scholar] [CrossRef]
  20. Skalska, A.; Wizner, B.; Piotrowicz, K.; Klich-Rączka, A.; Klimek, E.; Mossakowska, M.; Rowiński, R.; Kozak-Szkopek, E.; Jóźwiak, A.; Gąsowski, J.; et al. The prevalence of falls and their relation to visual and hearing impairments among a nation-wide cohort of older Poles. Exp. Gerontol. 2013, 48, 140–146. [Google Scholar] [CrossRef] [PubMed]
  21. Yoshida, S. A Global Report on Falls Prevention: Epidemiology of Falls; Technical Report; World Health Organization: Geneva, Switzerland, 2007. [Google Scholar]
  22. Black, A.; Wood, J. Vision and falls. Clin. Exp. Optom. 2005, 88, 212–222. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Taramasco, C.; Rodenas, T.; Martinez, F.; Fuentes, P.; Munoz, R.; Olivares, R.; De Albuquerque, V.H.C.; Demongeot, J. A Novel Monitoring System for Fall Detection in Older People. IEEE Access 2018, 6, 43563–43574. [Google Scholar] [CrossRef]
  24. Castellanos, J.; Bualaa, T.V. Ministry of Industry, Tourism and Commerce; BOE-A-2011-20329; TSI-090302-2011-0007 B80961279 BUALAA TV; Universidad Politécnica de Madrid: Madrid, Spain, 2011. [Google Scholar]
  25. Castellanos, J. Canal de Tod@s; TSI-020605-2012-40 B13420047; BOE-A-2013-1089; Ministry of Industry, Energy and Tourism, Universidad Politécnica de Madrid: Madrid, Spain, 2012.
  26. Zhao, Y.; Belkasim, S. Image Compression and Denoising Algorithm based on Multi-resolution Discrete Cosine Transform. In Proceedings of the International Conference on Image Processing, Computer Vision, and Pattern Recognition (IPCV), Las Vegas, NV, USA, 25–28 July 2016; pp. 110–116. [Google Scholar]
  27. Lee, S.; Zhao, Y.; Masoud, M.E.M.; Valero, M.; Kul, S.; Belkasim, S. Domain Specific Information Retrieval and Text Mining in Medical Document. In Proceedings of the 6th ACM Conference on Bioinformatics, Computational Biology and Health Informatics, Atlanta, GA, USA, 9–12 September 2015; ACM: New York, NY, USA, 2015; pp. 67–76. [Google Scholar] [CrossRef]
  28. Zhao, Y.; Belkasim, S.; Krishnamani, J. Image Compression Scheme Based on Multi-resolution Boundary-Based Shape Descriptors. In Proceedings of the 2013 IEEE/WIC/ACM International Joint Conferences on Web Intelligence (WI) and Intelligent Agent Technologies (IAT)—Volume 02, Atlanta, GA, USA, 17–20 November 2013; IEEE Computer Society: Washington, DC, USA, 2013; pp. 391–398. [Google Scholar] [CrossRef]
  29. Miller, B.; Shrestha, A.; Derby, J.; Olive, J.; Umapathy, K.; Li, F.; Zhao, Y. Digging into human rights violations: Data modelling and collective memory. In Proceedings of the 2013 IEEE International Conference on Big Data, Silicon Valley, CA, USA, 6–9 October 2013; IEEE Computer Society: Los Alamitos, CA, USA, 2013; pp. 37–45. [Google Scholar] [CrossRef]
  30. Fayyad, U.; Piatetsky-Shapiro, G.; Smyth, P. Knowledge Discovery and Data Mining: Towards a Unifying Framework. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, Portland, OR, USA, 2–4 August 1996; pp. 82–88. [Google Scholar]
  31. Krishnamani, J.; Zhao, Y.; Sunderraman, R. Forum Summarization Using Topic Models and Content-Metadata Sensitive Clustering. In Proceedings of the 2013 IEEE/WIC/ACM International Joint Conferences on Web Intelligence (WI) and Intelligent Agent Technologies (IAT)–Volume 03, WI-IAT’13, Atlanta, GA, USA, 17–20 November 2013; IEEE Computer Society: Washington, DC, USA, 2013; pp. 195–198. [Google Scholar] [CrossRef]
  32. He, R.; Zhu, Y.; Xiong, N.; Zhan, W.; Zhao, Y. Mobile Agent-Based Data Gathering for Web Image Retrieval. In Proceedings of the 12th IEEE International Conference on Computational Science and Engineering, CSE 2009, Vancouver, BC, Canada, 29–31 August 2009; pp. 1024–1029. [Google Scholar] [CrossRef]
  33. Zhao, Y.; Belkasim, S. Improving stability and invariance of Cartesian Zernike moments. In Proceedings of the IEEE Southwest Symposium on Image Analysis and Interpretation, Santa Fe, NM, USA, 22–24 April 2012; pp. 61–64. [Google Scholar] [CrossRef]
  34. Zhao, Y.; Belkasim, S. Multiresolution Fourier Descriptors for Multiresolution Shape Analysis. IEEE Signal Process. Lett. 2012, 19, 692–695. [Google Scholar] [CrossRef]
  35. Zhao, Y.; Zhang, Y.; Xiong, N. Biological data classification using rough sets and support vector machines. In Proceedings of the Annual Conference of the North American Fuzzy Information Processing Society—NAFIPS, Cincinnati, OH, USA, 14–17 June 2009. [Google Scholar] [CrossRef]
  36. Arteta, A.; Fernandez, L.; Gil, J. Algorithm for Application of Evolution Rules Based on Linear Diofantic Equations. In Proceedings of the 10th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing, SYNASC 2008, Timisoara, Romania, 26–29 September 2008; pp. 496–500. [Google Scholar] [CrossRef]
  37. Arteta, A.; Mingo, L.F.; Gómez, N. Membrane Systems Working with the P-Factor: Best Strategy to Solve Complex Problems. Adv. Sci. Lett. 2013, 19, 1490–1494. [Google Scholar] [CrossRef]
  38. de Mingo López, L.F.; Blas, N.G.; Albert, A.A. Multidimensional knapsack problem optimization using a binary particle swarm model with genetic operations. Soft Comput. 2018, 22, 2567–2582. [Google Scholar] [CrossRef]
  39. Albert, A.A.; Blas, N.G.; de Mingo López, L.F. Natural combination to trade in the stock market. Soft Comput. 2016, 20, 2433–2450. [Google Scholar] [CrossRef]
  40. de Mingo López, L.F.; Blas, N.G.; Arteta, A. Optimal Performance: Underlying Octahedron Graph of Evolutionary Processors. Comput. Inform. 2015, 34, 858–876. [Google Scholar]
  41. de Mingo López, L.F.; Blas, N.G.; Arteta, A. The optimal combination: Grammatical swarm, particle swarm optimization and neural networks. J. Comput. Sci. 2012, 3, 46–55. [Google Scholar] [CrossRef] [Green Version]
  42. de Frutos, J.A.; Arroyo, F.; Arteta, A. Usefulness States in New P System Communication Architectures. In Proceedings of the Membrane Computing—9th International Workshop, WMC 2008, Edinburgh, UK, 28–31 July 2008; pp. 169–186. [Google Scholar] [CrossRef]
  43. Council, N.R. Basic Research in Information Science and Technology for Air Force Needs; The National Academies Press: Washington, DC, USA, 2006. [Google Scholar] [CrossRef]
  44. Stuhlsatz, A.; Meyer, C.; Eyben, F.; Zielke, T.; Meier, G.; Schuller, B. Deep neural networks for acoustic emotion recognition: Raising the benchmarks. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 5688–5691. [Google Scholar] [CrossRef] [Green Version]
  45. Huttunen, H.; Yancheshmeh, F.S.; Chen, K. Car type recognition with Deep Neural Networks. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), Austin, TX, USA, 10–12 August 2016; pp. 1115–1120. [Google Scholar] [CrossRef] [Green Version]
  46. Yu, M.; Gong, L.; Kollias, S. Computer Vision Based Fall Detection by a Convolutional Neural Network. In Proceedings of the 19th ACM International Conference on Multimodal Interaction, ICMI ’17, Glasgow, Scotland, 13–17 November 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 416–420. [Google Scholar] [CrossRef]
Figure 1. System overview: scenarios, services and components (left to right).
Figure 1. System overview: scenarios, services and components (left to right).
Applsci 10 01142 g001
Figure 2. The architecture of the system. All actions are performed via API Rest calls to server (blue icon).
Figure 2. The architecture of the system. All actions are performed via API Rest calls to server (blue icon).
Applsci 10 01142 g002
Figure 3. Particular case of a given user. Dashboard with information of task to be done by patient (left), and historical data about patient (right).
Figure 3. Particular case of a given user. Dashboard with information of task to be done by patient (left), and historical data about patient (right).
Applsci 10 01142 g003
Figure 4. An example of CNN architecture. The way in which the layers are structured, which elements are used in each layer and how they are designed will often affect the speed and accuracy with which it can perform various tasks.
Figure 4. An example of CNN architecture. The way in which the layers are structured, which elements are used in each layer and how they are designed will often affect the speed and accuracy with which it can perform various tasks.
Applsci 10 01142 g004
Table 1. Experimental results show the proposed CNN classifier achieves a high fall detection accuracy.
Table 1. Experimental results show the proposed CNN classifier achieves a high fall detection accuracy.
ActivityNumberFallsNonfallsSuccess Rate
Falls98742475%
Walk (nonfall)9219175%
Sit (nonfall)863056
Bend (nonfall)1324587

Share and Cite

MDPI and ACS Style

Arteta Albert, A.; de Mingo López, L.F.; Gómez Blas, N. Tele-Treatment Application Design for Disable Patients with Wireless Sensors. Appl. Sci. 2020, 10, 1142. https://doi.org/10.3390/app10031142

AMA Style

Arteta Albert A, de Mingo López LF, Gómez Blas N. Tele-Treatment Application Design for Disable Patients with Wireless Sensors. Applied Sciences. 2020; 10(3):1142. https://doi.org/10.3390/app10031142

Chicago/Turabian Style

Arteta Albert, Alberto, Luis Fernando de Mingo López, and Nuria Gómez Blas. 2020. "Tele-Treatment Application Design for Disable Patients with Wireless Sensors" Applied Sciences 10, no. 3: 1142. https://doi.org/10.3390/app10031142

APA Style

Arteta Albert, A., de Mingo López, L. F., & Gómez Blas, N. (2020). Tele-Treatment Application Design for Disable Patients with Wireless Sensors. Applied Sciences, 10(3), 1142. https://doi.org/10.3390/app10031142

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop