26
|
Rahimzadeh V. A Policy and Practice Review of Consumer Protections and Their Application to Hospital-Sourced Data Aggregation and Analytics by Third-Party Companies. Front Big Data 2021; 3:603044. [PMID: 33693425 PMCID: PMC7931961 DOI: 10.3389/fdata.2020.603044] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2020] [Accepted: 12/21/2020] [Indexed: 11/30/2022] Open
Abstract
The Office of the National Coordinator for Health Information Technology estimates that 96% of all U.S. hospitals use a basic electronic health record, but only 62% are able to exchange health information with outside providers. Barriers to information exchange across EHR systems challenge data aggregation and analysis that hospitals need to evaluate healthcare quality and safety. A growing number of hospital systems are partnering with third-party companies to provide these services. In exchange, companies reserve the rights to sell the aggregated data and analyses produced therefrom, often without the knowledge of patients from whom the data were sourced. Such partnerships fall in a regulatory grey area and raise new ethical questions about whether health, consumer, or health and consumer privacy protections apply. The current opinion probes this question in the context of consumer privacy reform in California. It analyzes protections for health information recently expanded under the California Consumer Privacy Act ("CA Privacy Act") in 2020 and compares them to protections outlined in the Health Information Portability and Accountability Act ("Federal Privacy Rule"). Four perspectives are considered in this ethical analysis: 1) standards of data deidentification; 2) rights of patients and consumers in relation to their health information; 3) entities covered by the CA Privacy Act; 4) scope and complementarity of federal and state regulations. The opinion concludes that the CCPA is limited in its application when health information is processed by a third-party data aggregation company that is contractually designated as a business associate; when health information is deidentified; and when hospital data are sourced from publicly owned and operated hospitals. Lastly, the opinion offers practical recommendations for facilitating parity between state and federal health data privacy laws and for how a more equitable distribution of informational risks and benefits from the sale of aggregated hospital data could be fostered and presents ways both for-profit and nonprofit hospitals can sustain patient trust when negotiating partnerships with third-party data aggregation companies.
Collapse
|
27
|
Decentralized Privacy-Preserving Data Aggregation Scheme for Smart Grid Based on Blockchain. SENSORS 2020; 20:s20185282. [PMID: 32942782 PMCID: PMC7570635 DOI: 10.3390/s20185282] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/13/2020] [Revised: 09/03/2020] [Accepted: 09/10/2020] [Indexed: 11/30/2022]
Abstract
As a next-generation power system, the smart grid can implement fine-grained smart metering data collection to optimize energy utilization. Smart meters face serious security challenges, such as a trusted third party or a trusted authority being attacked, which leads to the disclosure of user privacy. Blockchain provides a viable solution that can use its key technologies to solve this problem. Blockchain is a new type of decentralized protocol that does not require a trusted third party or a central authority. Therefore, this paper proposes a decentralized privacy-preserving data aggregation (DPPDA) scheme for smart grid based on blockchain. In this scheme, the leader election algorithm is used to select a smart meter in the residential area as a mining node to build a block. The node adopts Paillier cryptosystem algorithm to aggregate the user’s power consumption data. Boneh-Lynn-Shacham short signature and SHA-256 function are applied to ensure the confidentiality and integrity of user data, which is convenient for billing and power regulation. The scheme protects user privacy data while achieving decentralization, without relying on TTP or CA. Security analysis shows that our scheme meets the security and privacy requirements of smart grid data aggregation. The experimental results show that this scheme is more efficient than existing competing schemes in terms of computation and communication overhead.
Collapse
|
28
|
Gagalova KK, Leon Elizalde MA, Portales-Casamar E, Görges M. What You Need to Know Before Implementing a Clinical Research Data Warehouse: Comparative Review of Integrated Data Repositories in Health Care Institutions. JMIR Form Res 2020; 4:e17687. [PMID: 32852280 PMCID: PMC7484778 DOI: 10.2196/17687] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2020] [Revised: 06/09/2020] [Accepted: 07/17/2020] [Indexed: 12/23/2022] Open
Abstract
Background Integrated data repositories (IDRs), also referred to as clinical data warehouses, are platforms used for the integration of several data sources through specialized analytical tools that facilitate data processing and analysis. IDRs offer several opportunities for clinical data reuse, and the number of institutions implementing an IDR has grown steadily in the past decade. Objective The architectural choices of major IDRs are highly diverse and determining their differences can be overwhelming. This review aims to explore the underlying models and common features of IDRs, provide a high-level overview for those entering the field, and propose a set of guiding principles for small- to medium-sized health institutions embarking on IDR implementation. Methods We reviewed manuscripts published in peer-reviewed scientific literature between 2008 and 2020, and selected those that specifically describe IDR architectures. Of 255 shortlisted articles, we found 34 articles describing 29 different architectures. The different IDRs were analyzed for common features and classified according to their data processing and integration solution choices. Results Despite common trends in the selection of standard terminologies and data models, the IDRs examined showed heterogeneity in the underlying architecture design. We identified 4 common architecture models that use different approaches for data processing and integration. These different approaches were driven by a variety of features such as data sources, whether the IDR was for a single institution or a collaborative project, the intended primary data user, and purpose (research-only or including clinical or operational decision making). Conclusions IDR implementations are diverse and complex undertakings, which benefit from being preceded by an evaluation of requirements and definition of scope in the early planning stage. Factors such as data source diversity and intended users of the IDR influence data flow and synchronization, both of which are crucial factors in IDR architecture planning.
Collapse
|
29
|
Effects of Data Aggregation on Time Series Analysis of Seasonal Infections. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2020; 17:ijerph17165887. [PMID: 32823719 PMCID: PMC7460497 DOI: 10.3390/ijerph17165887] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/30/2020] [Revised: 08/04/2020] [Accepted: 08/05/2020] [Indexed: 01/03/2023]
Abstract
Time series analysis in epidemiological studies is typically conducted on aggregated counts, although data tend to be collected at finer temporal resolutions. The decision to aggregate data is rarely discussed in epidemiological literature although it has been shown to impact model results. We present a critical thinking process for making decisions about data aggregation in time series analysis of seasonal infections. We systematically build a harmonic regression model to characterize peak timing and amplitude of three respiratory and enteric infections that have different seasonal patterns and incidence. We show that irregularities introduced when aggregating data must be controlled during modeling to prevent erroneous results. Aggregation irregularities had a minimal impact on the estimates of trend, amplitude, and peak timing for daily and weekly data regardless of the disease. However, estimates of peak timing of the more common infections changed by as much as 2.5 months when controlling for monthly data irregularities. Building a systematic model that controls for data irregularities is essential to accurately characterize temporal patterns of infections. With the urgent need to characterize temporal patterns of novel infections, such as COVID-19, this tutorial is timely and highly valuable for experts in many disciplines.
Collapse
|
30
|
Kenyeres M, Kenyeres J. Average Consensus over Mobile Wireless Sensor Networks: Weight Matrix Guaranteeing Convergence without Reconfiguration of Edge Weights. SENSORS (BASEL, SWITZERLAND) 2020; 20:s20133677. [PMID: 32630083 PMCID: PMC7374501 DOI: 10.3390/s20133677] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/19/2020] [Revised: 06/17/2020] [Accepted: 06/27/2020] [Indexed: 06/11/2023]
Abstract
Efficient data aggregation is crucial for mobile wireless sensor networks, as their resources are significantly constrained. Over recent years, the average consensus algorithm has found a wide application in this technology. In this paper, we present a weight matrix simplifying the average consensus algorithm over mobile wireless sensor networks, thereby prolonging the network lifetime as well as ensuring the proper operation of the algorithm. Our contribution results from the theorem stating how the Laplacian spectrum of an undirected simple finite graph changes in the case of adding an arbitrary edge into this graph. We identify that the mixing parameter of Best Constant weights of a complete finite graph with an arbitrary order ensures the convergence in time-varying topologies without any reconfiguration of the edge weights. The presented theorems and lemmas are verified over evolving graphs with various parameters, whereby it is demonstrated that our approach ensures the convergence of the average consensus algorithm over mobile wireless sensor networks in spite of no edge reconfiguration.
Collapse
|
31
|
Boudon S, Henry-Berger J, Cassar-Malek I. Aggregation of Omic Data and Secretome Prediction Enable the Discovery of Candidate Plasma Biomarkers for Beef Tenderness. Int J Mol Sci 2020; 21:E664. [PMID: 31963926 PMCID: PMC7013622 DOI: 10.3390/ijms21020664] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2019] [Revised: 01/15/2020] [Accepted: 01/16/2020] [Indexed: 12/18/2022] Open
Abstract
Beef quality is a complex phenotype that can be evaluated only after animal slaughtering. Previous research has investigated the potential of genetic markers or muscle-derived proteins to assess beef tenderness. Thus, the use of low-invasive biomarkers in living animals is an issue for the beef sector. We hypothesized that publicly available data may help us discovering candidate plasma biomarkers. Thanks to a review of the literature, we built a corpus of articles on beef tenderness. Following data collection, aggregation, and computational reconstruction of the muscle secretome, the putative plasma proteins were searched by comparison with a bovine plasma proteome atlas and submitted to mining of biological information. Of the 44 publications included in the study, 469 unique gene names were extracted for aggregation. Seventy-one proteins putatively released in the plasma were revealed. Among them 13 proteins were predicted to be secreted in plasma, 44 proteins as hypothetically secreted in plasma, and 14 additional candidate proteins were detected thanks to network analysis. Among these 71 proteins, 24 were included in tenderness quantitative trait loci. The in-silico workflow enabled the discovery of candidate plasma biomarkers for beef tenderness from reconstruction of the secretome, to be examined in the cattle plasma proteome.
Collapse
|
32
|
Nassar A, Torres-Rua A, Kustas W, Nieto H, McKee M, Hipps L, Stevens D, Alfieri J, Prueger J, Alsina MM, McKee L, Coopmans C, Sanchez L, Dokoozlian N. Influence of Model Grid Size on the Estimation of Surface Fluxes Using the Two Source Energy Balance Model and sUAS Imagery in Vineyards. REMOTE SENSING 2020; 12:342. [PMID: 32355571 PMCID: PMC7192008 DOI: 10.3390/rs12030342] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Evapotranspiration (ET) is a key variable for hydrology and irrigation water management, with significant importance in drought-stricken regions of the western US. This is particularly true for California, which grows much of the high-value perennial crops in the US. The advent of small Unmanned Aerial System (sUAS) with sensor technology similar to satellite platforms allows for the estimation of high-resolution ET at plant spacing scale for individual fields. However, while multiple efforts have been made to estimate ET from sUAS products, the sensitivity of ET models to different model grid size/resolution in complex canopies, such as vineyards, is still unknown. The variability of row spacing, canopy structure, and distance between fields makes this information necessary because additional complexity processing individual fields. Therefore, processing the entire image at a fixed resolution that is potentially larger than the plant-row separation is more efficient. From a computational perspective, there would be an advantage to running models at much coarser resolutions than the very fine native pixel size from sUAS imagery for operational applications. In this study, the Two-Source Energy Balance with a dual temperature (TSEB2T) model, which uses remotely sensed soil/substrate and canopy temperature from sUAS imagery, was used to estimate ET and identify the impact of spatial domain scale under different vine phenological conditions. The analysis relies upon high-resolution imagery collected during multiple years and times by the Utah State University AggieAir™ sUAS program over a commercial vineyard located near Lodi, California. This project is part of the USDA-Agricultural Research Service Grape Remote Sensing Atmospheric Profile and Evapotranspiration eXperiment (GRAPEX). Original spectral and thermal imagery data from sUAS were at 10 cm and 60 cm per pixel, respectively, and multiple spatial domain scales (3.6, 7.2, 14.4, and 30 m) were evaluated and compared against eddy covariance (EC) measurements. Results indicated that the TSEB2T model is only slightly affected in the estimation of the net radiation (R n ) and the soil heat flux (G) at different spatial resolutions, while the sensible and latent heat fluxes (H and LE, respectively) are significantly affected by coarse grid sizes. The results indicated overestimation of H and underestimation of LE values, particularly at Landsat scale (30 m). This refers to the non-linear relationship between the land surface temperature (LST) and the normalized difference vegetation index (NDVI) at coarse model resolution. Another predominant reason for LE reduction in TSEB2T was the decrease in the aerodynamic resistance (R a ), which is a function of the friction velocity F*) that varies with mean canopy height and roughness length. While a small increase in grid size can be implemented, this increase should be limited to less than twice the smallest row spacing present in the sUAS imagery. The results also indicated that the mean LE at field scale is reduced by 10% to 20% at coarser resolutions, while the with-in field variability in LE values decreased significantly at the larger grid sizes and ranged between approximately 15% and 45%. This implies that, while the field-scale values of LE are fairly reliable at larger grid sizes, the with-in field variability limits its use for precision agriculture applications.
Collapse
|
33
|
Sennan S, Balasubramaniyam S, Luhach AK, Ramasubbareddy S, Chilamkurti N, Nam Y. Energy and Delay Aware Data Aggregation in Routing Protocol for Internet of Things. SENSORS 2019; 19:s19245486. [PMID: 31842437 PMCID: PMC6961041 DOI: 10.3390/s19245486] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/18/2019] [Revised: 12/03/2019] [Accepted: 12/05/2019] [Indexed: 11/16/2022]
Abstract
Energy conservation is one of the most critical problems in Internet of Things (IoT). It can be achieved in several ways, one of which is to select the optimal route for data transfer. IPv6 Routing Protocol for Low Power and Lossy Networks (RPL) is a standardized routing protocol for IoT. The RPL changes its path frequently while transmitting the data from source to the destination, due to high data traffic in dense networks. Hence, it creates data traffic across the nodes in the networks. To solve this issue, we propose Energy and Delay Aware Data aggregation in Routing Protocol (EDADA-RPL) for IoT. It has two processes, namely parent selection and data aggregation. The process of parent selection uses routing metric residual energy (RER) to choose the best possible parent for data transmission. The data aggregation process uses the compressed sensing (CS) theory in the parent node to combine data packets from the child nodes. Finally, the aggregated data transmits from a downward parent to the sink. The sink node collects all the aggregated data and it performs the reconstruction operation to get the original data of the participant node. The simulation is carried out using the Contiki COOJA simulator. EDADA-RPL’s performance is compared to RPL and LA-RPL. The EDADA-RPL offers good performance in terms of network lifetime, delay, and packet delivery ratio.
Collapse
|
34
|
Stöhr MR, Günther A, Majeed RW. Verifying Data Integration Configurations for Semantical Correctness and Completeness. Stud Health Technol Inform 2019; 267:66-73. [PMID: 31483256 DOI: 10.3233/shti190807] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Data integration is the problem of combining data residing at different sources and providing the user with a unified view of these data. In medical informatics, such a unified view enables retrospective analyses based on more facts and prospective recruitment of more patients than any single data collection by itself. The technical part of data integration is based on rules interpreted by software. These rules define how to perform the translation of source database schemata into the target database schema. Translation rules are formulated by data managers who usually do not have the knowledge about meaning and acquisition methods of the data they handle. The professionals (data providers) collecting the source data who have the respective knowledge again usually have no sufficient technical background. Since data providers are neither able to formulate the transformation rules themselves nor able to validate them, the whole process is fault-prone. Additionally, in continuous development and maintenance of (meta-) data repositories, data structures underlie changes, which may lead to outdated transformation rules. We did not find any technical solution, which enables data providers to formulate transformation rules themselves or which provides an understandable reflection of given rules. Our approach is to enable data providers understand the rules regarding their own data by presenting rules and available context visually. Context information is fetched from a metadata repository. In this paper, we propose a software tool that builds on existing data integration infrastructures. The tool provides a visually supported validation routine for data integration rules. In a first step towards its evaluation, we implement the tool into the DZL data integration process and verify the correct presentation of transformation rules.
Collapse
|
35
|
Coenen VA, Schlaepfer TE, Varkuti B, Schuurman PR, Reinacher PC, Voges J, Zrinzo L, Blomstedt P, Fenoy AJ, Hariz M. Surgical decision making for deep brain stimulation should not be based on aggregated normative data mining. Brain Stimul 2019; 12:1345-1348. [PMID: 31353286 DOI: 10.1016/j.brs.2019.07.014] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2019] [Revised: 07/02/2019] [Accepted: 07/18/2019] [Indexed: 11/25/2022] Open
|
36
|
Scher E, Cohen SB, Sanguinetti G. PartCrafter: find, generate and analyze BioParts. Synth Biol (Oxf) 2019; 4:ysz014. [PMID: 32995539 PMCID: PMC7445878 DOI: 10.1093/synbio/ysz014] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2018] [Revised: 05/14/2019] [Accepted: 05/22/2019] [Indexed: 06/11/2023] Open
Abstract
The field of Synthetic Biology is both practically and philosophically reliant on the idea of BioParts-concrete DNA sequences meant to represent discrete functionalities. While there are a number of software tools which allow users to design complex DNA sequences by stitching together BioParts or genetic features into genetic devices, there is a lack of tools assisting Synthetic Biologists in finding BioParts and in generating new ones. In practice, researchers often find BioParts in an ad hoc way. We present PartCrafter, a tool which extracts and aggregates genomic feature data in order to facilitate the search for new BioParts with specific functionalities. PartCrafter can also turn a genomic feature into a BioPart by packaging it according to any manufacturing standard, codon optimizing it for a new host, and removing forbidden sites. PartCrafter is available at partcrafter.com.
Collapse
|
37
|
Dankar FK, Madathil N, Dankar SK, Boughorbel S. Privacy-Preserving Analysis of Distributed Biomedical Data: Designing Efficient and Secure Multiparty Computations Using Distributed Statistical Learning Theory. JMIR Med Inform 2019; 7:e12702. [PMID: 31033449 PMCID: PMC6658266 DOI: 10.2196/12702] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2018] [Revised: 01/14/2019] [Accepted: 02/15/2019] [Indexed: 02/01/2023] Open
Abstract
Background Biomedical research often requires large cohorts and necessitates the sharing of biomedical data with researchers around the world, which raises many privacy, ethical, and legal concerns. In the face of these concerns, privacy experts are trying to explore approaches to analyzing the distributed data while protecting its privacy. Many of these approaches are based on secure multiparty computations (SMCs). SMC is an attractive approach allowing multiple parties to collectively carry out calculations on their datasets without having to reveal their own raw data; however, it incurs heavy computation time and requires extensive communication between the involved parties. Objective This study aimed to develop usable and efficient SMC applications that meet the needs of the potential end-users and to raise general awareness about SMC as a tool that supports data sharing. Methods We have introduced distributed statistical computing (DSC) into the design of secure multiparty protocols, which allows us to conduct computations on each of the parties’ sites independently and then combine these computations to form 1 estimator for the collective dataset, thus limiting communication to the final step and reducing complexity. The effectiveness of our privacy-preserving model is demonstrated through a linear regression application. Results Our secure linear regression algorithm was tested for accuracy and performance using real and synthetic datasets. The results showed no loss of accuracy (over nonsecure regression) and very good performance (20 min for 100 million records). Conclusions We used DSC to securely calculate a linear regression model over multiple datasets. Our experiments showed very good performance (in terms of the number of records it can handle). We plan to extend our method to other estimators such as logistic regression.
Collapse
|
38
|
A Secure Mutual Batch Authentication Scheme for Patient Data Privacy Preserving in WBAN. SENSORS 2019; 19:s19071608. [PMID: 30987177 PMCID: PMC6480194 DOI: 10.3390/s19071608] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/01/2019] [Revised: 03/18/2019] [Accepted: 03/26/2019] [Indexed: 11/16/2022]
Abstract
The current advances in cloud-based services have significantly enhanced individual satisfaction in numerous modern life areas. Particularly, the recent spectacular innovations in the wireless body area networks (WBAN) domain have made e-Care services rise as a promising application field, which definitely improves the quality of the medical system. However, the forwarded data from the limited connectivity range of WBAN via a smart device (e.g., smartphone) to the application provider (AP) should be secured from an unapproved access and alteration (attacker) that could prompt catastrophic consequences. Therefore, several schemes have been proposed to guarantee data integrity and privacy during their transmission between the client/controller (C) and the AP. Thereby, numerous effective cryptosystem solutions based on a bilinear pairing approach are available in the literature to address the mentioned security issues. Unfortunately, the related solution presents security shortcomings, where AP can with ease impersonate a given C. Hence, this existing scheme cannot fully guarantee C's data privacy and integrity. Therefore, we propose our contribution to address this data security issue (impersonation) through a secured and efficient remote batch authentication scheme that genuinely ascertains the identity of C and AP. Practically, the proposed cryptosystem is based on an efficient combination of elliptical curve cryptography (ECC) and bilinear pairing schemes. Furthermore, our proposed solution reduces the communication and computational costs by providing an efficient data aggregation and batch authentication for limited device's resources in WBAN. These additional features (data aggregation and batch authentication) are the core improvements of our scheme that have great merit for limited energy environments like WBAN.
Collapse
|
39
|
Rippinger C, Weibrecht N, Zechmeister M, Scheffel S, Urach C, Endel F. Health Care Atlases: Informing the General Public About the Situation of the Austrian Health Care System. Stud Health Technol Inform 2019; 260:49-56. [PMID: 31118318] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
There is a great number of complex data concerning the Austrian Health Care System. The goal was to process this data and present it to the general public on an easily accessible information platform. The platform focuses on data about the burden of disease of the Austrian Population, the available medical care and the services provided by the physicians. Due to the vast differences in the underlying source data, the methods used for the data acquisition range from statistical linkage over web scraping to aggregating data on the reimbursed services. The results are published on a website and are mainly displayed with interactive graphics. Overall, these dynamic and interactive websites provide a good overview of the situation of the Austrian Health Care System and presents the information in an intuitive and comprehensible manner. Furthermore, the information given in the atlases can contribute to the health care planning in order to identify distinctive service provision in Austria.
Collapse
|
40
|
CS-FCDA: A Compressed Sensing-Based on Fault-Tolerant Data Aggregation in Sensor Networks. SENSORS 2018; 18:s18113749. [PMID: 30400248 PMCID: PMC6263422 DOI: 10.3390/s18113749] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/03/2018] [Revised: 10/26/2018] [Accepted: 11/01/2018] [Indexed: 11/27/2022]
Abstract
When the nodes in the network are deployed in the target area with an appropriate density, the effective aggregation and transmission of the data gathered in the monitoring area remain to be solved. The existing Compressed Sensing (CS) based on data aggregation schemes are accomplished in a centralized manner and the Sink node achieves the task of data aggregation. However, these existing schemes may suffer from load imbalance and coverage void issues. In order to address these problems, we propose a Compressed Sensing based on Fault-tolerant Correcting Data Aggregation (CS-FCDA) scheme to accurately reconstruct the compressed data. Therefore, the network communication overhead can be greatly reduced while maintaining the quality of the reconstructed data. Meanwhile, we adopt the node clustering mechanism to optimize and balance the network load. It is shown via simulation results, compared with other data aggregation schemes, that the proposed scheme shows obvious improvement in terms of the Fault-tolerant correcting capability and the network energy efficiency of the data reconstruction.
Collapse
|
41
|
Zguira Y, Rivano H, Meddeb A. Internet of Bikes: A DTN Protocol with Data Aggregation for Urban Data Collection. SENSORS 2018; 18:s18092819. [PMID: 30150525 PMCID: PMC6163721 DOI: 10.3390/s18092819] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/17/2018] [Revised: 08/20/2018] [Accepted: 08/21/2018] [Indexed: 12/03/2022]
Abstract
Intelligent Transport Systems (ITS) are an essential part of the global world. They play a substantial role for facing many issues such as traffic jams, high accident rates, unhealthy lifestyles, air pollution, etc. Public bike sharing system is one part of ITS and can be used to collect data from mobiles devices. In this paper, we propose an efficient, “Internet of Bikes”, IoB-DTN routing protocol based on data aggregation which applies the Delay Tolerant Network (DTN) paradigm to Internet of Things (IoT) applications running data collection on urban bike sharing system based sensor network. We propose and evaluate three variants of IoB-DTN: IoB based on spatial aggregation (IoB-SA), IoB based on temporal aggregation (IoB-TA) and IoB based on spatiotemporal aggregation (IoB-STA). The simulation results show that the three variants offer the best performances regarding several metrics, comparing to IoB-DTN without aggregation and the low-power long-range technology, LoRa type. In an urban application, the choice of the type of which variant of IoB should be used depends on the sensed values.
Collapse
|
42
|
Privacy-Preserving Data Aggregation against False Data Injection Attacks in Fog Computing. SENSORS 2018; 18:s18082659. [PMID: 30104516 PMCID: PMC6111540 DOI: 10.3390/s18082659] [Citation(s) in RCA: 29] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/21/2018] [Revised: 07/26/2018] [Accepted: 08/07/2018] [Indexed: 11/16/2022]
Abstract
As an extension of cloud computing, fog computing has received more attention in recent years. It can solve problems such as high latency, lack of support for mobility and location awareness in cloud computing. In the Internet of Things (IoT), a series of IoT devices can be connected to the fog nodes that assist a cloud service center to store and process a part of data in advance. Not only can it reduce the pressure of processing data, but also improve the real-time and service quality. However, data processing at fog nodes suffers from many challenging issues, such as false data injection attacks, data modification attacks, and IoT devices' privacy violation. In this paper, based on the Paillier homomorphic encryption scheme, we use blinding factors to design a privacy-preserving data aggregation scheme in fog computing. No matter whether the fog node and the cloud control center are honest or not, the proposed scheme ensures that the injection data is from legal IoT devices and is not modified and leaked. The proposed scheme also has fault tolerance, which means that the collection of data from other devices will not be affected even if certain fog devices fail to work. In addition, security analysis and performance evaluation indicate the proposed scheme is secure and efficient.
Collapse
|
43
|
Adaptive Aggregation Routing to Reduce Delay for Multi-Layer Wireless Sensor Networks. SENSORS 2018; 18:s18041216. [PMID: 29659535 PMCID: PMC5948754 DOI: 10.3390/s18041216] [Citation(s) in RCA: 26] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/30/2018] [Revised: 04/12/2018] [Accepted: 04/12/2018] [Indexed: 11/17/2022]
Abstract
The quality of service (QoS) regarding delay, lifetime and reliability is the key to the application of wireless sensor networks (WSNs). Data aggregation is a method to effectively reduce the data transmission volume and improve the lifetime of a network. In the previous study, a common strategy required that data wait in the queue. When the length of the queue is greater than or equal to the predetermined aggregation threshold (Nt) or the waiting time is equal to the aggregation timer (Tt), data are forwarded at the expense of an increase in the delay. The primary contributions of the proposed Adaptive Aggregation Routing (AAR) scheme are the following: (a) the senders select the forwarding node dynamically according to the length of the data queue, which effectively reduces the delay. In the AAR scheme, the senders send data to the nodes with a long data queue. The advantages are that first, the nodes with a long data queue need a small amount of data to perform aggregation; therefore, the transmitted data can be fully utilized to make these nodes aggregate. Second, this scheme balances the aggregating and data sending load; thus, the lifetime increases. (b) An improved AAR scheme is proposed to improve the QoS. The aggregation deadline (Tt) and the aggregation threshold (Nt) are dynamically changed in the network. In WSNs, nodes far from the sink have residual energy because these nodes transmit less data than the other nodes. In the improved AAR scheme, the nodes far from the sink have a small value of Tt and Nt to reduce delay, and the nodes near the sink are set to a large value of Tt and Nt to reduce energy consumption. Thus, the end to end delay is reduced, a longer lifetime is achieved, and the residual energy is fully used. Simulation results demonstrate that compared with the previous scheme, the performance of the AAR scheme is improved. This scheme reduces the delay by 14.91%, improves the lifetime by 30.91%, and increases energy efficiency by 76.40%.
Collapse
|
44
|
Multi-Sensor Data Integration Using Deep Learning for Characterization of Defects in Steel Elements. SENSORS 2018; 18:s18010292. [PMID: 29351215 PMCID: PMC5795376 DOI: 10.3390/s18010292] [Citation(s) in RCA: 60] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/09/2017] [Revised: 01/16/2018] [Accepted: 01/16/2018] [Indexed: 12/25/2022]
Abstract
Nowadays, there is a strong demand for inspection systems integrating both high sensitivity under various testing conditions and advanced processing allowing automatic identification of the examined object state and detection of threats. This paper presents the possibility of utilization of a magnetic multi-sensor matrix transducer for characterization of defected areas in steel elements and a deep learning based algorithm for integration of data and final identification of the object state. The transducer allows sensing of a magnetic vector in a single location in different directions. Thus, it enables detecting and characterizing any material changes that affect magnetic properties regardless of their orientation in reference to the scanning direction. To assess the general application capability of the system, steel elements with rectangular-shaped artificial defects were used. First, a database was constructed considering numerical and measurements results. A finite element method was used to run a simulation process and provide transducer signal patterns for different defect arrangements. Next, the algorithm integrating responses of the transducer collected in a single position was applied, and a convolutional neural network was used for implementation of the material state evaluation model. Then, validation of the obtained model was carried out. In this paper, the procedure for updating the evaluated local state, referring to the neighboring area results, is presented. Finally, the results and future perspective are discussed.
Collapse
|
45
|
Friedman CP, Rubin JC, Sullivan KJ. Toward an Information Infrastructure for Global Health Improvement. Yearb Med Inform 2017; 26:16-23. [PMID: 28480469 PMCID: PMC6239237 DOI: 10.15265/iy-2017-004] [Citation(s) in RCA: 84] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2017] [Indexed: 11/24/2022] Open
Abstract
Profound global challenges to individual and population health, alongside the opportunities to benefit from digital technology, have spawned the concept of the Learning Health System. Learning Health Systems (LHSs)--which can function at organizational, network, regional, and national levels of scale--have the capability of continuous data-driven self-study that promotes change and improvement. The LHS concept, which originated in the U.S. in 2007, is rapidly gaining attention around the world. LHSs require, but also transcend, the secondary use of health data. This paper describes the key features of LHSs, argues that effective and sustainable LHSs must be supported by infrastructures that allow them to function with economies of scale and scope, and describes the services that such infrastructures must provide. While it is relatively straightforward to describe LHSs, achieving them at the high level of capability necessary to promote significant health benefits will require advancements in science and engineering, engaging the field of informatics among a wider range of disciplines. It also follows from this vision that LHSs cannot be built from an imposed blueprint; LHSs will more likely evolve from efforts at smaller scales that compose into larger systems.
Collapse
|
46
|
Wu YC, Wei NC, Hung JJ, Yeh YC, Su LJ, Hsu WH, Chou TY. Generating a robust prediction model for stage I lung adenocarcinoma recurrence after surgical resection. Oncotarget 2017; 8:79712-79721. [PMID: 29108351 PMCID: PMC5668084 DOI: 10.18632/oncotarget.19161] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2017] [Accepted: 06/28/2017] [Indexed: 01/11/2023] Open
Abstract
Lung cancer mortality remains high even after successful resection. Adjuvant treatment benefits stage II and III patients, but not stage I patients, and most studies fail to predict recurrence in stage I patients. Our study included 211 lung adenocarcinoma patients (stages I–IIIA; 81% stage I) who received curative resections at Taipei Veterans General Hospital between January 2001 and December 2012. We generated a prediction model using 153 samples, with validation using an additional 58 clinical outcome-blinded samples. Gene expression profiles were generated using formalin-fixed, paraffin-embedded tissue samples and microarrays. Data analysis was performed using a supervised clustering method. The prediction model generated from mixed stage samples successfully separated patients at high vs. low risk for recurrence. The validation tests hazard ratio (HR = 4.38) was similar to that of the training tests (HR = 4.53), indicating a robust training process. Our prediction model successfully distinguished high- from low-risk stage IA and IB patients, with a difference in 5-year disease-free survival between high- and low-risk patients of 42% for stage IA and 45% for stage IB (p < 0.05). We present a novel and effective model for identifying lung adenocarcinoma patients at high risk for recurrence who may benefit from adjuvant therapy. Our prediction performance of the difference in disease free survival between high risk and low risk groups demonstrates more than two fold improvement over earlier published results.
Collapse
|
47
|
Fortune N, Hardiker NR, Strudwick G. Embedding Nursing Interventions into the World Health Organization's International Classification of Health Interventions (ICHI). J Am Med Inform Assoc 2017; 24:722-728. [PMID: 28339684 PMCID: PMC7651898 DOI: 10.1093/jamia/ocw173] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2016] [Revised: 11/11/2016] [Accepted: 11/21/2016] [Indexed: 11/13/2022] Open
Abstract
OBJECTIVE The International Classification of Health Interventions, currently being developed, seeks to span all sectors of the health system. Our objective was to test the draft classification's coverage of interventions commonly delivered by nurses, and propose changes to improve the utility and reliability of the classification for aggregating and analyzing data on nursing interventions. MATERIALS AND METHODS A 2-phase content mapping method was used: (1) three coders independently applied the classification to a dataset comprising 100 high-frequency nursing interventions; (2) the coders reached consensus for each intervention and identified reasons for initial discrepancies. RESULTS A consensus code was found for 80 of the 100 source terms; for 34% of these, the code was semantically equivalent to the source term, and for 64% it was broader. Issues that contributed to discrepancies in Phase 1 coding results included concepts in source terms not captured by the classification, ambiguities in source terms, and uncertainty of semantic matching between "action" concepts in source terms and classification codes. DISCUSSION While the classification generally provides good coverage of nursing interventions, there remain a number of content gaps and granularity issues. Further development of definitions and coding guidance is needed to ensure consistency of application. CONCLUSION This study has produced a set of proposals concerning changes needed to improve the classification. The novel method described here will inform future health terminology and classification content coverage studies.
Collapse
|
48
|
Data Aggregation Based on Overlapping Rate of Sensing Area in Wireless Sensor Networks. SENSORS 2017; 17:s17071527. [PMID: 28661418 PMCID: PMC5539569 DOI: 10.3390/s17071527] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/26/2017] [Revised: 06/09/2017] [Accepted: 06/20/2017] [Indexed: 12/02/2022]
Abstract
Wireless sensor networks are required in smart applications to provide accurate control, where the high density of sensors brings in a large quantity of redundant data. In order to reduce the waste of limited network resources, data aggregation is utilized to avoid redundancy forwarding. However, most of aggregation schemes reduce information accuracy and prolong end-to-end delay when eliminating transmission overhead. In this paper, we propose a data aggregation scheme based on overlapping rate of sensing area, namely AggOR, aiming for energy-efficient data collection in wireless sensor networks with high information accuracy. According to aggregation rules, gathering nodes are selected from candidate parent nodes and appropriate neighbor nodes considering a preset threshold of overlapping rate of sensing area. Therefore, the collected data in a gathering area are highly correlated, and a large amount of redundant data could be cleaned. Meanwhile, AggOR keeps the original entropy by only deleting the duplicated data. Experiment results show that compared with others, AggOR has a high data accuracy and a short end-to-end delay with a similar network lifetime.
Collapse
|
49
|
O'Neill TA. An Overview of Interrater Agreement on Likert Scales for Researchers and Practitioners. Front Psychol 2017; 8:777. [PMID: 28553257 PMCID: PMC5427087 DOI: 10.3389/fpsyg.2017.00777] [Citation(s) in RCA: 39] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2017] [Accepted: 04/26/2017] [Indexed: 11/15/2022] Open
Abstract
Applications of interrater agreement (IRA) statistics for Likert scales are plentiful in research and practice. IRA may be implicated in job analysis, performance appraisal, panel interviews, and any other approach to gathering systematic observations. Any rating system involving subject-matter experts can also benefit from IRA as a measure of consensus. Further, IRA is fundamental to aggregation in multilevel research, which is becoming increasingly common in order to address nesting. Although, several technical descriptions of a few specific IRA statistics exist, this paper aims to provide a tractable orientation to common IRA indices to support application. The introductory overview is written with the intent of facilitating contrasts among IRA statistics by critically reviewing equations, interpretations, strengths, and weaknesses. Statistics considered include rwg, rwg*, r′wg, rwg(p), average deviation (AD), awg, standard deviation (Swg), and the coefficient of variation (CVwg). Equations support quick calculation and contrasting of different agreement indices. The article also includes a “quick reference” table and three figures in order to help readers identify how IRA statistics differ and how interpretations of IRA will depend strongly on the statistic employed. A brief consideration of recommended practices involving statistical and practical cutoff standards is presented, and conclusions are offered in light of the current literature.
Collapse
|
50
|
Yeoum S, Kang B, Lee J, Choo H. Channel and Timeslot Co-Scheduling with Minimal Channel Switching for Data Aggregation in MWSNs. SENSORS (BASEL, SWITZERLAND) 2017; 17:s17051030. [PMID: 28471416 PMCID: PMC5469635 DOI: 10.3390/s17051030] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/08/2017] [Revised: 04/26/2017] [Accepted: 05/01/2017] [Indexed: 06/07/2023]
Abstract
Collision-free transmission and efficient data transfer between nodes can be achieved through a set of channels in multichannel wireless sensor networks (MWSNs). While using multiple channels, we have to carefully consider channel interference, channel and time slot (resources) optimization, channel switching delay, and energy consumption. Since sensor nodes operate on low battery power, the energy consumed in channel switching becomes an important challenge. In this paper, we propose channel and time slot scheduling for minimal channel switching in MWSNs, while achieving efficient and collision-free transmission between nodes. The proposed scheme constructs a duty-cycled tree while reducing the amount of channel switching. As a next step, collision-free time slots are assigned to every node based on the minimal data collection delay. The experimental results demonstrate that the validity of our scheme reduces the amount of channel switching by 17.5%, reduces energy consumption for channel switching by 28%, and reduces the schedule length by 46%, as compared to the existing schemes.
Collapse
|