Probabilistic Data Allocation in Pervasive Computing Applications

被引:0
作者
Kolomvatsos, Kostas [1 ]
机构
[1] Univ Thessaly, Dept Informat & Telecommun, Papasiopoulou 2-4, Lamia 35131, Greece
来源
2020 IEEE INTL SYMP ON PARALLEL & DISTRIBUTED PROCESSING WITH APPLICATIONS, INTL CONF ON BIG DATA & CLOUD COMPUTING, INTL SYMP SOCIAL COMPUTING & NETWORKING, INTL CONF ON SUSTAINABLE COMPUTING & COMMUNICATIONS (ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM 2020) | 2020年
关键词
Pervasive Computing; Internet of Things; Edge Computing; Data Storage; Accuracy; Probabilistic Model;
D O I
10.1109/ISPA-BDCloud-SocialCom-SustainCom51426.2020.00152
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Pervasive Computing (PC) deals with the placement of services and applications around end users for facilitating their everyday activities. Current advances on the Internet of Things (IoT) and the Edge Computing (EC) provide the room for adopting their infrastructures and hosting the desired services for supporting PC applications. Numerous devices present in IoT and EC infrastructures give the opportunity to record and process data through the interaction with users and their environment. Upon these data, the appropriate processing should be realized as requested by end users or applications. It is efficient to process such requests as close as possible to end users to limit the latency in the provision of responses. The research community, identifying this need, proposes the use of the EC as the appropriate place to perform the discussed processing which has the form of tasks or queries. Tasks/queries set specific conditions for data they desire imposing a number of requirements for the dataset upon which the desired processing should be executed. It is wise to pre-process the data and detect their statistics to know beforehand if it is profitable to have any dataset as part of the requested processing. This paper focuses on a model that is responsible to efficiently distribute the collected data to the appropriate datasets. We store similar data to the same datasets and keep their statistics solid (i.e., we meet a low deviation) through the use of a probabilistic approach. The second part of the proposed approach is related to an aggregation scheme upon multiple outlier detection methods. We decide to transfer outliers to Cloud avoiding to store them locally as they will jeopardize the solidity of datasets. If data are going to be locally stored, we provide a mechanism for selecting the most appropriate dataset to host them while we perform a controlled replication to support a fault tolerant system. The performance of the proposed models is evaluated by a high number of experiments for different scenarios.
引用
收藏
页码:1006 / 1013
页数:8
相关论文
共 41 条
  • [1] Amrutha S., 2016, INDIAN J SCI TECHNOL, V9, P1, DOI DOI 10.17485/ijst/2016/v9i48/108022
  • [2] [Anonymous], 2011, ENCY MATH ITS APPL
  • [3] [Anonymous], 2009, IEEE ICPCC
  • [4] [Anonymous], APPL MULTIPLE REGRES
  • [5] Baszczynska A., 2016, STAT TRANSITION, V17, P541, DOI DOI 10.21307/STATTRANS-2016-037
  • [6] Baxter R., 2001, P DSC
  • [7] Bonvin N., 2010, Proceedings of the 1st ACM symposium on cloud computing, P205, DOI DOI 10.1145/1807128.1807162
  • [8] Energy-efficient data replication in cloud computing datacenters
    Boru, Dejene
    Kliazovich, Dzmitry
    Granelli, Fabrizio
    Bouvry, Pascal
    Zomaya, Albert Y.
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2015, 18 (01): : 385 - 402
  • [9] Cong G, 2007, P VLDB, P1
  • [10] DETECTION OF INFLUENTIAL OBSERVATION IN LINEAR-REGRESSION
    COOK, RD
    [J]. TECHNOMETRICS, 1977, 19 (01) : 15 - 18