Learning Global Models Based on Distributed Data Abstractions - IJCAI
Recommend Documents
global model learning based on solely local GMM ... as a rough ensemble illustration. Thus, by ... KL-divergence is adopted to compute Rlk (E-step), resulting in.
give them quick practical reasoning in complex and often ill-defi ned problem .... The quality of problems and the way they are presented have proved to be very ...
across actors over time, as well as what actions were taken by the actors to drive the ... ping the process of model building can greatly reduce the ... given time period: 0 means no conflict, 5 means all-out war. ... This news event example includes
The decision making process of an actor â how it decides to use its resources to ... average life expectancy to predict conflict [O'Brien, 2002]. Often, statistical models built from .... sults of the models. First are standard machine learning me-
In robotics, forward models enable a robot to predict the con- sequences of its actions: .... The method of choosing motor commands in the ex- periment in this ...
is controlled and explored using 'motor babbling'. Feedback about its motor system comes from com- puter vision techniques requiring no prior informa-.
multiple hospitals without data leaving the hospital â A real life proof of concept ...... source ETL solutions with Pentaho Data Integration. John Wiley & Sons ...
Mar 3, 2017 - 2 Department of Computer Science and Engineering, Waseda University. 3 HIIT .... like to use the aggregate data for learning, but the clients.
235-245, 1973. Z. Pylyshyn, What the Mind's Eye Tells the Mind's Brain, PSYCHOLOGICAL. BULLETIN 80, 1-24, 1973. C. Hewitt, Description and Theoretic-.
Again, we define a possibility distribu- tion allowing to determine the stable models. 1 Introduction. Answer set programming (ASP) is an appropriate formalism.
results illustrate the advantage of the proposed filters for sparse data recovery. ... tive radio [45], and spectrum est
vide original theoretical analysis that offers the basis of a sound and efficient algorithm for learning the structure of an. RCM from relational data. We describe ...
acquired on a higher programming level which is characterized through the ...... control rules for planning (PRODIGY) [Minton et al., 1989], PABS can acquire.
Nov 19, 2013 - Fothergill-Gillmore LA, Rigden DJ, Michels PA, Phillips SE: The structure of ... Bouckaert RR, Frank E, Hall MA, Holmes G, Pfahringer B, ...
Apr 27, 2018 - students of junior high school in Karanganyar under the academic year 2017/2018. Data collected in ... experienced by junior high school students in the district of Karanganyar. The low in learning ..... Edina, MN: Interaction.
Chung, 1997], where dmax is the maximum rowsum of A. Here, transition probabilities are ..... [Yu and Shi, 2001] Stella Yu and Jianbo Shi. Grouping with bias.
Aggregation versus selection bias, and relational neural networks. 22. Hendrik Blockeel ... Parameter estimation for stochastic context-free graph grammars. 98. Tim Oates .... COMPLEX. PHENOTYPE. MOTIF. C_ID. Figure 1: Example database. I ID, respect
Tom Croonenborghs, Jan Ramon, Hendrik Blockeel and Maurice Bruynooghe. K.U.Leuven, Dept. .... Reinforcement Learning (RL) [Sutton and Barto, 1998] is of-.
inter-related statistical query translation models. First, a semantic network is automatically generated from large-scale English-Chinese bilingual parallel corpora ...
Abstract. To apply hierarchical task network (HTN) plan- ning to real-world planning problems, one needs to encode the HTN schemata and action mod-.
for the amount of exposure radiation dose received by staff. In terms of exposed radiation dose of staff, we obtained the well agreement on distance factor ...
On the secondary of the coupling transformer shown in Fig. 6, a VSC is connected for the reactive power variation contro
rectangular heliostat (e.g. GM-100) and (b) circular stretched membrane .... The resulting Z-displacement of the mirror plane is: , ( ) . ped g z x elev x. â. = -â. (6).
electrochemical measurements. (Tafel experiments) were carried out with an EG&G model .... [6] Schultz M. F., Benjamin M. M., Ferguson J. F., Adsorption and.
Learning Global Models Based on Distributed Data Abstractions - IJCAI
Xiaofeng Zhang and William K. Cheung. Department of ... {xfzhang,william}@comp.hkbu.edu.hk ... ture model (GMM) [Bishop, 1995] is chosen for the local data.
Learning Global Models Based on Distributed Data Abstractions Xiaofeng Zhang and William K. Cheung Department of Computer Science Hong Kong Baptist University Kowloon Tong, Hong Kong {xfzhang,william}@comp.hkbu.edu.hk Abstract Due to the increasing demand of massive and distributed data analysis, achieving highly accurate global data analysis results with local data privacy preserved becomes an increasingly important research issue. In this paper, we propose to adopt a model-based method (Gaussian mixture model) for local data abstraction and aggregate the local model parameters for learning global models. To support global model learning based on solely local GMM parameters instead of virtual data generated from the aggregated local model, a novel EM-like algorithm is derived. Experiments have been performed using synthetic datasets and the proposed method was demonstrated to be able to achieve the global model accuracy comparable to that of using the data regeneration approach at a much lower computational cost.
1
Introduction
Most machine learning and data mining algorithms work with a rather basic assumption that all the data can first be pooled together in a centralized data repository. Recently, there exist a growing number of cases that the data have to be physically distributed due to some practical constraints, such as privacy concerns. The need to tackle them calls for the recent research interest in distributed data mining. A common methodology for distributed data mining is to first perform local data analysis and then combine the results to form a global one [Kargupta et al., 2000]. One major limitation is that local analysis may result in loss of information for global pattern discovery. In [Zhang et al., 2004], a model parameter exchange approach was proposed to improve the global model accuracy, which however is not a generic methodology which can easily be generalized to different types of global models. An alternative is to adopt a flexible enough model for local data abstraction and have the subsequent analysis based on the local model parameters, which is what we adopt in this paper. As inspired by [Merugu and Ghosh, 2003], Gaussian mixture model (GMM) [Bishop, 1995] is chosen for the local data abstraction. Via different parameter settings, GMM can interpolate a set of data with more detailed information as well
as a rough ensemble illustration. Thus, by sharing only local model parameters for global analysis, the degree of data privacy preservation and the bandwidth requirement in a distributed environment can readily be controlled. In this paper, an EM-like algorithm is proposed for learning a global GMM directly from the aggregated version of the local abstractions.