Fast transfer Gaussian process regression with large-scale sources

作者:

Highlights:

摘要

In transfer learning, we aim to improve the predictive modeling of a target output by using the knowledge from some related source outputs. In real-world applications, the data from the target domain is often precious and hard to obtain, while the data from source domains is plentiful. Thus, since the complexity of Gaussian process based multi-task/transfer learning approaches grows cubically with the total number of source+ target observations, the method becomes increasingly impractical for large (>104) source data inputs even with a small amount of target data. In order to scale known transfer Gaussian processes to large-scale source datasets, we propose an efficient aggregation model in this paper, which combines the predictions from distributed (small-scale) local experts in a principled manner. The proposed model inherits the advantages of single-task aggregation schemes, including efficient computation, analytically tractable inference, and straightforward parallelization during training and prediction. Further, a salient feature of the proposed method is the enhanced expressiveness in transfer learning — as a byproduct of flexible inter-task relationship modelings across different experts. When deploying such models in real-world applications, each local expert corresponds to a lightweight predictor that can be embedded in edge devices, thus catering to cases of online on-mote processing in fog computing settings.

论文关键词:Transfer learning,Large-scale,Gaussian process,Aggregation models,Edge intelligence

论文评审过程:Received 30 July 2018, Revised 18 November 2018, Accepted 19 November 2018, Available online 15 December 2018, Version of Record 7 January 2019.

论文官网地址:https://doi.org/10.1016/j.knosys.2018.11.029