References
- Blanchard, G., Lee, G., Scott, C. (2011). Generalizing from several related classification tasks to a new unlabeled sample. In: Advances in Neural Information Processing Systems, Vol. 24.
- Muandet, K., Balduzzi, D., Schölkopf, B. (2013). Domain generalization via invariant feature representation. In: International Conference on Machine Learning, PMLR, pp. 10–18.
- Blanchard, G., Deshmukh, A. A., Dogan, Ü., Lee, G., Scott, C. (2021). Domain generalization by marginal transfer learning. J. Mach. Learn. Res. 22(1):46–100.
- Baxter, J. (1998). Theoretical models of learning to learn. In: Learning to Learn. Springer, pp. 71–94.
- Maurer, A. (2005). Algorithmic stability and meta-learning. J. Mach. Learn. Res. 6:967–994.
- Evgeniou, T., Micchelli, C. A., Pontil, M., Shawe-Taylor, J. (2005). Learning multiple tasks with kernel methods. J. Mach. Learn. Res. 6(4):615–637.
- Mollenhauer, M., Mücke, N., Sullivan, T. (2022). Learning linear operators: Infinite-dimensional regression as a well-behaved non-compact inverse problem. arXiv preprint arXiv:2211.08875.
- Jin, J., Lu, Y., Blanchet, J., Ying, L. (2022). Minimax optimal kernel operator learning via multilevel training. arXiv preprint arXiv:2209.14430.
- Tong, H. Z., Hu, L. F., Ng, M. (2022). Non-asymptotic error bound for optimal prediction of function-on-function regression by RKHS approach. Acta Math. Sinica, English Ser. 38(4):777–796. DOI: 10.1007/s10114-021-9346-4.
- Szabó, Z., Sriperumbudur, B. K., Póczos, B., Gretton, A. (2016). Learning theory for distribution regression. J. Mach. Learn. Res. 17(1):5272–5311.
- Gretton, A., Borgwardt, K., Rasch, M., Schölkopf, B., Smola, A. (2006). A kernel method for the two-sample-problem. In: Advances in Neural Information Processing Systems, Vol. 19.
- Sriperumbudur, B. K., Gretton, A., Fukumizu, K., Schölkopf, B., Lanckriet, G. R. (2010). Hilbert space embeddings and metrics on probability measures. J. Mach. Learn. Res. 11:1517–1561.
- Dudley, R. M. (2002). Cambridge Studies in Advanced mathematics: Real Analysis and Probability, Vol. 74, 2nd ed. Cambridge: Cambridge University Press.
- Caponnetto, A., De Vito, E. (2007). Optimal rates for the regularized least-squares algorithm. Found. Comput. Math. 7(3):331–368. DOI: 10.1007/s10208-006-0196-8.
- Guo, Z.-C., Lin, S.-B., Zhou, D.-X. (2017). Learning theory of distributed spectral algorithms. Inverse Probl. 33(7):074009. DOI: 10.1088/1361-6420/aa72b2.
- Wolfer, G., Alquier, P. (2022). Variance-aware estimation of kernel mean embedding. arXiv preprint arXiv:2210.06672.
- Cucker, F., Smale, S. (2002). On the mathematical foundations of learning. Bull. Amer. Math. Soc. 39(1):1–49. DOI: 10.1090/S0273-0979-01-00923-5.
- Schölkopf, B., Herbrich, R., Smola, A. J. (2001). A generalized representer theorem. In: Annual Conference on Computational Learning Theory (COLT). Springer, pp. 416–426.
- Gizewski, E. R., Mayer, L., Moser, B. A., Nguyen, D. H., Pereverzyev Jr., S., Pereverzyev, S. V., Shepeleva, N., Zellinger, W. (2022). On a regularization of unsupervised domain adaptation in RKHS. Applied and Computational Harmonic Analysis, 57:201–227. DOI: 10.1016/j.acha.2021.12.002.
- Olver, F. W. J., Lozier, D. W., Boisvert, R. F., Clark, C. W. (2010). NIST Handbook of Mathematical Functions, 1 pap/cdr ed. Cambridge: Cambridge University Press.
- Engl, H. W., Hanke, M., Neubauer, A. (1996). Regularization of Inverse Problems, Vol. 375. Dordrecht: Springer.
- Pinelis, I. (1994). Optimum bounds for the distributions of martingales in Banach spaces. Ann. Probab. 22:1679–1706. DOI: 10.1214/aop/1176988477.