Pre-image calculation for random Fourier feature kernel machines

Research output: Contribution to conferencePaperpeer-review

Abstract

In this paper we address the pre-image problem for random Fourier feature based kernel machines. Full kernel methods require the calculation of the Gram matrix of scalar products between all training samples. This is not feasible for large sample sizes, and various approximation methods have been proposed previously. Calculation of the inverse transform, known as finding the pre-image, is frequently required, and has been tackled previously for the full kernel approach, but has been less well explored for approximation methods. In this paper we investigate two approaches suitable for finding the pre-image from random Fourier features. The first is a direct adaptation of a learning-based approach to the random features case. This requires the solution of a completely separate optimisation problem, with additional model parameters and a second pass through the data. The second approach only requires solving a single optimisation problem with an augmented vector and a single pass through the data samples. We compare the two methods on four datasets and demonstrate their effectiveness.
Original languageEnglish
Pages433-444
Number of pages12
Publication statusPublished - 01 Feb 2024
EventThe 22nd UK Workshop on Computational Intelligence - Aston University, Birmingham, United Kingdom of Great Britain and Northern Ireland
Duration: 06 Sept 202308 Sept 2023
Conference number: 22
https://www.uk-ci.org/home

Conference

ConferenceThe 22nd UK Workshop on Computational Intelligence
Abbreviated titleUKCI 2023
Country/TerritoryUnited Kingdom of Great Britain and Northern Ireland
CityBirmingham
Period06 Sept 202308 Sept 2023
Internet address

Keywords

  • kernel pre-image
  • radial basis functions
  • random Fourier features
  • kernel PCA

Fingerprint

Dive into the research topics of 'Pre-image calculation for random Fourier feature kernel machines'. Together they form a unique fingerprint.

Cite this