Skip to yearly menu bar Skip to main content


Fast and Adversarial Robust Kernelized SDU Learning

Yajing Fan · wanli shi · Yi Chang · Bin Gu

MR1 & MR2 - Number 81
[ ]
Thu 2 May 8 a.m. PDT — 8:30 a.m. PDT

Abstract: SDU learning, a weakly supervised learning problem with only pairwise similarities, dissimilarities data points and unlabeled data available, has many practical applications. However, it is still lacking in defense against adversarial samples, and its learning process can be expensive. To address this gap, we propose a novel adversarial training framework for SDU learning.Our approach reformulates the conventional minimax problem as an equivalent minimization problem based on the kernel perspective, departing from traditional confrontational training methods. Additionally, we employ the random gradient method and random features to accelerate the training process. Theoretical analysis shows that our method can converge to a stationary point at a rate of $\mathcal{O}(1/T^{1/4})$.Our experimental results show that our algorithm is superior to other adversarial training methods in terms of generalization, efficiency and scalability against various adversarial attacks.

Live content is unavailable. Log in and register to view live content