Sorry, you need to enable JavaScript to visit this website.

PRIVACY PROTECTION IN LEARNING FAIR REPRESENTATIONS

Citation Author(s):
Yulu Jin, Lifeng Lai
Submitted by:
Yulu Jin
Last updated:
5 May 2022 - 6:19pm
Document Type:
Presentation Slides
Document Year:
2022
Event:
Presenters:
Yulu Jin
Paper Code:
IFS-5.1
 

In this paper, we develop a framework to achieve a desirable trade-off between fairness, inference accuracy and privacy protection in the inference as service scenario. Instead of sending raw data to the cloud, we conduct a random mapping of the data, which will increase privacy protection and mitigate bias but reduce inference accuracy. To properly address the trade-off, we formulate an optimization problem to find the optimal transformation map. As the problem is nonconvex in general, we develop an iterative algorithm to find the desired map. Numerical examples show that the proposed method has better performance than gradient ascent in the convergence speed, solution quality and algorithm stability.

up
0 users have voted: