A research team from the University of Tokyo has outlined a new approach to training large language models that aims to curb sensitive data leakage while preserving performance, addressing one of the ...