industry news
Subscribe Now

Black-Box Forgetting: A New Method for Tailoring Large AI Models

Researchers develop a method to selectively remove specific information from models, enhancing efficiency and improving privacy
Pretrained large-scale AI models need to ‘forget’ specific information for privacy and computational efficiency, but no methods exist for doing so in black-box vision-language models, where internal details are inaccessible. Now, researchers from Japan addressed this issue through an innovative strategy based on latent context sharing, successfully getting an image classifier to forget multiple classes it was trained on. Their findings could expand the use cases of large-scale AI models while safeguarding end users’ privacy.
The capabilities of large-scale pre-trained AI models have recently skyrocketed, as demonstrated by large-scale vision-language models like CLIP or ChatGPT. These typical generalist models can perform reasonably well in tasks covering a large variety of fields, which has paved the way for their widespread adoption by the public. However, such versatility no doubt comes at a cost.
Training and operating large-scale models consume extreme amounts of energy and time, which goes against sustainability goals and limits the types of computers they can be deployed on. Moreover, in many practical applications, people want AI models to fulfil specific roles rather than be jacks-of-all-trades. In such cases, a model’s generalist capabilities might be useless and even counter-productive, reducing accuracy. Could there be a way to leverage large-scale pre-trained models more efficiently by having them ‘forget’ unnecessary information?
In a recent paper that will be presented in Neural Information Processing Systems (NeurIPS 2024), a research team led by Associate Professor Go Irie from Tokyo University of Science (TUS), Japan, sought to tackle this problem. They developed a methodology dubbed “black-box forgetting,” by which one can iteratively optimize the text prompts presented to a black-box vision-language classifier model to have it selectively ‘forget’ some of the classes it can recognize. Co-authors of this study included Mr. Yusuke Kuwana and Mr. Yuta Goto, both from TUS, as well as Dr. Takashi Shibata from NEC Corporation.
In practical applications, the classification of all kinds of object classes is rarely required. For example, in an autonomous driving system, it would be sufficient to recognize limited classes of objects such as cars, pedestrians, and traffic signs. We would not need to recognize food, furniture, or animal species,” explains Dr. Irie, “Retaining the classes that do not need to be recognized may decrease overall classification accuracy, as well as cause operational disadvantages such as the waste of computational resources and the risk of information leakage.
Although some methods for selective forgetting in pre-trained models do exist, these assume a white-box setting, where the user has access to the internal parameters and architecture of the model. More often than not, users deal with black-boxes; they do not have access to the model itself or most of its information due to commercial or ethical reasons. Thus, the researchers had to employ a so-called derivative-free optimization strategy—one that does not require access to the model’s gradients.
To this end, they extended a method known as CMA-ES, with the image classifier model CLIP as the target model for this study. This evolutionary algorithm involves sampling various candidate prompts to feed to the model and evaluating the results via predefined objective functions, updating a multivariate distribution based on the calculated values.
However, the performance of derivative-free optimization techniques deteriorates quickly for large-scale problems. As more classes need to be forgotten, the ‘latent context’ used to optimize the input prompts grows to unmanageable sizes. To address this issue, the research team came up with a new parametrization technique called ‘latent context sharing.’ This approach involves decomposing latent context derived from prompts into various smaller elements, which are considered to be ‘unique’ to a prompt token or ‘shared’ between multiple tokens. By optimizing aiming to optimize for these smaller units rather than large chunks of latent context, the dimensionality of the problem can be greatly reduced, making it much more tractable.
The researchers validated their approach using several benchmark image classification datasets, trying to get CLIP to ‘forget’ 40% of the classes in a given dataset. This marks the first study in which the goal is to have a pre-trained vision-language model fail to recognize specific classes under black-box conditions and, based on reasonable performance baselines, the results were very promising.
This innovative method has important implications in the field of artificial intelligence and machine learning. It could help large-scale models perform better in specialized tasks, extending their already astounding applicability. Another use, for example, would be to prevent image generation models from producing undesirable content by having them forget specific visual contexts.
In addition, the proposed method could help tackle privacy issues, which are a rising concern in the field. “If a service provider is asked to remove certain information from a model, this can be accomplished by retraining the model from scratch by removing problematic samples from the training data. However, retraining a large-scale model consumes enormous amounts of energy,” says Dr. Irie, “Selective forgetting, or so-called machine unlearning, may provide an efficient solution to this problem.” In other words, it could help develop solutions for protecting the so-called “Right to be Forgotten,” which is a particularly sensitive topic in healthcare and finances.
This groundbreaking approach not only empowers large-scale AI models but also safeguards end users, paving the way for seamless integration of AI into our daily lives!
 
***
 
About The Tokyo University of Science
Tokyo University of Science (TUS) is a well-known and respected university, and the largest science-specialized private research university in Japan, with four campuses in central Tokyo and its suburbs and in Hokkaido. Established in 1881, the university has continually contributed to Japan’s development in science through inculcating the love for science in researchers, technicians, and educators.
With a mission of “Creating science and technology for the harmonious development of nature, human beings, and society,” TUS has undertaken a wide range of research from basic to applied science. TUS has embraced a multidisciplinary approach to research and undertaken intensive study in some of today’s most vital fields. TUS is a meritocracy where the best in science is recognized and nurtured. It is the only private university in Japan that has produced a Nobel Prize winner and the only private university in Asia to produce Nobel Prize winners within the natural sciences field.
 
About Associate Professor Go Irie from Tokyo University of Science
Dr. Go Irie graduated in Engineering from Keio University in 2004 and obtained a PhD degree from The University of Tokyo in 2011. He specializes in pattern recognition, machine learning, and media understanding, with a strong focus on perceptual information processing. He has participated in numerous international conferences on these topics, including prestigious ones such as NeurIPS, CVPR, and ACM Multimedia.

Leave a Reply

featured blogs
Dec 19, 2024
Explore Concurrent Multiprotocol and examine the distinctions between CMP single channel, CMP with concurrent listening, and CMP with BLE Dynamic Multiprotocol....
Dec 20, 2024
Do you think the proton is formed from three quarks? Think again. It may be made from five, two of which are heavier than the proton itself!...

featured video

Introducing FPGAi – Innovations Unlocked by AI-enabled FPGAs

Sponsored by Intel

Altera Innovators Day presentation by Ilya Ganusov showing the advantages of FPGAs for implementing AI-based Systems. See additional videos on AI and other Altera Innovators Day in Altera’s YouTube channel playlists.

Learn more about FPGAs for Artificial Intelligence here

featured chalk talk

SiC-Based High-Density Industrial Charger Solutions
Sponsored by Mouser Electronics and onsemi
In this episode of Chalk Talk, Amelia Dalton and Prasad Paruchuri from onsemi explore the benefits of silicon carbide based high density industrial charging solutions. They investigate the topologies of Totem Pole PFC and Half Bridge LLC circuits, the challenges that bidirectional CLLC resonant DC-DC converters are solving today, and how you can take advantage of onsemi’s silicon carbide charging solutions for your next design.
May 21, 2024
37,625 views