The current modus operandi in adapting pre-trained models involves updating all the backbone parameters, ie, full fine-tuning. This paper introduces Visual Prompt Tuning (VPT) as an efficient and effective alternative to full fine-tuning for large-scale Transformer models in vision. Taking inspiration from recent advances in efficiently tuning large language models, VPT introduces only a small amount (less than 1% of model parameters) of trainable parameters in the input space while keeping the model backbone frozen. Via extensive experiments on a wide variety of downstream recognition tasks, we show that VPT achieves significant performance gains compared to other parameter efficient tuning protocols. Most importantly, VPT even outperforms full fine-tuning in many cases across model capacities and training data scales, while reducing per-task storage cost. 
                        more » 
                        « less   
                    This content will become publicly available on June 23, 2026
                            
                            To Protect the LLM Agent Against the Prompt Injection Attack with Polymorphic Prompt
                        
                    More Like this
- 
            
- 
            The recent multi-messenger and multi-wavelength observations of gamma-ray bursts (GRBs) have encouraged renewed interest in these energetic events. In spite of the substantial amount of data accumulated during the past few decades, the nature of the prompt emission remains an unsolved puzzle. We present an overview of the leading models for their prompt emission phase, focusing on the perspective opened by future missions.more » « less
- 
            Large language models can perform downstream tasks in a zero-shot fashion, given natural language prompts that specify the desired behavior. Such prompts are typically hand engineered, but can also be learned with gradient-based methods from labeled data. However, it is underexplored what factors make the prompts effective, especially when the prompts are in natural language. In this paper, we investigate common attributes shared by effective prompts in classification problems. We first propose a human readable prompt tuning method (FluentPrompt) based on Langevin dynamics that incorporates a fluency constraint to find a distribution of effective and fluent prompts. Our analysis reveals that effective prompts are topically related to the task domain and calibrate the prior probability of output labels. Based on these findings, we also propose a method for generating prompts using only unlabeled data, outperforming strong baselines by an average of 7.0{\%} accuracy across three tasks.more » « less
 An official website of the United States government
An official website of the United States government 
				
			 
					 
					
