Abstract:
We propose two novel transferability metrics fast optimal transport-based conditional entropy (F-OTCE) and joint correspondence OTCE (JC-OTCE) to evaluate how much the so...Show MoreMetadata
Abstract:
We propose two novel transferability metrics fast optimal transport-based conditional entropy (F-OTCE) and joint correspondence OTCE (JC-OTCE) to evaluate how much the source model (task) can benefit the learning of the target task and to learn more generalizable representations for cross-domain cross-task transfer learning. Unlike the original OTCE metric that requires evaluating the empirical transferability on auxiliary tasks, our metrics are auxiliary-free such that they can be computed much more efficiently. Specifically, F-OTCE estimates transferability by first solving an optimal transport (OT) problem between source and target distributions and then uses the optimal coupling to compute the negative conditional entropy (NCE) between the source and target labels. It can also serve as an objective function to enhance downstream transfer learning tasks including model finetuning and domain generalization (DG). Meanwhile, JC-OTCE improves the transferability accuracy of F-OTCE by including label distances in the OT problem, though it incurs additional computation costs. Extensive experiments demonstrate that F-OTCE and JC-OTCE outperform state-of-the-art auxiliary-free metrics by 21.1% and 25.8%, respectively, in correlation coefficient with the ground-truth transfer accuracy. By eliminating the training cost of auxiliary tasks, the two metrics reduce the total computation time of the previous method from 43 min to 9.32 and 10.78 s, respectively, for a pair of tasks. When applied in the model finetuning and DG tasks, F-OTCE shows significant improvements in the transfer accuracy in few-shot classification experiments, with up to 4.41% and 2.34% accuracy gains, respectively.
Published in: IEEE Transactions on Neural Networks and Learning Systems ( Volume: 36, Issue: 2, February 2025)
Funding Agency:
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Transfer Learning ,
- Computation Time ,
- Source Model ,
- Optimal Transport ,
- Domain Generalization ,
- Target Label ,
- Target Task ,
- Conditional Entropy ,
- Fine-tuned Model ,
- Negative Entropy ,
- Auxiliary Task ,
- Optimal Coupling ,
- Accurate Transfer ,
- Few-shot Classification ,
- Pair Of Tasks ,
- Data Sources ,
- Accuracy Of Model ,
- Target Domain ,
- Target Data ,
- Source Characteristics ,
- Source Task ,
- Source Domain ,
- Stochastic Gradient Descent Optimizer ,
- Transfer Performance ,
- Source Dataset ,
- Universal Model ,
- Wasserstein Distance ,
- Domain-specific Models ,
- Unseen Domains ,
- Support Set
- Author Keywords
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Transfer Learning ,
- Computation Time ,
- Source Model ,
- Optimal Transport ,
- Domain Generalization ,
- Target Label ,
- Target Task ,
- Conditional Entropy ,
- Fine-tuned Model ,
- Negative Entropy ,
- Auxiliary Task ,
- Optimal Coupling ,
- Accurate Transfer ,
- Few-shot Classification ,
- Pair Of Tasks ,
- Data Sources ,
- Accuracy Of Model ,
- Target Domain ,
- Target Data ,
- Source Characteristics ,
- Source Task ,
- Source Domain ,
- Stochastic Gradient Descent Optimizer ,
- Transfer Performance ,
- Source Dataset ,
- Universal Model ,
- Wasserstein Distance ,
- Domain-specific Models ,
- Unseen Domains ,
- Support Set
- Author Keywords