PowerPoint 演示文稿

1 of
Published on Video
Go to video
Download PDF version
Download PDF version
Embed video
Share video
Ask about this video

Page 1 (0s)

Chengming Xu cmxu18@fudan.edu.cn chmxu.github.io.

Page 2 (8s)

Personal Experience. Research Interests Few-shot learning, action understanding, foundation model adaptation.

Page 3 (29s)

Few-shot Learning. Normal deep learning training Large-scale dataset Sufficient data for each category.

Page 4 (40s)

Few-shot Learning. Normal deep learning training Large-scale dataset Sufficient data for each category.

Page 5 (49s)

Few-shot Learning. Base data. Novel data.

Page 6 (57s)

FSL Applications. Lee X Y, Vidyaratne L, Alam M, et al. XDNet: A Few-Shot Meta-Learning Approach for Cross-Domain Visual Inspection[C]//CVPRW2023.

Page 7 (1m 12s)

FSL Applications. svpdouov øyo uy sofieull Indul.

Page 8 (1m 29s)

FSL & meta-learning. Stn.

Page 9 (1m 37s)

Developing FSL algorithms. Deep learning based FSL.

Page 10 (1m 49s)

Developing FSL algorithms. Deep learning based FSL.

Page 11 (1m 58s)

General Augmentation. image Mixup Cutout CutMix.

Page 12 (2m 6s)

Augment training data: PatchMix. FSL methods generally suffer from problem of distribution shift and spurious correlation.

Page 13 (2m 22s)

Augment training data: PatchMix. Xu C, Liu C, Sun X, et al. PatchMix Augmentation to Identify Causal Features in Few-shot Learning[J]. TPAMI2022..

Page 14 (2m 35s)

Augment training data: PatchMix. Xu C, Liu C, Sun X, et al. PatchMix Augmentation to Identify Causal Features in Few-shot Learning[J]. TPAMI2022..

Page 15 (2m 44s)

Augment training data: PatchMix. Papyan V, Han X Y, Donoho D L. Prevalence of neural collapse during the terminal phase of deep learning training[J]. PNAS2020..

Page 16 (3m 6s)

Augment training data: PatchMix. Xu C, Liu C, Sun X, et al. PatchMix Augmentation to Identify Causal Features in Few-shot Learning[J]. TPAMI2022..

Page 17 (3m 24s)

Augment training data: PatchMix. Xu C, Liu C, Sun X, et al. PatchMix Augmentation to Identify Causal Features in Few-shot Learning[J]. TPAMI2022..

Page 18 (3m 43s)

Augment training data: PatchMix. Xu C, Liu C, Sun X, et al. PatchMix Augmentation to Identify Causal Features in Few-shot Learning[J]. TPAMI2022..

Page 19 (5m 22s)

Cross-domain settings on CUB, Cars, Places, Plantae Unsupervised setting on miniImageNet.

Page 20 (5m 32s)

Developing FSL algorithms. Deep learning based FSL.

Page 21 (5m 46s)

cÄc:Ä Soft Nearest Neighbor Classification Scores (a) Instance Embedding Embedding Adaptation Set-to-Set Function Soft Nearest Neighbor Test Instance Task Agnostic Embedding Task Specific Em bedding I Classification Scores (b) Embedding Adaptation.

Page 22 (6m 8s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 23 (6m 30s)

Strengthen model: DMF. A novel dynamic meta-filter for feature alignment in FSL. Dynamic sampling and grouping strategy to improved flexibility and efficiency. Neural ODE for flexible feature alignment.

Page 24 (6m 48s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 25 (6m 55s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 26 (7m 9s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 27 (7m 17s)

Strengthen model: DMF. Not flexible: The alignment can be written as where ? is alignment function. By gradually decreasing the time step ?, such a recursive residual equation can be transformed into an ODE whose solution is the final alignment feature..

Page 28 (7m 35s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 29 (7m 45s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 30 (7m 55s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 31 (8m 7s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 32 (8m 20s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 33 (9m 26s)

Strengthen model: DMF. Xu C, Fu Y, Liu C, et al. Learning dynamic alignment via meta-filter for few-shot learning[C]//CVPR2021..

Page 34 (9m 34s)

Developing FSL algorithms. Deep learning based FSL.

Page 35 (9m 48s)

Test-time finetune. External data Pre-trained backbone Class 1 Domain A Class 4 Domain B Class 2 Class 5 Class 3 Class 6 Meta-trained backbone Support set Augmented support set Task-specifically fine-tuned backbone.

Page 36 (10m 3s)

Test-time finetune. ViTs have been less touched in FSL. PMF: naïve pretrain  meta-train  finetune pipeline Heavy computation: updating whole network parameters.

Page 37 (10m 14s)

Test-time finetune. ViTs have been less touched in FSL. PMF: naïve pretrain  meta-train  finetune pipeline Potential overfitting: cannot handle extremely small number of labeled images.

Page 38 (10m 27s)

Efficient Transformer Tuning. We propose a novel efficient Transformer Tuning (eTT) that is tailored for few-shot learning. Fewer learnable parameters, while being flexible and effective enough.

Page 39 (10m 47s)

Efficient Transformer Tuning. Xu C, Yang S, Wang Y, et al. Exploring Efficient Few-shot Adaptation for Vision Transformers[J]//TMLR 2022..

Page 40 (10m 55s)

Efficient Transformer Tuning. Domain residual adapter Adapter but light structure with only offset Represent gap between source and target domains, and transfer the original manifold to a more appropriate one.

Page 41 (11m 8s)

Efficient Transformer Tuning. Xu C, Yang S, Wang Y, et al. Exploring Efficient Few-shot Adaptation for Vision Transformers[J]//TMLR 2022..

Page 42 (12m 13s)

eTT. Xu C, Yang S, Wang Y, et al. Exploring Efficient Few-shot Adaptation for Vision Transformers[J]//TMLR 2022..

Page 43 (12m 26s)

Developing FSL algorithms. Deep learning based FSL.

Page 44 (12m 41s)

Worst-case risk minimization. Lei J, Li D, Xu C, et al. Worst-case Feature Risk Minimization for Data-Efficient Learning[J]//TMLR 2023..

Page 45 (12m 56s)

Worst-case risk minimization. Lei J, Li D, Xu C, et al. Worst-case Feature Risk Minimization for Data-Efficient Learning[J]//TMLR 2023..

Page 46 (13m 3s)

Worst-case risk minimization. Lei J, Li D, Xu C, et al. Worst-case Feature Risk Minimization for Data-Efficient Learning[J]//TMLR 2023..

Page 47 (14m 0s)

FSL in current research. Deep learning based FSL.

Page 48 (14m 14s)

Meta-learning in current research. Bar A, Gandelsman Y, Darrell T, et al. Visual prompting via image inpainting[J]. NIPS2022.

Page 49 (14m 28s)

FSL in current research. Deep learning based FSL.

Page 50 (14m 43s)

FSL in current research. Jia M, Tang L, Chen B C, et al. Visual prompt tuning[C]//ECCV2022. Zhang R, Fang R, Zhang W, et al. Tip-adapter: Training-free clip-adapter for better vision-language modeling[J]. arXiv preprint arXiv:2111.03930, 2021..