showSidebars ==
showTitleBreadcrumbs == 1
node.field_disable_title_breadcrumbs.value ==

Pre-conference talk by CHEN Zhaozheng and LUO Zilin

Please click here if you are unable to view this page.

 

Pre-conference talk by CHEN Zhaozheng and LUO Zilin
DATE : 14 June 2023, Wednesday
TIME : 10:00am - 10:40am
VENUE : Meeting room 4.4, Level 4
School of Computing and Information Systems 1,
Singapore Management University,
80 Stamford Road,
Singapore 178902
Please register by 13 June 2023

 

 

 

There are 2 talks in this session, each talk is approximately 20 minutes. 
All sessions are for pre-conference talk for IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR 2023).

About the Talk (s)

Talk #1: Extracting Class Activation Maps from Non-Discriminative Features as well
by CHEN Zhaozheng, PhD Candidate

Extracting class activation maps (CAM) from a classification model often results in poor coverage on foreground objects, i.e., only the discriminative region (e.g., the “head” of “sheep”) is recognized and the rest (e.g., the “leg” of “sheep”) mistakenly as background. The crux behind is that the weight of the classifier (used to compute CAM) captures only the discriminative features of objects. We tackle this by introducing a new computation method for CAM that explicitly captures non-discriminative features as well, thereby expanding CAM to cover whole objects. Specifically, we omit the last pooling layer of the classification model, and perform clustering on all local features of an object class, where “local” means “at a spatial pixel position”. We call the resultant K cluster centers local prototypes — represent local semantics like the “head”, “leg”, and “body” of “sheep”. Given a new image of the class, we compare its unpooled features to every prototype, derive K similarity matrices, and then aggregate them into a heatmap (i.e., our CAM). Our CAM thus captures all local features of the class without discrimination. We evaluate it in the challenging tasks of weakly-supervised semantic segmentation (WSSS), and plug it in multiple state-of-the-art WSSS methods, such as MCTformer and AMN, by simply replacing their original CAM with ours. Our extensive experiments on standard WSSS benchmarks (PASCAL VOC and MS COCO) show the superiority of our method: consistent improvements with little computational overhead.

Talk #2: Class-Incremental Exemplar Compression for Class-Incremental Learning  
by LUO Zilin, PhD Candidate

Exemplar-based class-incremental learning (CIL) finetunes the model with all samples of new classes but few-shot exemplars of old classes in each incremental phase, where the “few-shot” abides by the limited memory budget. In this paper, we break this “few-shot” limit based on a simple yet surprisingly effective idea: compressing exemplars by downsampling non-discriminative pixels and saving “many-shot” compressed exemplars in the memory. Without needing any manual annotation, we achieve this compression by generating 0-1 masks on discriminative pixels from class activation maps (CAM). We propose an adaptive mask generation model called class-incremental masking (CIM) to explicitly resolve two difficulties of using CAM: 1) transforming the heatmaps of CAM to 0-1 masks with an arbitrary threshold leads to a trade-off between the coverage on discriminative pixels and the quantity of exemplars, as the total memory is fixed; and 2) optimal thresholds vary for different object classes, which is particularly obvious in the dynamic environment of CIL. We optimize the CIM model alternatively with the conventional CIL model through a bilevel optimization problem. We conduct extensive experiments on high-resolution CIL benchmarks including Food-101, ImageNet-100, and ImageNet-1000, and show that using the compressed exemplars by CIM can achieve a new state-of-the-art CIL accuracy, e.g., 4.8 percentage points higher than FOSTER on 10-Phase ImageNet-1000.

About the Speaker (s)
 

CHEN Zhaozheng is a Ph.D. candidate in the School of Computing and Information Systems at Singapore Management University under the supervision of Assistant Professor SUN Qianru. His general research interests lie in the field of computer vision, with a particular focus on weakly-supervised semantic segmentation.

 
 

LUO Zilin is a Ph.D. candidate in the School of Computing and Information Systems at Singapore Management University under the supervision of Assistant Professor, supervised by Asst. Prof. SUN Qianru. His primary research interest is in the field of computer vision, with a specific focus on  incremental learning.