Challenge of Remembering from fewer frames

The video instances contain a temporal dimension size that could show large variability, and some action classes do not require considering all frames to understand them. To favor fair comparisons between methods and datasets and reduce the huge memory consumption, we define the working memory size of the rehearsal methods in terms of stored frames. This creates a new unique scenario of Class Incremental Learning in video data, in which rehearsal methods must decide first what subset of frames should be selected and then decide what video to store according to selected frames per video. We encourage future works on Class Incremental Learning for Video Understanding to focus on this and beat our Temporal Consistency Regulation strategy.

ModelFrames per videoKineticsActivityNet-TrimUCF101
Mem. Frame CapacityAccBWFMem. Frame CapacityAccBWFMem. Frame CapacityAccBWF
iCaRL43.2 × 10430.73%40.36%1.6 × 10421.63%36.98%8.08 × 10380.32%17.13%
iCaRL86.4 × 10432.04%38.48%3.2 × 10421.54%33.41%16.16 × 10381.12%18.25%
iCaRL1612.8 × 10431.36%38.74%6.4 × 10425.27%29.71%32.32 × 10381.06%18.23%
iCaRLALL2 × 10632.04%38.74%15.5 × 10648.53%19.72%3.69 × 10580.97%18.11%
iCaRL+TC43.2 × 10435.32%34.07%1.6 × 10442.99%23.82%8.08 × 10373.85%26.35%
iCaRL+TC86.4 × 10436.24%33.83%3.2 × 10445.73%18.90%16.16 × 10374.25%25.27%
iCaRL+TC1612.8 × 10436.54%33.53%6.4 × 10444.04%22.82%32.32 × 10375.84%23.23%