Online Continual Learning through Mutual Information Maximization_윤형석발표
페이지 정보
작성자 최고관리자 댓글 조회 작성일 24-01-08 13:07본문
This paper proposes a new online continual learning technique called OCM based on mutual information maximization. It achieves two objectives
that are critical in dealing with catastrophic forgetting (CF). (1) It reduces feature bias caused
by cross entropy (CE) as CE learns only discriminative features for each task, but these features
may not be discriminative for another task. To
learn a new task well, the network parameters
learned before have to be modified, which causes
CF. The new approach encourages the learning
of each task to make use of holistic representations or the full features of the task training
data. (2) It encourages preservation of the previously learned knowledge when training a new
batch of incrementally arriving data. Empirical
evaluation shows that OCM substantially outperforms the online CL baselines. For example, for
CIFAR10, OCM improves the accuracy of the
best baseline by 13.1% from 64.1% (baseline) to
77.2% (OCM)
첨부파일
- 20230517_윤형석.pptx (2.7M) 0회 다운로드 | DATE : 2024-01-08 13:07:20
댓글목록
등록된 댓글이 없습니다.