This talk is on the analysis and applications of a constructive architecture for training Deep Neural Networks (DNNs), which are usually trained by End-to-End (E2E) gradient propagation with fixed depths. E2E training of DNNs has proven to offer impressive performances in a number of applications such as computer vision, machine translation and in playing complex games such as GO. However, the massive cost in computing and memory hinders its applications in many areas, such as portable medical devices. Moreover, the majority of DNNs are data hungry which raises further barriers of applications in regions where data collection or labelling is expensive. As an alternative, Cascade Learning (CL), the approach of interest here, trains networks in a layer-wise fashion and has been demonstrated to achieve satisfactory performance in large scale tasks such as the popular ImageNet benchmark dataset, at substantially reduced computing and memory requirements. Here we focus on the nature of features extracted from CL. By attempting to explain the process of learning using the Information Bottleneck theory, an empirical rule (Information Transition Ratio) is derived to automatically determine a satisfactory depth for Deep Neural Networks. We suggest that CL packs information in a hierarchical manner, with coarse features in early layers and more task specific features in later layers. This is verified by considering Transfer Learning whereby features learned from a data-rich source domain assist in learning a data-sparse target domain. Using a wide range of inference problems in medical imaging, human activity recognition and inference from single cell gene expression between mice and humans, Transfer Learning from a cascade trained model shows significant advantages in small data regime.
The seminar will be held in a hybrid format. We strongly encourage you to participate in person at MR 2 , Centre of Mathematical Sciences, CB3 0WA . Althernatively you can join using the following Zoom link:
Meeting ID: 998 0160 5037
Security Passcode: 435654
This is a seminar held jointly by CMIH and Cambridge Image Analysis (CIA). Please note the change of date; it will take place on Friday instead of the normal Wednesday.
- Speaker: Doris Xin Du
- Friday 10 June 2022, 13:00–14:00
- Venue: This talk will be held in a hybrid format at MR 2, Centre of Mathematical Sciences, CB3 0WA. Alternatively you can also join with Zoom (see abstract for Zoom link)..
- Series: CMIH Hub seminar series; organiser: Yuan Huang.