Readings
Week 2:
-
Paper A: Foundations & Recent Trends in Multimodal Machine Learning Definitions, Challenges, & Open Questions - Section 1, Section 2, Section 3
-
Paper B: Foundations & Recent Trends in Multimodal Machine Learning Definitions, Challenges, & Open Questions - Section 1, Section 2, Section 4
-
Paper C: Foundations & Recent Trends in Multimodal Machine Learning Definitions, Challenges, & Open Questions - Section 1, Section 2, Section 5
-
Paper D: Foundations & Recent Trends in Multimodal Machine Learning Definitions, Challenges, & Open Questions - Section 1, Section 2, Section 6
- Paper E: Foundations & Recent Trends in Multimodal Machine Learning Definitions, Challenges, & Open Questions - Section 1, Section 2, Section 7
- Paper F: Foundations & Recent Trends in Multimodal Machine Learning Definitions, Challenges, & Open Questions - Section 1, Section 2, Section 8
Week 3:
- Zeiler and Fergus, Visualizing and Understanding Convolutional Networks. ECCV 2014
- Selvaraju et al., Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. ICCV 2017
- Karpathy et al., Visualizing and Understanding Recurrent Networks. arXiv 2015
- Khandelwal et al., Sharp Nearby, Fuzzy Far Away: How Neural Language Models Use Context. ACL 2018
- (optional reading) Learning Translation Invariance in CNNs
Week 5: For this week, you are expected to read one of the following papers:
- Paper A: Multimodal Learning using Optimal Transport for Sarcasm and Humor Detection
- Paper B: Deep Multimodal Clustering for Unsupervised Audiovisual Learning
- Paper C: On the Benefits of Early Fusion in Multimodal Representation Learning
- Paper D: Improving Multimodal fusion via Mutual Dependency Maximisation
(Optional Readings)