Korthals, Timo: Deep generative models for multi-modal perception under the influence of ambiguity. 2021
Inhalt
- Introduction
- Contributions and Outline
- Deep Generative Models
- Deep Neural Network
- The Structure of Deep Neural Networks
- Learning Objectives
- Training Deep Neural Networks
- Further Remarks on Training Deep Neural Networks
- Generative Models
- Generative Model Framework and Learning
- Graphical Model
- Training of Generative Models with Latent Variables
- Linking Deep Neural Networks and Generative Models
- Multi-Modal Perception
- Multi-Modal Machine Learning – Definition and Taxonomy
- Multi-Modal Properties
- Heterogeneity of Multi-Modal Data
- Correlations between Modalities, Classes, and Attributes
- Requirements for Multi-Modal Data and Observation Ambiguities
- Conclusion and Challenges faced in this Work
- Deep Multi-Modal Machine Learning
- Multi-Modal Variational Autoencoder
- Preliminary Approaches
- M²VAE
- Derivation of the Bi-Modal M²VAE
- Extension to three Modalities
- Derivation of a General Expression for an Arbitrary Number of Modalities
- Realization as Deep Neural Network
- Conscious vs. Unconscious M²VAE
- Auto Re-Encoding
- Multi-Modal Data Sets and their Properties
- Review of Available Data Sets
- Proposed Data Sets
- Discussion and Choice of Suitable Data Sets
- Metrics, Evaluations, and Results
- Scores and Metrics
- Results
- Discussion
- Applications
- Active Sensing through Ambiguity
- Embedding M²VAE in a Learning Framework
- Evaluation
- Rubiks
- The Rubiks data set
- Intrinsic Curiosity Module (ICM) vs. Multi-Modal Variational Autoencoder (M²VAE)
- Results
- Active Sensing with Distributed and Heterogeneous Robots
- Discussion
- Conclusion and Outlook
- Acronyms and Abbreviations
- List of Figures
- List of Tables
- Bibliography
- Most Influential Conferences and Journals
- Supplemental Material
- List of Applied Software
- Architecture Setups and Assets
- VAE Training Setup
- CVAE Training Setup
- Re-Encoding Training Setup
- eMNIST CVAE Training Setup
- Hyperparameter Dependencies Training Setup
- XOR Training Setup
- MoG Training Setup
- Shared Weights Training Setup
- Competitive Evaluation Setup
- Rubiks Data Set Evaluation Setup
- AMiRo-CITrack Evaluation Setup
- Latent Space Statistics
- Semi-Supervised VAE
- List of Data Set Websites
- Alternative nomenclature for the Variational Autoencoder
- Derivation for the Joint Multi-Modal VAE via Variation of Information
- Derivation of KLD-Derivative Inequality
- Variation of Information
- Lie Groups
- CITrack & AMiRo
- Mathematical Foundations
- Expected Value of a Random Variable
- Further Quantities of a Random Variable
- Entropy
- Cross Entropy
- Kullback–Leibler Divergence (KLD)
- Jensen–Shannon Divergence (JSD)
- KLD for two Gaussian distributions
- Positiveness of Entropy for a Gaussian Distribution
- Jensen's Inequality
- Mixture of Gaussian versus Gaussian - Derivation via Jensen's Inequality
