Towards Grounding Conceptual Spaces in Neural Representations - - PowerPoint PPT Presentation
Towards Grounding Conceptual Spaces in Neural Representations - - PowerPoint PPT Presentation
Towards Grounding Conceptual Spaces in Neural Representations Lucas Bechberger and Kai-Uwe Khnberger https://www.lucas-bechberger.de The Different Layers of Representation x :apple ( x ) red ( x ) Symbolic Layer Formal Logics
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 2
The Different Layers of Representation
Symbolic Layer Subsymbolic Layer [0.42; -1.337] ∀ x:apple( x)⇒red(x) Formal Logics Sensor Values, Machine Learning
?
Conceptual Layer Geometric Representation
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 3
Dimensions & Domains
- Quality dimensions
- Different ways stimuli are judged to be similar or different
- Interpretable by a human
- E.g., temperature, weight, brightness, pitch
- Domain
- Set of dimensions that inherently belong together
- Color: hue, saturation, and brightness
- Distance in this space is inversely related to similarity
- Geometric betweenness represents semantic betweenness
- Concepts: regions in this space
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 4
Example: The Color Domain
https://en.wikipedia.org/wiki/HSL_and_HSV#/media/File:HSL_color_solid_dblcone_chroma_gray.png
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 5
Towards A Thorough Implementation
Symbolic Layer Subsymbolic Layer Conceptual Layer
Manually define regions Manually define dimensions
representation learning
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 6
InfoGAN – Architecture
- Information Maximizing Generative Adversarial Networks
- X. Chen et al., “InfoGAN: Interpretable Representation Learning by Information Maximizing
Generative Adversarial Nets”, Advances in Neural Information Processing Systems, 2016
D
z c
G
G(z) x c ?
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 7
InfoGAN – MNIST Results
- X. Chen et al., “InfoGAN: Interpretable Representation Learning by Information Maximizing
Generative Adversarial Nets”, Advances in Neural Information Processing Systems, 2016
- Three latent variables
- Categorical (10 classes)
- Continuous (uniform)
- Continuous (uniform)
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 8
Domains and InfoGAN
Domains in CS framework
- Interpretable dimensions
- Distance-based notion of
semantic similarity
- Geometric betweenness
represents semantic betweenness Latent Space of InfoGAN
- Tends to be the case
- Smoothness assumption
- Interpolations in latent
space describe a meaningful morph
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 9
Our Proposal
- Use InfoGAN to learn dimensions of “difficult” domains
- For starters: shape domain
- Hyperparameters
- Number of latent variables
→ as few as possible, as much as necessary
- Type of latent variables
→ continuous (i.e., uniform or Gaussian)
- Make sure that we learn new dimensions
- Pre-select training data (if possible)
- Additional term in loss function: correlation to other dimensions
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 10
First Preliminary Results
- Data set of right-angled triangles, rectangles, and ellipses
- 2 continuous varibles (uniform distribution), 500 epochs
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 11
Towards A Thorough Implementation
Symbolic Layer Subsymbolic Layer Conceptual Layer
Manually define regions Manually define dimensions
clustering representation learning
Towards Grounding Conceptual Spaces in Neural Representations / Lucas Bechberger 12
Overall Envisioned System
apple banana color domain shape domain
ANN HSB
Symbolic layer Conceptual layer Subsymbolic layer