Quantitative and Computational Systems Science Center
print

Links and Functions

Breadcrumb Navigation


Content

Learning with Memory Embeddings

Prof. Dr. Volker Tresp

Embedding learning, a.k.a. representation learning, has been shown to be able to model large-scale semantic knowledge graphs. A key concept is a mapping of the knowledge graph to a tensor representation whose entries are predicted by models using latent representations of generalized entities. Latent variable models are well suited to deal with the high dimensionality and sparsity of typical knowledge graphs and have successfully been employed in knowledge graph completion and fact extraction from the Web. We have extended the approach to also consider temporal evolutions, temporal patterns and subsymbolic representations, which permits us to model medical decision processes. In addition, we consider embedding approaches to be a possible basis for modeling cognitive memory functions, in particular semantic and concept memory, episodic memory, sensory memory, short-term memory, and working memory.

Back to QCSSC Lecture Series