Learning with Memory Embeddings and its Application in the Digitalization of Healthcare

Keynote

Embedding learning, a.k.a. representation learning,  has been shown to be able to model large-scale semantic knowledge graphs. A  key concept is a mapping of the knowledge graph to  a tensor representation whose entries are predicted by models using latent representations of generalized entities.  Latent variable models are well suited to deal with the high dimensionality and sparsity of typical knowledge graphs and have successfully been employed in knowledge graph completion and fact extraction from the Web.  We have extended the approach to also consider temporal evolutions, temporal patterns and subsymbolic representations, which permits us to model medical decision processes. In addition, we consider embedding approaches to be a possible basis for modeling cognitive memory functions, in particular semantic and concept memory, episodic memory, sensory memory, short-term memory, and working memory.

Speakers: