Portrait photo Position: Assistant Professor in Intelligent Systems with spec. in Machine Learning
Division of Speech, Music and Hearing (TMH)
School of Electrical Engineering and Computer Science
KTH Royal Institute of Technology
E-mail: ghe at kth se
Address: KTH Royal Institute of Technology
Lindstedtsvägen 24, office 518
SE-100 44 Stockholm
SWEDEN

Research

My research interests are probabilistic modelling and machine learning for data-generation tasks, most prominently speech synthesis as well as motion and gesture. Particular interests include deep generative modelling, the perceptual effects of modelling assumptions and design decisions, statistical robustness, and how to perform and analyse subjective evaluations.

Publications

A list of my publications is provided on a separate page, grouped by type:

Degrees and positions

Students

I am fortunate to be supervising the following doctoral students:

I previously supervised the following students:

Awards and recognition

Awards and recognition since my return to KTH in 2018:

Selected presentation materials

  1. Move over, MSE! – New probabilistic models of motion
    A presentation most recently given at the Department of Computer Science, University of California, Davis, USA, in April 2021.
    [ video presentation | slides w/o video (22.3 MiB) | abstract ]
  2. Practical classifier design
    A lecture from the M.Sc. course EQ2341 Pattern Recognition and Machine Learning at KTH Royal Institute of Technology, Stockholm, Sweden, most recently given in May 2019.
    [ slides ]
  3. Cyborgs and other controllable synthesisers: an update on past and planned research
    A presentation given at the Centre for Speech Technology Research, the University of Edinburgh, UK, in November 2018.
    [ slides w/o audio | abstract ]
  4. Wagging speech by the tail: The case for robust data generation
    A presentation most recently given at the ACCESS Linnaeus Centre, KTH Royal Institute of Technology, Stockholm, Sweden, in October 2018.
    [ slides w/ audio (2.6 MiB) | slides w/o audio | abstract ]
  5. Introduction to modern and controllable speech synthesis
    Excerpt from a presentation given at the annual EACare project retreat in Sigtuna, Sweden, in April 2018.
    [ slides w/ audio (6.6 MiB) | slides w/o audio ]
  6. Perceptual debugging of speech synthesis: Using speech from the future to pinpoint the design decisions that matter most
    A presentation most recently given at the Tokuda and Nankaku Laboratory, Nagoya Institute of Technology, Japan, in January 2018.
    [ slides w/ audio (7.6 MiB) | slides w/o audio | abstract ]

Note: Some audio files have been removed from the slides and presentations on this page due to strict Japanese regulations regarding personal data. Audio playback from pdf is known to work in Adobe Acrobat Reader (from trusted documents), but not in Preview on macOS.

Data releases

  1. Taras Kucherenko, Patrik Jonell, Youngwoo Yoon, Pieter Wolfert, and Gustav Eje Henter (2020)
    Data from the GENEA Challenge 2020
    Motion data, video stimuli, visualisation code, evaluation code, user responses, analysis code, and scientific papers from the first-ever challenge in automatic, speech-driven gesture generation.
    [ main page | workshop page | gesture data (direct link) | gesture data readme file | Zenodo repository | .bib | publication ]
  2. Gustav Eje Henter, Thomas Merritt, Matt Shannon, Catherine Mayo, and Simon King (2014)
    Repeated Harvard Sentence Prompts (REHASP) corpus version 0.5
    Studio recordings of 30 Harvard sentences, each read aloud 40 times by a female native speaker of British English. Released under a CC BY 4.0 license.
    [ official repository | readme file | .bib | publication ]

Events organised

Links and additional information

This page was last updated 2023-10-06.

[ back to the top of the page | contact the author ]