Portrait of Dr. Henter Position: Assistant Professor in Intelligent Systems with spec. in Machine Learning
Division of Speech, Music and Hearing (TMH)
School of Electrical Engineering and Computer Science
KTH Royal Institute of Technology
E-mail: ghe at kth se
Address: KTH Royal Institute of Technology
Lindstedtsvägen 24, office 518
SE-100 44 Stockholm
SWEDEN

Research

My research interests are probabilistic modelling and machine learning for data generation tasks, most prominently statistical speech synthesis but also motion and gesture. My interests extend to the perceptual effects of modelling assumptions and design decisions, statistical robustness, and how to perform and analyse subjective tests.

Bio

I have a Ph.D. in electrical engineering (telecommunications) from KTH. Following graduation, I was a post-doc at the Centre for Speech Technology Research (CSTR) at the University of Edinburgh, UK, (2013–2016) then in Prof. Junichi Yamagishi's lab at the National Institute of Informatics, Tokyo, Japan, (2016–2018), and then at Speech, Music and Hearing (TMH) at KTH in 2018, prior to being promoted to my current position in January 2020.

Students

I am fortunate to be supervising the following doctoral students:

Publications

A list of my publications is provided on a separate page, grouped by type:

Selected presentation materials

  1. Practical classifier design
    A lecture from the M.Sc. course EQ2341 Pattern Recognition and Machine Learning at KTH Royal Institute of Technology, Stockholm, Sweden, most recently given in May 2019.
    [ slides ]
  2. Cyborgs and other controllable synthesisers: an update on past and planned research
    A presentation given at the Centre for Speech Technology Research, the University of Edinburgh, UK, in November 2018.
    [ slides w/o audio | abstract ]
  3. Wagging speech by the tail: The case for robust data generation
    A presentation most recently given at the ACCESS Linnaeus Centre, KTH Royal Institute of Technology, Stockholm, Sweden, in October 2018.
    [ slides w/ audio (2.6 MiB) | slides w/o audio | abstract ]
  4. Introduction to modern and controllable speech synthesis
    Excerpt from a presentation given at the annual EACare project retreat in Sigtuna, Sweden, in April 2018.
    [ slides w/ audio (6.6 MiB) | slides w/o audio ]
  5. Perceptual debugging of speech synthesis: Using speech from the future to pinpoint the design decisions that matter most
    A presentation most recently given at the Tokuda and Nankaku Laboratory, Nagoya Institute of Technology, Japan, in January 2018.
    [ slides w/ audio (7.6 MiB) | slides w/o audio | abstract ]

Note: Some audio files have been removed from the slides and presentations on this page due to strict Japanese regulations regarding personal data. Audio playback from pdf is known to work in Adobe Acrobat Reader (from trusted documents), but not in Preview on macOS.

Data releases

  1. Gustav Eje Henter, Thomas Merritt, Matt Shannon, Catherine Mayo, and Simon King (2014)
    Repeated Harvard Sentence Prompts (REHASP) corpus version 0.5
    Studio recordings of 30 Harvard sentences, each read aloud 40 times by a female native speaker of British English. Released under a CC BY 4.0 license.
    [ official repository | readme file | .bib | publication ]

Links and additional information

This page was last updated 2019-12-16.

[ back to the top of the page | contact the author ]