no code implementations • 25 Apr 2020 • Anthony Bagnall, Paul Southam, James Large, Richard Harvey
Given the massive volume of luggage that needs to be screened for this threat, the best way to automate the detection is to first filter whether a bag contains an electric device or not, and if it does, to identify the number of devices and their location.
no code implementations • 8 May 2018 • Helen L. Bear, Richard Harvey
Not only is this ambiguity damaging to the performance of audio-visual classifiers operating on real expressive speech, there is also considerable choice between possible mappings.
no code implementations • 8 May 2018 • Helen L. Bear, Richard Harvey
Visual lip gestures observed whilst lipreading have a few working definitions, the most common two are; `the visual equivalent of a phoneme' and `phonemes which are indistinguishable on the lips'.
no code implementations • 8 May 2018 • Kwanchiva Thangthai, Helen L. Bear, Richard Harvey
We compare the performance of a lipreading system by modeling visual speech using either 13 viseme or 38 phoneme units.
no code implementations • 3 Oct 2017 • Helen L. Bear, Richard Harvey, Barry-John Theobald, Yuxuan Lan
Visual-only speech recognition is dependent upon a number of factors that can be difficult to control, such as: lighting; identity; motion; emotion and expression.
no code implementations • 3 Oct 2017 • Helen L. Bear, Gari Owen, Richard Harvey, Barry-John Theobald
In the quest for greater computer lip-reading performance there are a number of tacit assumptions which are either present in the datasets (high resolution for example) or in the methods (recognition of spoken visual units called visemes for example).
no code implementations • 3 Oct 2017 • Helen L. Bear, Richard Harvey
To undertake machine lip-reading, we try to recognise speech from a visual signal.