Venue: Student Hall, Talbot Campus
Event Schedule is as follows:
Monday (9th March)
12:30 Registration
12:50 Opening address
13:00 – 14:30 BSL session – BSL teachers: Lynn Preston & Richard Neale
14:30 – 15:00 Break
15:00 – 16:30 BSL session – BSL teachers: Lynn Preston & Richard Neale
16:30 Closing
Tuesday (10th March)
9:30 Registration
9:50 Opening address
10:00 – 11:00 AI Translation between Sign and spoken language
Presenter: Richard Bowden is Professor of Computer Vision and Machine Learning within the Centre for Vision Speech and Signal Processing at the University of Surrey. He is also co-founder of Signapse. His research centres on the use of computer vision to locate, track, and understand humans with a specific interest and focus on sign languages. Having worked on many areas of computer vision, including robotics, autonomous vehicles and the tracking and analysis of humans. He has been developing computational approaches to sign language recognition and translation for over 30 years. Over the last few years Richard has refocused his entire research group on Sign Language Translation, with recent funding including a large grant from Google to reduce the cost of translation through AI and SignGPT, which aims to develop the next generation of AI tools for sign languages.
11:00 – 11:15 Break
11:15 – 12:00 A History of the Development and Access to British Sign Language
Presenter: George Raggett has been deaf from the age of eight through a yet undiagnosed hereditary condition. George acquired good speaking and reading skills prior to his hearing loss and prefer to communicate using these methods, supplemented by hearing aids and naturally developed lip-reading skills. George has developed Sign Language skills since 1987 and has been teaching Sign Language and Deaf Awareness & Communication courses to all ages since qualifying as a teacher in 2002. In 2007, George became a trustee of The Wiltshire and Dorset Deaf Association, a charity set up to promote and support the Deaf and Hard of Hearing Communities of both counties, in 2004
12:00 – 13:30 Break
13:30 – 14:15 Relevant Annotation for Multimodal Interaction Analysis: AI Bridging Interaction Studies of Gestures and Signs
Presenter: Mayuni Bono is an Associate Professor at the National Institute of Informatics (NII), Tokyo, Japan. She received her Ph.D. in Applied Linguistics from Kobe University in 2005. In her Ph.D. project, she demonstrated how to build a machine-readable model of the ‘Participation Framework,’ initially provided by Canadian-American sociologist Erving Goffman. After receiving her Ph.D., she got a position in informatics at ATR Media Information Science Laboratories, Kyoto University, and the National Institute of Informatics based on fruitful collaborations with engineering and informatics researchers. Currently, she is conducting several research projects of building sharable spoken and sign language multimodal corpora in an open science framework for all academic researchers interested in human communication using artificial intelligence techniques such as machine learning.
14:15 – 14:45 J-Shuwa: A Large-Scale Web-Collected Japanese Sign Language-Japanese Parallel Corpus
Presenter: Junwen Mo is a PhD candidate in the Nakayama Lab at the University of Tokyo. His research focuses on sign language understanding, with an emphasis on Japanese Sign Language translation. Recently, he has also developed an interest in alternative computational paradigms for artificial intelligence.
14:45 – Break
15:00 – 15:30 Between Representation and Embodiment: Perspectives on Sign Language Avatars
Presenter: Yi Wen is a first-year PhD researcher at Bournemouth University, where she previously completed her Master’s in Artificial Intelligence. She has a background in the game industry, working on large-scale production pipelines and collaborative projects. Her current research interests include sign language avatars, particularly exploring how different representational approaches may influence how signing is generated and experienced.
15:30 – 16:00 Continuous Sign language Recognition (CSLR): Feature Extraction from an Image Procesing Perspective
Presenter: Xinyu Zhang is an artificial intelligence researcher working at the intersection of medical imaging, representation learning, and algorithmic alignment. Her research advances contemporary medical AI by promoting approaches that reduce dependence on clinical expert teams and high computational resources, thereby enhancing scalability, reproducibility, and accessibility.
16:30 Closing











From Sustainable Research to Sustainable Research Lives: Reflections from the SPROUT Network Event
REF Code of Practice consultation is open!
BU Leads AI-Driven Work Package in EU Horizon SUSHEAS Project
ECR Funding Open Call: Research Culture & Community Grant – Apply now
ECR Funding Open Call: Research Culture & Community Grant – Application Deadline Friday 12 December
MSCA Postdoctoral Fellowships 2025 Call
ERC Advanced Grant 2025 Webinar
Update on UKRO services
European research project exploring use of ‘virtual twins’ to better manage metabolic associated fatty liver disease