supported by gold sponsor basic sponsor media partner
Transcript
supported by gold sponsor basic sponsor media partner
SUPPORTED BY GOLD SPONSOR BASIC SPONSOR MEDIA PARTNER 5th International Symposium Respeaking, Live Subtitling and Accessibility TABLE OF CONTENTS The Symposium 2 Programme6-7 Abstracts11 Speakers35 INDICE Il Simposio 4 Programma8-9 Abstracts11 Oratori35 1 5th International Symposium Respeaking, Live Subtitling and Accessibility 5th International Symposium Respeaking, Live Subtitling and Accessibility. 12 June 2015 UNINT - Università degli Studi Internazionali di Roma is pleased to host the 5th International Symposium on Respeaking, Live Subtitling and Accessibility. This is the fifth symposium in the series, after Forlì (2005), Barcelona (2007), Antwerp (2011) and Barcelona (2013). These events aim to bring together academic researchers, software developers, broadcasters, consumers and anyone interested in recent developments in media accessibility. In February 2013 the European Parliament Committee on Culture and Education published a Report on the transposition of Directive 2010/13/EU on Audiovisual Media Services in EU countries. The Report found that progress is still slow and uneven throughout the EU and that further developments are needed to ensure wider access to programmes and information for all citizens via audio description, audio/spoken subtitles and sign language. As regards live subtitling, over the past 10 years respeaking by means of speakerdependent speech recognition (SR) technology has become the preferred method to provide live subtitles on TV and in live events in many countries. At the same time, there have been significant improvements in speaker-independent speech recognition technology, with interesting applications in automatic subtitling and automatic speech reporting. Moreover, the deployment of the HbbTV (Hybrid Broadcast Broadband TV) standard in Europe is opening up more opportunities for the delivery of audiovisual content to consumers, ranging from traditional broadcast TV, video on-demand and catch-up TV services, like BBC iPlayer, as well as interactive services and Internet applications. Clearly, these developments are also creating more demand for accessibility services. The 5th Symposium will be an opportunity to look at areas of current research and future prospects in SR, live subtitling, and accessibility in general. 2 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Scientific Committee Carlo Eugeni (Roma Tre University) Pilar Orero (Universitat Autònoma de Barcelona) Aline Remael (University of Antwerp) Pablo Romero Fresco (University of Roehampton) Annalisa Sandrelli (UNINT University) Local Organising Committee Annalisa Sandrelli Serena Scaldaferri Claudio Russello Symposium website: http://www.unint.eu/it/component/content/article/8-pagina/494-respeaking-livesubtitling-and-accessibility.html Facebook page: https://www.facebook.com/respeakingsimposium?fref=ts 3 5th International Symposium Respeaking, Live Subtitling and Accessibility Quinto Simposio Internazionale Respeaking, Sottotitolazione in diretta e Accessibilità 12 giugno 2015 La UNINT - Università degli Studi Internazionali di Roma è lieta di ospitare il Quinto Simposio Internazionale su Respeaking, Sottotitolazione in diretta e Accessibilità. Si tratta della quinta edizione di un ciclo iniziato a Forlì (2005) e proseguito poi a Barcellona (2007), Anversa (2011) e ancora Barcellona (2013). Questi eventi costituiscono un punto di incontro per ricercatori, studiosi, aziende informatiche, broadcaster, utenti e chiunque si interessi di accessibilità e delle applicazioni delle tecnologie di riconoscimento del parlato. Nel febbraio 2013 la commissione Cultura e Istruzione del Parlamento europeo ha pubblicato una Relazione sulla trasposizione della Direttiva 2010/13/UE sui servizi audiovisivi. La Relazione ha evidenziato che I progressi sono ancora lenti e poco uniformi nei vari paesi dell’Unione e che sono necessari ulteriori sviluppi per assicurare un più ampio accesso all’intrattenimento e alle informazioni a tutti i cittadini, tramite audio descrizione, audiotitoli, sottotitoli e lingue dei segni. Per quanto riguarda la sottotitolazione in diretta, negli ultimi 10 anni il respeaking mediante la tecnologia di riconoscimento del parlato speaker-dependent è divenuto il metodo più comune in molti paesi per produrre sottotitoli in diretta per la TV e gli spettacoli dal vivo. Contemporaneamente, la tecnologia di riconoscimento del parlato speaker-independent ha fatto grandi passi avanti, con interessanti applicazioni alla sottotitolazione automatica e alla resocontazione automatica. Inoltre, l’avvento dello standard HbbTV (Hybrid Broadcast Broadband TV) nel sistema televisivo europeo sta aprendo nuove opportunità ai fornitori di contenuti audiovisivi, dalle trasmissioni televisive in modalità tradizionale al video on-demand e catch-up TV (come l’iPlayer della BBC), fino ad arrivare ai servizi interattivi e alle applicazioni Internet. Ovviamente tutti questi sviluppi contribuiscono ad accrescere anche la domanda di servizi di accessibilità per tutti i consumatori. Il Quinto Simposio sarà l’occasione per poter riflettere sulla ricerca attuale e sulle prospettive future delle tecnologie di riconoscimento del parlato, della sottotitolazione in diretta e dell’accessibilità in generale. 4 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Comitato scientifico Carlo Eugeni (Università Roma Tre) Pilar Orero (Università Autonoma di Barcellona) Aline Remael (Università di Anversa) Pablo Romero Fresco (Università di Roehampton) Annalisa Sandrelli (Università degli Studi Internazionali di Roma - UNINT) Comitato organizzatore locale Annalisa Sandrelli Serena Scaldaferri Claudio Russello Sito web del Simposio: http://www.unint.eu/it/calendario-eventi/-22.html Pagina Facebook: https://www.facebook.com/respeakingsimposium?fref=ts 5 5th International Symposium Respeaking, Live Subtitling and Accessibility PROGRAMME Aula Magna Morning session 6 08.30 - 10.00 Registration 09.00 - 09.15 Welcome and conference opening 09.15 - 09.30 Roxana Widmer-Iliescu, Unione Internazionale delle Telecomunicazioni “ITU work on ICT Accessibility” Theme 1. Moderator: Live subtitling on television: focus on quality Gion Linder, Swiss Text 09.30 - 09.50 Pablo Romero Fresco, “Live subtitling in the UK: an official review” 09.50 - 10.10 Aleš Pražák, “Live TV subtitling in Czech” 10.10 - 10.30 Zoe Moores, “The language of respeaking” 10.30 - 10.50 Juan Martínez Pérez, “New Approaches to Improve the Quality of Live Subtitling on TV” 10.50 - 11.00 Q & A Session 11.00 - 11.30 Coffee break Theme 2. Moderator: Respeaking: Experimental Studies Pablo Romero Fresco, University of Roehampton 11.30 - 11.50 Katarzyna Stachowiak, “More than meets the ear: Do subtitles matter for the well-educated, speaking hearing impaired?” 11.50 - 12.10 Agnieszka Szarkowska, Łukasz Dutka, “Are interpreters better respeakers? An exploratory study on respeaking competences” 12.10 - 12.30 Agnieszka Chmiel, Agnieszka Lijewska, Łukasz Dutka, Agnieszka Szarkowska, Isabela Krejtz, Krzysztof Krejtz, “Tapping the linguistic competence in the respeaking process: comparing intralingual paraphrasing done by interpreting trainees, translation trainees and bilinguals” 12.30 - 12.40 Q & A Session 13.00 - 14.30 Lunch break 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Afternoon session Theme 3. Moderator: New applications of technology Pilar Orero, Universitat Autònoma de Barcelona 14.30 - 14.50 Andrew Lambourne, “Pushing forward the frontiers of access technology” 14.50 - 15.10 Carlo Aliprandi, “Evaluating performances of automatic speech recognition for assisted subtitling” 15.10 - 15.30 Andreas Søeborg Kirkedal, Julián Zapata, “Translating and respeaking in a native or foreign language and from different environments: a prepilot experiment with a laptop and a tablet” 15.30 - 15.50 Lukasz Daniluk, Anna Matamala, Pablo Romero-Fresco, “Transcribing documentaries: can respeaking be used efficiently?” 15.50 - 16.00 Q & A Session Theme 4. Moderator: Respeaking as a professional opportunity for the blind Federico Spoletti, Sub-Ti & FRED.fm 16.00 - 16.20 Carlo Eugeni “Teaching live subtitling for the deaf through respeaking - the potential for blind respeakers and paraplegic live editors” 16.20 - 16.40 Robin Ribback, “Vocational training for blind speech-to-text-reporters” 16.40 - 16.50 Q & A Session 16.50 - 17.00 Coffee break ROUND TABLE: Focus on Italy Moderator: Carlo Eugeni, Università Roma Tre 17.00 - 18.30 Giacomo Pirelli, Luca Iacovino, “Proposta di aggiornamento delle legislazioni italiane relative all’accessibilità dei programmi delle emittenti televisive per i cittadini sordi” Samuel Zambelli, “La Meeting Industry in Italia oggi può offrire attraverso il respeaking nuove opportunità sia per gli organizzatori di congressi ed eventi che per i professionisti della documentazione scritta?” Lorenzo Di Ciaccio, “La telefonata sottotitolata” Saveria Arma, “New frontiers of respeaking: live subtitles delivered on smartphones, tablets and electronic glasses” 18.30 Closing 7 5th International Symposium Respeaking, Live Subtitling and Accessibility PROGRAMMA Aula Magna Mattina 08.30 - 10.00 Registrazione partecipanti 09.00 - 09.15 Saluti e apertura del Simposio 09.15 - 09.30 Roxana Widmer-Iliescu, International Telecommunications Union “ITU work on ICT Accessibility” Tema 1. La sottotitolazione in diretta in televisione: la qualità Moderatore: Gion Linder, Swiss Text 09.30 - 09.50 Pablo Romero Fresco, “Live subtitling in the UK: an official review” 09.50 - 10.10 Aleš Pražák, “Live TV subtitling in Czech” 10.10 - 10.30 Zoe Moores, “The language of respeaking” 10.30 - 10.50 Juan Martínez Pérez, “New Approaches to Improve the Quality of Live Subtitling on TV” 10.50 - 11.00 Dibattito 11.00 - 11.30 Pausa caffè Tema 2. Respeaking: studi sperimentali Moderatore: Pablo Romero Fresco, University of Roehampton 11.30 - 11.50 Katarzyna Stachowiak, “More than meets the ear: Do subtitles matter for the well-educated, speaking hearing impaired?” 11.50 - 12.10 Agnieszka Szarkowska, Łukasz Dutka, “Are interpreters better respeakers? An exploratory study on respeaking competences” 12.10 - 12.30 Agnieszka Chmiel, Agnieszka Lijewska, Łukasz Dutka, Agnieszka Szarkowska, Isabela Krejtz, Krzysztof Krejtz, “Tapping the linguistic competence in the respeaking process: comparing intralingual paraphrasing done by interpreting trainees, translation trainees and bilinguals” 12.30 - 12.40 Dibattito 13.00 - 14.30 Pausa pranzo 8 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Pomeriggio Tema 3. Moderatore: Nuove applicazioni della tecnologia Pilar Orero, Universitat Autònoma de Barcelona 14.30 - 14.50 Andrew Lambourne, “Pushing forward the frontiers of access technology” 14.50 - 15.10 Carlo Aliprandi, “Evaluating performances of automatic speech recognition for assisted subtitling” 15.10 - 15.30 Andreas Søeborg Kirkedal, Julián Zapata, “Translating and respeaking in a native or foreign language and from different environments: a pre-pilot experiment with a laptop and a tablet” 15.30 - 15.50 Lukasz Daniluk, Anna Matamala, Pablo Romero-Fresco, “Transcribing documentaries: can respeaking be used efficiently?” 15.50 - 16.00 Dibattito Tema 4. Moderatore: Il respeaking come sbocco professionale per i non vedenti Federico Spoletti, Sub-Ti & FRED.fm 16.00 - 16.20 Carlo Eugeni, “Teaching live subtitling for the deaf through respeaking- the potential for blind respeakers and paraplegic live editors” 16.20 - 16.40 Robin Ribback, “Vocational training for blind speech-to-text-reporters” 16.40 - 16.50 Dibattito 16.50 - 17.00 Pausa caffè TAVOLA ROTONDA: Uno sguardo sull’Italia Moderatore: Carlo Eugeni, Università Roma Tre 17.00 - 18.30 Giacomo Pirelli, Luca Iacovino, “Proposta di aggiornamento delle legislazioni italiane relative all’accessibilità dei programmi delle emittenti televisive per i cittadini sordi” Samuel Zambelli, “La Meeting Industry in Italia oggi può offrire attraverso il respeaking nuove opportunità sia per gli organizzatori di congressi ed eventi che per i professionisti della documentazione scritta?” Lorenzo Di Ciaccio, “La telefonata sottotitolata” Saveria Arma, “New frontiers of respeaking: live subtitles delivered on smartphones, tablets and electronic glasses” 18.30 Chiusura Simposio 9 5th International Symposium Respeaking, Live Subtitling and Accessibility Abstracts 10 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility SYMPOSIUM OPENING ITU work on ICT Accessibility Roxana Widmer-Iliescu - ITU Abstract: The presentation will share the work of the International Telecommunication Union (ITU) in promoting ICT accessibility. ITU is the leading United Nations agency for Information and Communication Technologies (ICTs) which includes 193 Member States. The presentation will highlight ITU’s work in promoting ICT accessibility in its technical standards, its policy and regulatory advice to Member States, capacity building initiatives, such as developing accessible audio visual services, and will introduce ITU’s own accessibility policy which covers accessible meeting services such as captioning for the deaf and hard of hearing. ITU adopted the first international technical accessibility standard in 1994 and developed the Telecommunications Accessibility Checklist (2006) to assist experts developing technical standards to ensure that they take ICT accessibility for persons with disabilities into account when developing technical standards. ITU Member States adopted the Connect 2020 targets which call for all countries to have ICT accessibility policies in place by 2020. ITU is supporting its Member States to achieve this goal through the work of its Study Groups and Regional initiatives including the European Regional Initiative, Ensuring access to telecommunications/ICTs, in particular for persons with disabilities. Participants will learn about the range of ITU resources to promote ICT accessibility including its reports Making Mobile Phones and Services Accessible, Making Television Accessible and the ITU Model ICT Accessibility Policy Report which regulators and policy makers can use in developing their own national ICT accessibility policies. These policies cover public access points, television/video programming, mobile phones, web sites and public procurement. Multi-stakeholder cooperation, including operators and organizations of persons with disabilities, is essential for the successful definition and implementation of ICT accessibility policies. References ITU http://www.itu.int/en/Pages/default.aspx; www.itu150.org FSTP-TACL Telecommunications Accessibility Checklist (2006): http://www.itu.int/pub/T-TUT-FSTP-2006-TACL/en European Regional Initiative, Ensuring access to telecommunications/ICTs, in particular for persons with disabilities. http://www.itu.int/en/ITU-D/Regional-Presence/Europe/Pages/ActivitiesandProjects.aspx ITU Report (2012) Making Mobile Phones and Services Accessible for Persons with Disabilities. http://www. itu.int/ITU-D/sis/PwDs/Documents/Mobile_Report.pdf ITU Report (2011) Making Television Accessible http://www.itu.int/ITU-D/sis/PwDs/Documents/Making_ TV_Accessible-FINAL-WithAltTextInserted.pdf ITU Report (2014) Model ICT Accessibility Policy Report http://www.itu.int/en/ITU-D/Digital-Inclusion/Persons-with-Disabilities/Documents/ICT%20Accessibility%20Policy%20Report.pdf Roxana Widmer-Iliescu, [email protected] Senior Programme Officer and BDT( ITU Development Bureau) Focal Point for Study Group on Question 7 “Access to telecommunication /ICT services by Persons with Disabilities and with specific needs” International Telecommunication Union - Switzerland 11 5th International Symposium Respeaking, Live Subtitling and Accessibility THEME 1 LIVE SUBTITLING ON TELEVISION: FOCUS ON QUALITY Live Subtitling in the UK: An Official Review Pablo Romero-Fresco - University of Roehampton Abstract: In May 2013, Ofcom, the governmental communications regulator in the UK, published a consultation document in order to gather views from broadcasters, subtitle providers and user associations as to how to improve the quality of live captioning on UK TV to benefit deaf and hard-of-hearing viewers. Following this consultation, Ofcom decided that broadcasters should be required to assess the speed, reduction rate, latency and accuracy of their live captions using the NER model (Romero-Fresco and Martínez 2015). Ofcom also decided that the University of Roehampton (London) would review the measurements provided by broadcasters from a third party standpoint in four stages over two years between 2013 and 2015. So far, three of the four measurements have been conducted, including the analysis of 216 programmes belonging to three genres (news, entertainment and chat shows) and broadcast by the BBC, ITV, Channel 4, Five and Sky. The review has covered 365.000 words, 55.000 subtitles and 50 hours of TV material. The aim of this presentation is to discuss the results of this review, focusing on the different issues involved in the quality of live subtitles, such as accuracy (including accuracy rates, types of errors and reasons why they are occurring), delay, speed and format. It will also deal with the challenges posed by the hybrid mode (the combination of live subtitles and prerecorded subtitles cued live) introduced in the UK as a result of the Ofcom review in order to improve the quality of subtitling. Initially welcome as a means to introduce 100% accurate and perfectly synchronised subtitles in live programmes, this hybrid mode is also bringing about subtitles with speeds exceeding 400wpm and even 500wpm regularly. Despite not being legible for the viewers, these subtitles contribute to increasing the overall accuracy rate and to decreasing the average delay of the broadcasters’ output. This presentation will offer some suggestions as to how this issue may be tackled by the subtitling companies and how it can be factored in the accuracy analysis using the NER model. References Romero-Fresco, Pablo & Martínez Pérez, Juan (2015) “Accuracy in Live Subtitling- The NER model”, in Jorge Díaz Cintas & Rocío Baños (eds) Audiovisual Translation in a Global Context: Mapping an Ever-changing Landscape, London: Palgrave. Pablo Romero-Fresco, [email protected] Reader in Audiovisual Translation and Filmmaking University of Roehampton - Department of Media, Culture and Language - United Kingdom 12 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Live TV subtitling in Czech Aleš Pražák - SpeechTech, s.r.o. Abstract: We would like to introduce our solution for live TV subtitling that we provide as a service for the Czech Television, the public service broadcaster in the Czech Republic. We operate a fully automated subtitling system (without a re-speaker) for the subtitling of meetings of the Parliament of the Czech Republic directly from the real acoustic track, using specifically designed speaker-independent speech recognition. At the same time, since we develop both the speech recognition and live subtitling technologies, we implemented a novel approach to live subtitling through re-speaking. We use only one re-speaker doing all the tasks with very close interaction with the recognition system, i.e. instant corrections, speaker colouring, online word additions, etc. Moreover, our subtitling platform is completely distributed – our re-speakers operate at home with bidirectional connection to the television company, so the subtitle latency is significantly reduced. We also provide supplementary services, such as language model adaptation for sport programmes, an automatic daily updating of speech recognizer vocabulary for news subtitling, correction and automatic timing of subtitles for programme rerun or re-speaker training using special four-phase training application. We have several years of experience in live TV subtitling, including subtitling of sport events such as Winter Olympic Games or FIFA World Cup. Now we have 8 skilled re-speakers who subtitled almost 2.000 hours in 2014. References Pražák, A., Loose, Z., Trmal, J., Psutka, J. V. & Psutka, J. (2012) “Novel Approach to Live Captioning Through Re-speaking: Tailoring Speech Recognition to Re-speaker’s Needs”. INTERSPEECH, ISCA. Trmal, J.; Pražák, A.; Loose, Z. & Psutka, J. (2010) “Online TV Captioning of Czech Parliamentary Sessions”. In: Petr Sojka, Ales Horák, Ivan Kopecek & Karel Pala (eds) LNCS (LNAI), vol. 6231, pp. 416422, Springer, Heidelberg. Aleš Pražák, [email protected] Project manager SpeechTech, s.r.o. - Czech Republic 13 5th International Symposium Respeaking, Live Subtitling and Accessibility The Language of Respeaking Zoe Moores - University of Roehampton Abstract: Subtitle providers in the UK are under particular scrutiny from the communications regulator Ofcom over the accuracy of their live subtitling. Since the very nature of respeaking means it is done in an intense, high-pressure environment, any guidance available on how to achieve more accurate subtitles will benefit trainers, individual subtitlers and deaf and hard of hearing viewers, for whom the degree of accessibility provided depends on the quality of the subtitles they receive. In this paper I present the results of a study into the errors that occur whilst respeaking in English and their causes. The corpus consists of 47 weather forecasts, sampled from the major television channels and totalling almost an hour of respeaking. The style and speed of this programme, along with the variable content that appears within the chatty asides, makes it particularly challenging for the respeaker and a relevant focus for a study on accuracy. In the first half of the corpus, I respoke each clip into a new voice model and replicated key features of in-house subtitling software. Errors were then analysed, potential solutions tested and the clip respoken for a second time. From this detailed analysis, it was possible to create a taxonomy of the English used in respeaking, which revealed the lexical categories errors fell into, the parts of speech that were most affected and the cause of each error. The key features of a respeaking style which maximizes accuracy and reduces the likelihood of errors were revealed. In the second half, consideration was given to whether these findings could be applied to a range of respeakers working with professional software and to the balance between the degree of editing and the impact on the content that the SDH viewer accesses. This presentation concludes with some recommendations based on the findings obtained. References Greenberg, Steven & Fosler-Lussier, Eric (2000) “The Uninvited Guest: Information’s Role in Guiding the Production of Spontaneous Speech.” In: Proceedings Of The CREST Workshop On Models Of Speech Production: Motor Planning And Articulatory Modelling. Kloster Seeon, Germany, May 1-4. Moores, Zoe (2014) Respeaking The Weather Accurately - Why The Nation’s Favourite Topic is a Respeaker’s Nightmare. UK, London: University of Roehampton (Unpublished dissertation). Roach, Peter (2009) English Phonetics and Phonology A Practical Course Fourth Edition. UK: Cambridge University Press. Romero-Fresco, Pablo (2009) “More Haste Less Speed: Edited Versus Verbatim Respoken Subtitles.” Vigo International Journal of Applied Linguistics, vol.6, pp. 109-133. Romero-Fresco, Pablo & Martínez, Juan (2014). Accuracy Rate in Live Subtitling - The NER Model. https://roehampton.openrepository.com/roehampton/bitstream/10142/141892/1/NER-English.pdf Shockey, Linda (2003) Sound Patterns Of Spoken English. UK: Blackwell Publishing. Zoe Moores, [email protected] - Graduate Student University of Roehampton - Department of Media, Culture and Language - United Kingdom 14 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility New Approaches to Improve the Quality of Live Subtitling on TV Juan Martínez Pérez - TransMedia Catalonia, UAB, CaiaC Abstract: The aim of this presentation is twofold. First, to define certain quality-related aspects of live subtitling on TV and how to benchmark them, and second to describe the NERstar Editor. This tool is an advanced application for quality measurement and data generation, which has been developed by the author on behalf of SWISS TXT in close collaboration with the German company VerbaVoice. Regarding the first point, two questions arise: what is meant by “quality-related aspects of live subtitling on TV” and how to define quality benchmarking. Attention should also be paid to the different preferences and expectations of the countries with respect to quality-related criteria in live subtitling on TV, and the many differences in legislation, funding and production models. Language always works within a culture, hence distinct cultural background will also have an impact in quality benchmarking. In other words: quality is a highly complex interplay of very contrasting ideas and often controversial issues. As to the second point, the aim is to present NERstar Editor, a software solution that has specifically been developed for assessing and improving the quality in live subtitling on TV linked to accuracy, delay, speed and display issues. References Romero-Fresco, Pablo & Martínez, Juan (2014) Accuracy Rate in Live Subtitling - The NER Model. https://roehampton.openrepository.com/roehampton/bitstream/10142/141892/1/NER-English.pdf Juan Martínez Pérez, [email protected] TransMedia Catalonia, CaiaC, UAB, Barcelona - Spain 15 5th International Symposium Respeaking, Live Subtitling and Accessibility THEME 2 RESPEAKING: EXPERIMENTAL STUDIES More than meets the ear: Do subtitles matter for the well-educated, speaking hearing-impaired? Katarzyna Stachowiak - Adam Mickiewicz University Abstract: The common belief is that the main target group of interlingual subtiling or closed captions for the deaf and hard of hearing are individuals whose command of spoken language is poor or who do not have that command (Neves 2008). On the other hand, well-educated patients native in a spoken language or with excellent command of it may benefit greatly from subtitles. The aim of the study was to verify to what extent patients with hearing aids and cochlear implants benefit from subtitled media content, and more specifically: live debates. There were three groups of participants with hearing loss: the ones with mild to moderate (up to 55 dB), moderately severe to severe (up to 90 dB) and profound (> 90 dB) hearing loss. Additionally, there were groups of participants unilaterally and bilaterally implanted (implanted for 2 years or more). Particular groups included 40 participants, resulting in the total number of participants equal to 200. The subjects were aged 2443 and had higher education. There were 42-66% of females in the groups. The level of understanding speech was above 80% in silence and above 50% in noise in all the participants. The subjects’ native language and the language of daily use was spoken Polish. The participants were sent two 5-minute films (live debates engaging 3 speakers), one subtitled in Polish. The participants were asked to fill in a questionnaire composed of 20 questions assessing the level of understanding the content with or without subtitling and the satisfaction from the subtitles. The results revealed a significant need for and satisfaction from subtitling, and no correlation with age and the level of understanding speech in silence, nor noise, in all the participants. The fact of having a hearing loss or deafness correlated positively with the level of satisfaction, excluding subjects with mild to moderate hearing impairment. The results shed new light on the debate about potential beneficiaries of subtitling and the need to have media content subtitled, as well as the needs of audience with hearing loss or deafness. References Neves Josélia (2008) “10 fallacies about Subtitling for the d/Deaf and the hard of hearing”, The Journal of Specialised Translation, vol. 10: 128. Katarzyna Stachowiak, [email protected] Adam Mickiewicz University in Poznań Department of Psycholinguistic Studies, Faculty of English - Poland 16 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Are interpreters better respeakers? An exploratory study on respeaking competences Agnieszka Szarkowska - University of Warsaw Łukasz Dutka - University of Warsaw Abstract: In this paper we present preliminary data from the research project “Respeaking competences, process, quality”. In the course of respeaking, a number of complex cognitive processes take place, requiring the use of multiple cognitive resources. The main goal of the project is to study these processes and to obtain insight into their cognitive intricacies. In the project, three groups of participants (trainee interpreters, trainee translators and bilinguals with no background in interpreting or translation) are subjected to a number of psychophysiological tests: working memory capacity tests (see Baddeley 1986), paraphrasing tests (see Christoffels and de Groot 2004) and respeaking tests. During the respeaking tests, participants’ brain activity is monitored by EEG (Emotiv), their eye movements are recorded by an eye tracker, while their computer activity is examined through screen recording software. These tools serve to understand the respeaking process by analysing cognitive resources used by participants while respeaking. EEG enables us to verify the level of the engagement of different parts of the brain during respeaking and the lateralisation of brain functions (Fabbro et al. 1990). Eye tracking helps identify prominent areas which attract the respeaker’s visual attention. The respoken subtitles created by the study participants undergo a qualitative and quantitative analysis using the NERstar software (Martínez Pérez 2013). We hypothesise that interpreting trainees - as opposed to trainee translators - are better predisposed to respeaking owing to a number of similarities between interpreting and respeaking. References Baddeley, Alan D. (1986) Working Memory. Oxford: Oxford University Press. Christoffels, Ingrid K., Annette M. B. De Groot, Judith F. Kroll (2006) “Memory and language skills in simultaneous interpreters: The role of expertise and language proficiency”. Journal of Memory and Language 54 (3), 324-345. Martínez Pérez, Juan (2013) “NERStar. Speech technology accuracy rate”. Paper presented at the Media for All 5 conference in Dubrovnik, September 2013. Romero-Fresco, Pablo (2011) Subtitling through Speech Recognition: Respeaking. Manchester: St Jerome. Romero-Fresco, Pablo (2012) “Respeaking in Translator Training Curricula. Present and Future Prospects”. The Interpreter and Translator Trainer, 6(1), 91-112. Agnieszka Szarkowska, [email protected] Assistant Professor - University of Warsaw - Institute of Applied Linguistics - Poland Łukasz Dutka, [email protected] - University of Warsaw Institute of Applied Linguistics - Poland 17 5th International Symposium Respeaking, Live Subtitling and Accessibility Tapping the linguistic competence in the respeaking process: comparing intralingual paraphrasing done by interpreting trainees, translation trainees and bilinguals Agnieszka Chmiel, Agnieszka Lijewska - Adam Mickiewicz University Łukasz Dutka, Agnieszka Szarkowska - University of Warsaw Izabela Krejtz, Krzysztof Krejtz - University of Social Sciences and Humanities Abstract: Respeakers need to master a number of linguistic and technical competences (Arumí Ribas & Romero Fresco 2008, Romero Fresco 2012, Eugeni 2008). Among them is the ability to listen and speak at the same time in the same language or to translate the words spoken in a live programme into another language. Other necessary linguistic skills are the ability to paraphrase and condense the text (Luycks et al. 2010, Sandrelli 2013) as well as the ability to simultaneously listen to other speakers and to one’s own voice in order to control the intonation and correctness of the respoken text (Romero Fresco 2011). Thus, respeaking involves a re-expression skill similar to the one applied in simultaneous interpreting (Romero-Fresco 2011: 45). We have conducted a study to test whether due to their interpreting skills, interpreting trainees would perform significantly better in a paraphrasing task understood as a subcomponent of respeaking. 15 interpreting trainees, 15 translation trainees and 6 non-interpreting and non-translating bilinguals participated in a two-day intensive respeaking course and then in an intralingual paraphrasing study. They were asked to paraphrase three types of sentences. Type 1 included semantic redundancies (which could be replaced by hyperonyms), type 2 included oral discourse markers (which could be omitted without sense loss) and type 3 included false starters (which could be normalized). The participants performed the paraphrasing in a delayed condition (respeaking a given sentence after they have finished listening to it) or in a simultaneous condition (respeaking a given sentence while listening to it). In the analysis we looked at group differences in the ear-voice span and correlated the quantitative data with the quality of the paraphrases and the participants’ memory scores. We hypothesize that interpreting trainees – as opposed to trainee translators and bilinguals – will be better predisposed to respeaking thanks a number of similarities between interpreting and respeaking, and also thanks to their well-trained working memory. 18 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility References Arumí Ribas, Marta & Romero-Fresco, Pablo (2008) “A Practical Proposal for the Training of Respeakers”. The Journal of Specialised Translation 10, pp. 106-127. Eugeni, Carlo (2008) Le sous-titrage en direct: aspects théoretiques, professionels et didactiques. Macerata: Eum. Luyckx, Bieke, Delbeke, Tijs, Van Waes, Luuk, Leijten, Marielle & Remael, Aline (2010) “Live subtitling with speech recognition. Causes and consequences of text reduction”. Romero-Fresco, Pablo (2011) Subtitling through Speech Recognition: Respeaking. Manchester: St Jerome. Romero-Fresco, Pablo (2012) “Respeaking in Translator Training Curricula. Present and Future Prospects”. The Interpreter and Translator Trainer 6(1), pp. 91-112. Romero-Fresco, Pablo (2013) “Quality and NER in the UK”. Paper presented at the 4th International Symposium on Live Subtitling in Barcelona, March 2013. Sandrelli, Annalisa (2013) “Reduction strategies and accuracy rate in live subtitling of weather forecasts: a case study”. Paper presented at the 4th International Symposium on Live Subtitling in Barcelona, March 2013. Agnieszka Chmiel, [email protected] Assistant Professor Adam Mickiewicz University - Faculty of English - Department of Translation Studies - Poland Agnieszka Lijewska, [email protected] Assistant Professor Adam Mickiewicz University - Faculty of English - Department of Psycholinguistic Studies - Poland Łukasz Dutka, [email protected] Consultant University of Warsaw - Institute of Applied Linguistics - Poland Agnieszka Szarkowska, [email protected] Assistant Professor University of Warsaw - Institute of Applied Linguistics - Poland Izabela Krejtz, [email protected] Assistant Professor University of Social Sciences and Humanities Interdisciplinary Center for Applied Cognitive Studies - Poland Krzysztof Krejtz, [email protected] Assistant Professor University of Social Sciences and Humanities - Poland 19 5th International Symposium Respeaking, Live Subtitling and Accessibility THEME 3 NEW APPLICATIONS OF TECHNOLOGY Pushing forward the frontiers of access technology Andrew Lambourne - Screen Systems Abstract: Tools and techniques for production of access subtitles have been in a continuous state of development since the early days of such services in the 1980s. Respeaking is the latest in a series of techniques which have been used to solve the difficult problem of providing subtitles for real-time broadcasts: other methods have included Stenographic transcription systems, ordinary keyboards, multiplexed dual keyboards, syllabic chord keyboards and automatic speech recognition. Fundamentally all these systems face the same challenge - how to transcribe and present real-time unscripted speech with minimum delay and maximum quality? This paper describes new tools and techniques targeted at reducing the toughest challenges faced by a real-time subtitler – be they a trained respeaker or someone using any other real-time transcription approach – and explores not only how these challenges can be minimised, but what further practical research could beneficially be conducted. Missing vocabulary, homophones, out-of-language words, number grammars, corrections, and throughput delay are all examples of problems which can lead to a reduction in quality. We examine practical ways in which these problems can be reduced, and suggest further research directions. Additionally, we look at two opportunities relating to IP delivery. Firstly, the HbbTV system, which permits traditional broadcast material to be supplemented by IP-delivered content. One application for this is in personalisation, and new proposals for delivering audio could open up possibilities for assistance to hard-of-hearing viewers by enabling them to select and personalise the mix of audio they receive. This would not only help to solve the problem of intrusive music beneath dialogue, but would actually assist in the production of access subtitles by making it easier to transcribe programmes for which no script was available. Secondly, we look at new tools which have been developed to enable live webcasts to carry high quality properly synchronised live captions. Andrew Lambourne, [email protected] Business Development Director - Screen System - United Kingdom 20 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Evaluating performances of Automatic Speech Recognition for Assisted Subtitling Carlo Aliprandi - Synthema Srl Abstract: The contribution will address recent advances in Automatic Speech Recognition (ASR), presenting and evaluating a new Speaker Independent ASR technology that has been developed in the SAVAS project. The demand for Access Services has quickly grown in the years, mainly due to National and International laws. This trend is expected to consolidate for subtitling in particular. However, the path of manual subtitling is no longer feasible, due to increased costs and reduced production times. In this contribution we will focus on an emerging technique, Assisted Subtitling. Assisted Subtitling consists on the application of Automatic Speech Recognition (ASR) to generate transcripts of programs and to use the transcripts as the basis for subtitles. We will present recent advances, based on a new ASR technology that has been that has been developed in the SAVAS project. We will describe three systems based on SAVAS: • S .Scribe! is the system for pre-recorded content, capable of automatically transcribing audio and video content into time-aligned subtitles. • S .Align! is the system for automatic synchronisation, capable of auto-timing audio and video content to script and subtitles. • S .Live! is the first-of-a-kind Speaker Independent Transcription System, with realtime performances for online subtitling. Eventually, we will give an overview of the methodology, quality features and metrics employed to evaluate the SAVAS technology. The considered features include parameters linked to subtitle layout, duration and text editing such as speaker colours, delay, persistence on screen, capitalization, punctuation, splitting or the use of acronyms, apostrophes and numerals. An extended version of the NER model that also considers errors related to the automatic allocation of speaker colours, time stamps and segmentations will be described. The quality performance achieved by the SAVAS applications will be presented together with results of experiment testing the productivity gains achieved in post-editing automatic subtitles. Finally, the main challenges that still remain to be solved will be highlighted. 21 5th International Symposium Respeaking, Live Subtitling and Accessibility References Aliprandi, Carlo & Verruso, Fabrizio (2006), “Tecnologie del Linguaggio Naturale e sottotitolazione multilingue diretta”, inTRAlinea, Special Issue: Respeaking, ISSN 1827-000X, 2006 http://www.intralinea. it/specials/respeaking/ita_more.php?id=453_0_41_0_M Aliprandi, Carlo (2013) “Speech Recognition with a focus for quality Subtitling of minor languages”, 4th Symposium on Live Subtitling, March 2013, Barcelona. Aliprandi, Carlo (2013) “The Savas project – Speech Recognition for Respeaking and Live Subtitling”. In Carlo Eugeni & Luigi Zambelli (a cura di) Respeaking. On-line. Special Issue n. 1. June 2013. www. accademia-aliprandi.it pp. 117-119. Álvarez, Aitor, Aliprandi, Carlo, Gallucci, Isabella, Piccinini, Nicola, Raffaelli, Matteo, del Pozo, Arantza, Cassaca, Renato, Neto, Joao, Mendes, Carlos, Viveiros, Marcio forthcoming, “Automating Live and Batch Subtitling of Multimedia Contents for several European Languages”, Multimedia Tools and Applications (MTAP), An International Journal from Springer. Álvarez, Aitor, Aliprandi, Carlo, Gallucci, Isabella, Piccinini, Nicola, Raffaelli, Matteo, del Pozo, Arantza, Cassaca, Renato, Neto, Joao, Mendes, Carlos, Viveiros, Marcio (2014) “Assisted Subtitling: a new opportunity for Access Services”, Paper presented at the International Broadcasting Conference 2014 (IBC2014), September 10-15 2014, Amsterdam, The Netherlands. Arantza del Pozo, Carlo Aliprandi, Aitor Álvarez, Carlos Mendes, Joao Neto, Sérgio Paulo, Nicola Piccinini, Matteo Raffaelli (2014) “SAVAS: Collecting, Annotating and Sharing Audiovisual Language Resources for Automatic Subtitling”, International Conference on Language Resources and Evaluation (LREC2014), May 2014, Reykjavik. C. Aliprandi, I. Gallucci, N. Piccinini, M. Raffaelli, A. del Pozo, A. Álvarez, R. Cassaca, J. Neto, C. Mendes, M. Viveiros, Carlo Aliprandi, Cristina Scudellari, Isabella Gallucci, Nicola Piccinini, Matteo Raffaelli, Arantza del Pozo, Aitor Álvarez, Haritz Arzelus, Renato Cassaca, Tiago Luis, Joao Neto, Carlos Mendes, Sérgio Paulo, Marcio Viveiros (2104) “Automatic Live Subtitling: state of the art, expectations and current trends”, NAB Broadcast Engineering Conference, April 5-10 2014, Las Vegas. Eugeni, Carlo (2006). “Introduzione al rispeakeraggio televisivo”, inTRAlinea, Special Issue: Respeaking, ISSN 1827-000X, 2006, URL: http://www.intralinea.it/specials/respeaking/ita_more. php?id=444_0_41_0_M Lambourne, Andrew (2006) “Subtitle respeaking: A new skill for a new age”, inTRAlinea, Special Issue: Respeaking, ISSN 1827-000X, 2006, URL: http://www.intralinea.it/specials/respeaking/eng_more. php?id=447_0_41_0. Martínez, Juan (2103) “Tools for assessing Accuracy Rate of Live Subtitles”, 4th Symposium on Live Subtitling, March 2013, Barcelona. Carlo Aliprandi, [email protected] Speech Solutions Manager - SyNTHEMA Srl - Italy 22 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Beyond PC-based speaker-adapted speech recognition: Exploring the potential and challenges of cloud-based speaker-independent speech recognition systems Andreas Søeborg Kirkedal - Copenhagen Business School Julián Zapata - University of Ottawa Abstract: Automatic speech recognition (ASR) technology gives a computer system the ability to recognize and process human speech. Developed and used for several decades, speech-enabled systems and devices are proven better than keyboard-based interfaces for tasks for which full natural language communication is useful, such as medical dictation, translation dictation, respeaking and (live) subtitling. Our presentation will describe a pilot experiment aiming to assess and compare the performance of two off-the-shelf ASR systems: a speaker-adapted (SA) system on a Windows laptop, and a speaker-independent (SI) system on an Android tablet. The experiment was carried out with speakers of three major languages and with different linguistic profiles: non-native English (Indian-accented and Spanish-accented) speakers, non-native French (Russian-accented and Spanish-accented) speakers, and native Spanish (Iberian-accented and Latin-American-accented) speakers. Five tasks were involved in this experiment: (1) typing, (2) reading aloud, (3) dictating with a commercial PC-based SA ASR system on a laptop, (4) dictating with a commercial cloud-based SI ASR system on a tablet, and (5) manually fixing ASR errors using a laptop keyboard. The focus of this experiment was on task times and ASR performance (i.e., word accuracy rates) across various users, languages and devices. No translation or respeaking task was involved in this pilot experiment. However, results highlight the potential advantages and drawbacks of a particular ASR approach (i.e., PC-based SA vs. cloud-based SI) for specific users, who have different accents and levels of proficiency in their working languages, and who may have different levels of competence and experience in translation, respeaking and (live) subtitling, and with ASR technology. Lastly, we enumerate several innovative areas for future work. References Garcia-Martínez, Mercedes, Singla, Karan, Tammewar, Aniruddha, Mesa-Lao, Bartolomé, Thakur, Ankita, Anusuya, M. A., Carl, Michael & Bangalore, Srinivas (2014) “SEECAT: ASR & Eye-tracking Enabled Computer-Assisted Translation”. In Proceedings of the 17th Annual Conference of the European Association for Machine Translation , pp. 81-88. Rodríguez, Luis, Reddy, Aarthy & Rose, Richard (2012) “Efficient Integration of Translation and Speech Models in Dictation Based Machine Aided Human Translation” In Proceedings of the IEEE 2012 International Conference on Acoustics, Speech, and Signal Processing, pp. 4949-4952. 23 5th International Symposium Respeaking, Live Subtitling and Accessibility Romero-Fresco, Pablo (2011) Subtitling Through Speech Recognition: Respeaking. Manchester: St. Jerome. Zapata, Julián (2014) “Exploring Multimodality for Translator-Computer Interaction”. In Proceedings of the 16th International Conference on Multimodal Interaction, pp. 339-343. Andreas Søeborg Kirkedal, [email protected] - PhD Fellow - Copenhagen Business School Spoken NLP Researcher - Mirsk Digital ApS - Denmark Julián Zapata, [email protected] - PhD (candidate), Translation Studies, School of Translation and Interpretation - University of Ottawa - Canada 24 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Transcribing documentaries: can respeaking be used efficiently? Lukasz Daniluk - University of Roehampton Anna Matamala - Universitat Autònoma de Barcelona Pablo Romero-Fresco - University of Roehampton Abstract: Respeaking is increasingly used to offer live intralingual subtitles for the deaf and hard-of-hearing (Romero-Fresco 2011), hence guaranteeing media accessibility to a wider section of the population. However, our hypothesis is that respeaking could be efficiently used for other tasks, such as transcribing non-fictional products that do not have an already created script. This could benefit transcribers in their jobs but would also allow to (semi-)automatically generate an input for machine translation processes, to name just one example. This paper aims to present the results of an experiment in which three scenarios were compared: a) manual transcription b) revision of an automatically generated transcription c) respeaking Three comparable 4-minute clips were selected, and 15 professional English transcribers were asked to do three different tasks: manually transcribe an excerpt, revise an automatically generated transcription (ASR), and respeak another excerpt. The order of the tasks and excerpt was randomized. The time spent in each of the tasks was monitorized, and data on the perceived effort and the transcriber experience were collected. This paper will describe the experimental set-up as well as the results of this test. This experiment is part of the ALST project (2013-2015), funded by the Spanish Ministry of Economy (FFI2012-31024), which aims to investigate the application of speech and translation technologies in the audio description of films and in the voice-over of non-fictional genres. (Matamala et al. 2013) References Matamala, Anna, Fernández-Torné, Anna, Ortiz-Boix, Carla (2013) “Enhancing sensorial and linguistic accessibility: further developments in the TECNACC and ALST projects”. 5th International Conference Media for All. Audiovisual Translation: Expanding Borders. Dubrovnik, 25-27/09/13. Romero-Fresco, Pablo (2011) Subtitling through speech recognition: Respeaking. Manchester: St. Jerome. Lukasz Daniluk, [email protected] - Ph. D. Student University of Roehampton - United Kingdom Anna Matamala, [email protected] - Full-time Senior Lecturer Universitat Autònoma de Barcelona - Spain Pablo Romero-Fresco, P. [email protected] - Reader in Audiovisual Translation and Filmmaking University of Roehampton - Department of Media, Culture and Language - United Kingdom 25 5th International Symposium Respeaking, Live Subtitling and Accessibility THEME 4 RESPEAKING AS A PROFESSIONAL OPPORTUNITY FOR THE BLIND Teaching live subtitling for the deaf through respeaking the potential of blind respeakers and paraplegic live editors Carlo Eugeni - Roma Tre University Abstract: Live subtitling through respeaking is still one of the most sectorial professions in Audiovisual Translation (Orero 2006). Teaching respeaking is all the more rare, especially at academic level (Remael & van der Veer 2006, Romero Fresco 2011). Among the reasons for this to happen, the organization of such courses in universities is hindered by a threefold lack: of the minimum technological requirements; of teachers being able to teach it; and of a real interest from both potential respeakers and the job market (Baaring 2006, Lambourne 2006). In this framework, disabled people are normally considered as the end users of the live subtitling service and, in terms of accessibility, they just receive the message and never give a real contribution in providing the service (Eugeni 2008). To try and bridge some of these gaps, a course in live subtitling through respeaking and live editing was organized by the association of disabled people “Cooperativa Progetto il Seme Onlus” based in Milan. The project was aimed at introducing disabled people in the job market, by giving them the possibility to exploit the skills they normally use as daily users of the PC workstation. In particular, blind people have been trained to professional respeaking and paraplegic people to professional live editing for three months by professional trainers. The trainees were finally tested by the Italian Academy of Multimedia Writing and Communication “Giuseppe Aliprandi - Flaviano Rodriguez”. In this speech, the didactic solutions that have been adopted during the course will be analyzed on the basis of the needs and expectations of both blind people and paraplegic people before the course itself. After that, their attitudes as trainees in respeaking and live editing will be discussed in view of perfecting the didactics of live subtitling in general (Russello 2013). Finally, the result of the tests will be analyzed and overall considerations will be made about the skills and competences of disabled people as professional live subtitlers. 26 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility References Baaring, I. (2006) “Respeaking-based online subtitling in Denmark”. In Eugeni, C. & G. Mack (eds.) InTRAlinea Special Issue: http://www.intralinea.org/specials/article/1685 Eugeni, C. (2008) Le sous-titrage en direct : aspects théoriques, professionnels et didactiques. Macerata: EUM. Lambourne, A. (2006) “Subtitle respeaking: A new skill for a new age”. In Eugeni, C. & G. Mack (eds.) InTRAlinea Special Issue: http://www.intralinea.org/specials/article/1686 Orero, P. (2006) “Real-time subtitling in Spain”. In Eugeni, C. & G. Mack (eds InTRAlinea Special Issue: http://www.intralinea.org/specials/article/1689 Remael, A. & B. van der Veer (2006) “Real-Time Subtitling in Flanders: Needs and Teaching”. In Eugeni, C. & G. Mack (eds.) InTRAlinea Special Issue: http://www.intralinea.org/specials/article/1702 Romero-Fresco, P. (2011) Subtitling through Speech Recognition: Respeaking. Manchester: St Jerome. Russello, C. (2013) “Aspetti didattici”. In Eugeni, C. & L. Zambelli (eds.) Respeaking. Specializzazione online. www.accademia-aliprandi.it pp. 44-53. Carlo Eugeni, [email protected] Università Roma Tre - Italy 27 5th International Symposium Respeaking, Live Subtitling and Accessibility Vocational Training for Blind Speech-to-Text Reporters Robin Ribback - VerbaVoice Abstract: The project set up by a consortium of international research and technology partners aims to set up vocational training for blind or partially sighted people in different European countries. The training on the use of speech recognition technology is set to provide live and remote speech-to-text services for hard of hearing, deaf and non-native people in the fields of inclusive education, live events, political debates or broadcast media. It enables twofold inclusion for visual and hearing impaired people. A remote interpreting system permits blind interpreters to control their output using a braille display connected to the cloud-based online platform both interpreters and clients are using. The multinational project team consists of the social enterprise VerbaVoice GmbH (accessible technologies for deaf and hard of hearing people, Germany), the Munich University of Applied Languages (SDI Munich, Germany), the BFW Würzburg (vocational training institute for blind and partially sighted people, Germany), GIB-BLWG Nuremberg (institute for the communication of hearing impaired people, Germany), the Universidad Autònoma Barcelona (UAB, Spain), the University of Roehampton (UoR, UK), subtitling company SUBTI (UK) and multimedia service provider SWISS TXT (Switzerland). More partners are most likely to join in. The project scope includes ICT-based teaching, training and assessment practices particularly for audiences of disadvantaged backgrounds facilitating transitions from education to the world of work for disadvantaged learners. The training program will be based on previous experiences made during training a training program for (non-blind) speechto-text-reporters held at the University of Applied Languages Munich in 2014 and one for visually impaired speech-to-text reporters organized by VerbaVoice and BFW Würzburg in 2015. The methodology, curriculum and management of the project will be adapted to its multinational character. 28 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility References Linder, Gion & Martínez, Juan (2013) “Access the Accuracy for Live Subtitles in a Quick and Effective Way.” ITU Workshop on Making Media Accessible to all: the options and economics. Electronic version available at: http://www.itu.int/en/ITU-T/Workshops-and-Seminars/mmaa/201310/Pages/ programme.aspx Martínez Pérez, Juan & Romero-Fresco, Pablo (2011) “Apples and Oranges: A Comparative Study of Accuracy in Respeaking”. 4th International Media for All Conference, Imperial College London. Mikul, Chris (2014) “Caption Quality: Approaches to Standards and Measurement.” Electronic version available at: http://www.mediaaccess.org.au/research-policy/white-papers/caption-quality-international-approaches-to-standards-and-measurement Romero-Fresco, Pablo (2011) “The NER Model to Assess Accuracy in Respeaking”. ITU Workshop on Telecommunications Relay Services for Persons with Disabilities. Electronic version available at: http:// www.itu.int/ITU-T/worksem/trs/programme.html Robin Ribback, [email protected] CIO VerbaVoice GmbH - Germany 29 5th International Symposium Respeaking, Live Subtitling and Accessibility ROUND TABLE: FOCUS ON ITALY Proposta di aggiornamento delle legislazioni italiane relative all’accessibilità dei programmi delle emittenti televisive per i cittadini sordi Giacomo Pirelli - On A.I.R. - Intersteno Italia Luca Iacovino - Presidente ANS Abstract: L’attuale Contratto di Servizio per la sottotitolazione televisiva lega la RAI alla Commissione di Vigilanza Rai, ma tale vincolo non si estende alle altre emittenti televisive. Tale situazione non tiene conto della liberalizzazione delle bande televisive e del conseguente numero maggiore di interlocutori. Con l’occasione del rinnovo di tale contratto alla sua scadenza triennale, sarebbe opportuno aggiornarlo ed estenderlo alle altre emittenti televisive. Il nuovo Contratto di Servizio e l’aggiornamento del Decreto Legislativo 31 luglio 2005, n. 177 Testo unico della radiotelevisione, dovrebbero specificare le esigenze delle persone sorde riguardo una fruizione accessibile dei programmi televisivi. Tale Contratto dovrebbe coinvolgere non solo la RAI, ma anche le reti televisive private, come Mediaset, LA7, Sky, ecc., per una piena applicabilità dei diritti fondamentali di accesso all’informazione da parte degli utenti sordi. Ciò sarebbe in conformità all’articolo 30 della Convenzione ONU sui Diritti Umani delle Persone con Disabilità, ratificata in Italia nel 2009, ed alle normative di diversi paesi europei, in particolare Regno Unito e Francia. Le emittenti televisive sarebbero così obbligate a sottotitolare una percentuale dei programmi, per esempio al minimo dal 60%. Esse dovrebbero inoltre definire con le associazioni delle linee guida sulla qualità dei sottotitoli nei programmi televisivi, sia in forma preregistrata che in tempo reale, per tutti i canali televisivi, seguendo gli esempi delle migliori prassi (per esempio le linee guida approvate dall’OFCOM per la BBC e tutti gli altri canali televisivi statali e privati nel Regno Unito e quelle approvate dal Conseil Supérieur de l’Audiovisuel in Francia nel dicembre 2011 per tutti i canali). 30 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility References ITALIA: articoli 4 e 45 nel Decreto Legislativo 31 luglio 2005, n. 177 Testo unico della radiotelevisione. Schema di contratto di servizio tra il Ministero dello sviluppo economico e la RAI-Radiotelevisione Italiana Spa, per il triennio 2013-2015 http://www.senato.it/leg/17/BGT/Schede/docnonleg/25703.htm FRANCIA: Article 74 - LOI n° 2005-102 du 11 février 2005 pour l’égalité des droits et des chances, la participation et la citoyenneté des personnes handicapées Charte relative a la qualité du sous-titrage á destination des personnes sourdes ou malentendantes (Linee Guida di sottotitolazione televisiva), Conseil Supérieur de L’Audiovisuel 2011: http://www. csa.fr/Espace-juridique/Chartes/Charte-relative-a-la-qualite-du-sous-titrage-a-destination-des-personnes-sourdes-ou-malentendantes-Decembre-2011 REGNO UNITO: Section 303 - Communications Act 2003 Code on Television Access Standard: http://stakeholders.ofcom.org.uk/binaries/broadcast/other-codes/ctas.pdf Linee Guida di sottotitolazione televisiva dell’OFCOM (Office of Communication) http://www.ofcom.org.uk/static/archive/itc/itc_publications/codes_guidance/standards_for_subtitling/subtitling_1.asp.html Aggiornamento linee guida OFCOM (2013): http://consumers.ofcom.org.uk/2013/05/improving-the-quality-of-live-tv-subtitles/ Giacomo Pirelli, [email protected] onA.I.R. - Intersteno Italia - Italy Luca Iacovino, [email protected] Presidente ANS - Italy 31 5th International Symposium Respeaking, Live Subtitling and Accessibility La Meeting Industry in Italia oggi può offrire attraverso il respeaking nuove opportunità sia per gli organizzatori di congressi ed eventi che per i professionisti della documentazione scritta Samuel Zambelli - Studio Acta Abstract: Il V° Simposio Internazionale sul respeaking, sottotitolazione in diretta e accessibilità offre agli operatori della comunicazione/documentazione scritta un’importante occasione per riflettere sullo stato dell’arte della ricerca e, nel contempo, fornisce elementi di analisi delle problematiche connesse con un ulteriore sviluppo dei software di riconoscimento del parlato e, conseguentemente, dei servizi di produzione di testi scritti. Nel corso della mia attività professionale ho potuto seguire l’evolversi degli strumenti utilizzati dai professionisti nei diversi ambiti operativi: un repertorio di sussidi sempre più sofisticati, che hanno comportato significative ricadute a livello operativo. La rapidissima evoluzione tecnologica ha avuto anche forti implicazioni sul fronte della formazione permettendo a soggetti dotati di specifiche competenze e riconosciute abilità di tipo tecnico di proporre servizi innovativi nell’area della documentazione scritta. Una carente azione di informazione e sensibilizzazione di quanti, ai diversi livelli, possono trarre vantaggi dall’introduzione delle nuove tecniche e software di produzione del testo scritto, ha pesantemente condizionato il loro sviluppo mentre – anche grazie agli spunti che emergeranno in questo convegno - si possono sicuramente prefigurare significativi sviluppi nell’uso professionale dei software di riconoscimento del parlato e, in particolare, del respeaking, ovvero della tecnica di produzione del testo scritto attraverso l’intermediazione di un professionista. Mi limiterò a proporre qualche riflessione unicamente sulla tecnica del respeaking nella sua applicazione nel mondo congressuale. Non mi occuperò dei software “speaker indipendent” che, a mio parere, se possono trovare applicazione in aree ben limitate della documentazione scritta (refertazione medica, trascrizione brevi messaggi, etc.), non hanno ancora raggiunto un grado di precisione che permetta di fornire dei servizi di trascrizione, né tanto meno di resocontazione, con un accettabile standard qualitativo. Gli stimoli che questo Simposio internazionale ci propone necessitano quindi di un’attenta e non rituale analisi, sia sul piano accademico/didattico, sia sul piano operativo, dei multiformi aspetti del respeaking come tecnica privilegiata per la resocontazione e per la sottotitolatura. È questa un’importante occasione per inaugurare un rapporto di sinergica collaborazione tra il mondo della ricerca e gli attori privilegiati dell’area congressuale e convegnistica in senso lato. Samuel Zambelli, [email protected] Sales & Marketing Manager - Studio Acta - Italy 32 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility La telefonata sottotitolata Lorenzo Di Ciaccio - Pedius Abstract: La sottotitolazione ha da sempre avuto impiego nell’ambito dell’informazione e dell’intrattenimento. Partendo dal cinema fino alla televisione, il sottotitolo ha agevolato la fruizione a tutta la fascia della popolazione con problemi uditivi. Oggi grazie alle nuove tecnologie è possibile introdurre la sottotitolazione in tempo reale anche all’interno delle telefonate agevolando la comunicazione tra persone sorde e persone udenti. Pedius consente alle persone sorde di telefonare grazie alle le tecnologie di riconoscimento del parlato e di sintesi vocale. Attraverso l’applicazione mobile gli utenti possono avviare una telefonata verso un numero telefonico. Durante la telefonata l’utente scrive come in una normale chat e una voce artificiale pronuncerà i messaggi. In tempo reale la risposta vocale della persona chiamata viene trasformata in testo ed inviata all’utente. Lorenzo Di Ciaccio, [email protected] CEO Pedius - Italy 33 5th International Symposium Respeaking, Live Subtitling and Accessibility New frontiers of respeaking: live subtitles delivered on smartphones, tablets and electronic glasses Saveria Arma - Artis-Project / CulturAbile / MovieReading Abstract: Subtitling through respeaking is a relatively well-known technique for live TV programs and news. However, when it comes to conference respeaking, literature and practices become much less abundant. This contribution represents an introduction to respeaking-based live subtitling for the deaf and the hard of hearing in a conference setting. In particular, we will focus on new technical and technological solutions for providing and reading respeaking-based subtitles: from classical “PC-to-beamer” configuration with subtitles produced and projected by an operator onsite, to remote respeaking operated by a team of geographically distant respeakers delivered on individual devices (smartphones, tablets and electronic glasses) in one or more languages. The individual distribution of accessible content is becoming one of the major frontiers in accessibility as a whole, not only intended as a tool for including people with sensory impairment but - more and more often - as a cutting-edge cost-effective solution for conferences and events attended by people who speak different languages or do not have sufficient knowledge of the working language (this solution is also being considered as an alternative to traditional simultaneous interpreting). We will present a new solution for producing real-time respeaking-based subtitles (onsite/remote captioning) and to distribute them on multiple platforms, including electronic augmented-reality glasses. We will draw on preliminary evidence and results in terms of subtitles production and delivery processes, as well as of feedback from the hearing and hearing-impaired audience as to the type of device, readability of subtitles and “social” function. References Eugeni, Carlo & Mack, Gabriele (eds) (2006) Intralinea, Respeaking Special Issue. Proceedings of the First International Seminar on real-time intralingual subtitling, University of Bologna - Forlì Campus. 17th November 2006. Available at: http://www.intralinea.org/specials/respeaking Romero-Fresco, Pablo (2011) Subtitling through Speech Recognition: Respeaking. Manchester: St. Jerome Publishing. Saveria Arma, [email protected] Artis-Project / CulturAbile / MovieReading - Italy 34 5th In lity 5th International Symposium Respeaking, Live Subtitling and Accessibility Elenco degli oratori - List of speakers Name Page Aliprandi, Carlo 18-19 Arma, Saveria 30 Chmiel, Agnieszka 15-16 Daniluk, Lukasz, 21-22 Di Ciaccio, Lorenzo 29 Dutka, Łukasz 14, 15-16 Eugeni, Carlo 23-24 Iacovino, Luca 26-27 Krejtz, Izabela 15-16 Krejtz, Krzysztof 15-16 Lambourne, Andrew 17 Lijewska, Agnieszka 15-16 Martínez Pérez, Juan 12 Matamala, Anna 21-22 Moores, Zoe 11 Pirelli, Giacomo 26-27 Pražák, Aleš 10 Ribback, Robin 25 Romero-Fresco, Pablo 9, 21-22 Søeborg Kirkedal, Andreas 20 Stachowiak, Katarzyna 13 Szarkowska, Agnieszka 14, 15-16 Widmer-Iliescu, Roxana 8 Zambelli, Samuel 28 Zapata, Julián 20 35 5th International Symposium Respeaking, Live Subtitling and Accessibility 36