supported by gold sponsor basic sponsor media partner

Transcript

supported by gold sponsor basic sponsor media partner
SUPPORTED BY
GOLD SPONSOR
BASIC SPONSOR
MEDIA PARTNER
5th International Symposium Respeaking, Live Subtitling and Accessibility
TABLE OF CONTENTS
The Symposium
2
Programme6-7
Abstracts11
Speakers35
INDICE
Il Simposio
4
Programma8-9
Abstracts11
Oratori35
1
5th International Symposium Respeaking, Live Subtitling and Accessibility
5th International Symposium
Respeaking, Live Subtitling and Accessibility.
12 June 2015
UNINT - Università degli Studi Internazionali di Roma is pleased to host the 5th International Symposium on Respeaking, Live Subtitling and Accessibility. This is the fifth
symposium in the series, after Forlì (2005), Barcelona (2007), Antwerp (2011) and Barcelona (2013). These events aim to bring together academic researchers, software developers, broadcasters, consumers and anyone interested in recent developments in
media accessibility.
In February 2013 the European Parliament Committee on Culture and Education
published a Report on the transposition of Directive 2010/13/EU on Audiovisual
Media Services in EU countries. The Report found that progress is still slow and uneven
throughout the EU and that further developments are needed to ensure wider access
to programmes and information for all citizens via audio description, audio/spoken
subtitles and sign language.
As regards live subtitling, over the past 10 years respeaking by means of speakerdependent speech recognition (SR) technology has become the preferred method to
provide live subtitles on TV and in live events in many countries. At the same time,
there have been significant improvements in speaker-independent speech recognition
technology, with interesting applications in automatic subtitling and automatic speech
reporting. Moreover, the deployment of the HbbTV (Hybrid Broadcast Broadband TV)
standard in Europe is opening up more opportunities for the delivery of audiovisual
content to consumers, ranging from traditional broadcast TV, video on-demand and
catch-up TV services, like BBC iPlayer, as well as interactive services and Internet
applications. Clearly, these developments are also creating more demand for
accessibility services.
The 5th Symposium will be an opportunity to look at areas of current research and
future prospects in SR, live subtitling, and accessibility in general.
2
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Scientific Committee
Carlo Eugeni (Roma Tre University)
Pilar Orero (Universitat Autònoma de Barcelona)
Aline Remael (University of Antwerp)
Pablo Romero Fresco (University of Roehampton)
Annalisa Sandrelli (UNINT University)
Local Organising Committee
Annalisa Sandrelli
Serena Scaldaferri
Claudio Russello
Symposium website:
http://www.unint.eu/it/component/content/article/8-pagina/494-respeaking-livesubtitling-and-accessibility.html
Facebook page:
https://www.facebook.com/respeakingsimposium?fref=ts
3
5th International Symposium Respeaking, Live Subtitling and Accessibility
Quinto Simposio Internazionale
Respeaking, Sottotitolazione in diretta e Accessibilità
12 giugno 2015
La UNINT - Università degli Studi Internazionali di Roma è lieta di ospitare il Quinto
Simposio Internazionale su Respeaking, Sottotitolazione in diretta e Accessibilità. Si
tratta della quinta edizione di un ciclo iniziato a Forlì (2005) e proseguito poi a Barcellona (2007), Anversa (2011) e ancora Barcellona (2013). Questi eventi costituiscono un
punto di incontro per ricercatori, studiosi, aziende informatiche, broadcaster, utenti e
chiunque si interessi di accessibilità e delle applicazioni delle tecnologie di riconoscimento del parlato.
Nel febbraio 2013 la commissione Cultura e Istruzione del Parlamento europeo ha
pubblicato una Relazione sulla trasposizione della Direttiva 2010/13/UE sui servizi
audiovisivi. La Relazione ha evidenziato che I progressi sono ancora lenti e poco
uniformi nei vari paesi dell’Unione e che sono necessari ulteriori sviluppi per assicurare
un più ampio accesso all’intrattenimento e alle informazioni a tutti i cittadini, tramite
audio descrizione, audiotitoli, sottotitoli e lingue dei segni.
Per quanto riguarda la sottotitolazione in diretta, negli ultimi 10 anni il respeaking
mediante la tecnologia di riconoscimento del parlato speaker-dependent è divenuto
il metodo più comune in molti paesi per produrre sottotitoli in diretta per la TV e gli
spettacoli dal vivo. Contemporaneamente, la tecnologia di riconoscimento del parlato
speaker-independent ha fatto grandi passi avanti, con interessanti applicazioni alla
sottotitolazione automatica e alla resocontazione automatica. Inoltre, l’avvento dello
standard HbbTV (Hybrid Broadcast Broadband TV) nel sistema televisivo europeo sta
aprendo nuove opportunità ai fornitori di contenuti audiovisivi, dalle trasmissioni
televisive in modalità tradizionale al video on-demand e catch-up TV (come l’iPlayer
della BBC), fino ad arrivare ai servizi interattivi e alle applicazioni Internet. Ovviamente
tutti questi sviluppi contribuiscono ad accrescere anche la domanda di servizi di
accessibilità per tutti i consumatori.
Il Quinto Simposio sarà l’occasione per poter riflettere sulla ricerca attuale e sulle
prospettive future delle tecnologie di riconoscimento del parlato, della sottotitolazione
in diretta e dell’accessibilità in generale.
4
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Comitato scientifico
Carlo Eugeni (Università Roma Tre)
Pilar Orero (Università Autonoma di Barcellona)
Aline Remael (Università di Anversa)
Pablo Romero Fresco (Università di Roehampton)
Annalisa Sandrelli (Università degli Studi Internazionali di Roma - UNINT)
Comitato organizzatore locale
Annalisa Sandrelli
Serena Scaldaferri
Claudio Russello
Sito web del Simposio:
http://www.unint.eu/it/calendario-eventi/-22.html
Pagina Facebook:
https://www.facebook.com/respeakingsimposium?fref=ts
5
5th International Symposium Respeaking, Live Subtitling and Accessibility
PROGRAMME
Aula Magna
Morning session
6
08.30 - 10.00
Registration
09.00 - 09.15
Welcome and conference opening
09.15 - 09.30
Roxana Widmer-Iliescu, Unione Internazionale delle
Telecomunicazioni “ITU work on ICT Accessibility”
Theme 1.
Moderator:
Live subtitling on television: focus on quality
Gion Linder, Swiss Text
09.30 - 09.50
Pablo Romero Fresco, “Live subtitling in the UK: an official review”
09.50 - 10.10
Aleš Pražák, “Live TV subtitling in Czech”
10.10 - 10.30
Zoe Moores, “The language of respeaking”
10.30 - 10.50
Juan Martínez Pérez, “New Approaches to Improve the Quality of Live
Subtitling on TV”
10.50 - 11.00
Q & A Session
11.00 - 11.30
Coffee break
Theme 2.
Moderator:
Respeaking: Experimental Studies
Pablo Romero Fresco, University of Roehampton
11.30 - 11.50
Katarzyna Stachowiak, “More than meets the ear: Do subtitles matter
for the well-educated, speaking hearing impaired?”
11.50 - 12.10
Agnieszka Szarkowska, Łukasz Dutka, “Are interpreters better respeakers?
An exploratory study on respeaking competences”
12.10 - 12.30
Agnieszka Chmiel, Agnieszka Lijewska, Łukasz Dutka, Agnieszka
Szarkowska, Isabela Krejtz, Krzysztof Krejtz, “Tapping the linguistic
competence in the respeaking process: comparing intralingual
paraphrasing done by interpreting trainees, translation trainees and
bilinguals”
12.30 - 12.40
Q & A Session
13.00 - 14.30
Lunch break
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Afternoon session
Theme 3.
Moderator:
New applications of technology
Pilar Orero, Universitat Autònoma de Barcelona
14.30 - 14.50
Andrew Lambourne, “Pushing forward the frontiers of access technology”
14.50 - 15.10
Carlo Aliprandi, “Evaluating performances of automatic speech
recognition for assisted subtitling”
15.10 - 15.30
Andreas Søeborg Kirkedal, Julián Zapata, “Translating and respeaking
in a native or foreign language and from different environments: a prepilot experiment with a laptop and a tablet”
15.30 - 15.50
Lukasz Daniluk, Anna Matamala, Pablo Romero-Fresco, “Transcribing
documentaries: can respeaking be used efficiently?”
15.50 - 16.00
Q & A Session
Theme 4.
Moderator:
Respeaking as a professional opportunity for the blind
Federico Spoletti, Sub-Ti & FRED.fm
16.00 - 16.20
Carlo Eugeni “Teaching live subtitling for the deaf through respeaking
- the potential for blind respeakers and paraplegic live editors”
16.20 - 16.40
Robin Ribback, “Vocational training for blind speech-to-text-reporters”
16.40 - 16.50
Q & A Session
16.50 - 17.00
Coffee break
ROUND TABLE: Focus on Italy
Moderator:
Carlo Eugeni, Università Roma Tre
17.00 - 18.30
Giacomo Pirelli, Luca Iacovino, “Proposta di aggiornamento delle
legislazioni italiane relative all’accessibilità dei programmi delle
emittenti televisive per i cittadini sordi”
Samuel Zambelli, “La Meeting Industry in Italia oggi può offrire attraverso
il respeaking nuove opportunità sia per gli organizzatori di congressi ed
eventi che per i professionisti della documentazione scritta?”
Lorenzo Di Ciaccio, “La telefonata sottotitolata”
Saveria Arma, “New frontiers of respeaking: live subtitles delivered
on smartphones, tablets and electronic glasses”
18.30
Closing
7
5th International Symposium Respeaking, Live Subtitling and Accessibility
PROGRAMMA
Aula Magna
Mattina
08.30 - 10.00 Registrazione partecipanti
09.00 - 09.15
Saluti e apertura del Simposio
09.15 - 09.30
Roxana Widmer-Iliescu, International Telecommunications Union
“ITU work on ICT Accessibility”
Tema 1.
La sottotitolazione in diretta in televisione: la qualità
Moderatore: Gion Linder, Swiss Text
09.30 - 09.50
Pablo Romero Fresco, “Live subtitling in the UK: an official review”
09.50 - 10.10
Aleš Pražák, “Live TV subtitling in Czech”
10.10 - 10.30
Zoe Moores, “The language of respeaking”
10.30 - 10.50
Juan Martínez Pérez, “New Approaches to Improve the Quality of
Live Subtitling on TV”
10.50 - 11.00
Dibattito
11.00 - 11.30 Pausa caffè
Tema 2.
Respeaking: studi sperimentali
Moderatore: Pablo Romero Fresco, University of Roehampton
11.30 - 11.50
Katarzyna Stachowiak, “More than meets the ear: Do subtitles matter for the well-educated, speaking hearing impaired?”
11.50 - 12.10
Agnieszka Szarkowska, Łukasz Dutka, “Are interpreters better respeakers? An exploratory study on respeaking competences”
12.10 - 12.30
Agnieszka Chmiel, Agnieszka Lijewska, Łukasz Dutka, Agnieszka
Szarkowska, Isabela Krejtz, Krzysztof Krejtz, “Tapping the linguistic competence in the respeaking process: comparing intralingual
paraphrasing done by interpreting trainees, translation trainees and
bilinguals”
12.30 - 12.40
Dibattito
13.00 - 14.30 Pausa pranzo
8
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Pomeriggio
Tema 3.
Moderatore:
Nuove applicazioni della tecnologia
Pilar Orero, Universitat Autònoma de Barcelona
14.30 - 14.50
Andrew Lambourne, “Pushing forward the frontiers of access technology”
14.50 - 15.10
Carlo Aliprandi, “Evaluating performances of automatic speech recognition for assisted subtitling”
15.10 - 15.30
Andreas Søeborg Kirkedal, Julián Zapata, “Translating and respeaking in a native or foreign language and from different environments:
a pre-pilot experiment with a laptop and a tablet”
15.30 - 15.50
Lukasz Daniluk, Anna Matamala, Pablo Romero-Fresco, “Transcribing documentaries: can respeaking be used efficiently?”
15.50 - 16.00
Dibattito
Tema 4.
Moderatore:
Il respeaking come sbocco professionale per i non vedenti
Federico Spoletti, Sub-Ti & FRED.fm
16.00 - 16.20
Carlo Eugeni, “Teaching live subtitling for the deaf through respeaking- the potential for blind respeakers and paraplegic live editors”
16.20 - 16.40
Robin Ribback, “Vocational training for blind speech-to-text-reporters”
16.40 - 16.50
Dibattito
16.50 - 17.00 Pausa caffè
TAVOLA ROTONDA: Uno sguardo sull’Italia
Moderatore: Carlo Eugeni, Università Roma Tre
17.00 - 18.30
Giacomo Pirelli, Luca Iacovino, “Proposta di aggiornamento delle legislazioni italiane relative all’accessibilità dei programmi delle emittenti televisive per i cittadini sordi”
Samuel Zambelli, “La Meeting Industry in Italia oggi può offrire attraverso il respeaking nuove opportunità sia per gli organizzatori di congressi ed
eventi che per i professionisti della documentazione scritta?”
Lorenzo Di Ciaccio, “La telefonata sottotitolata”
Saveria Arma, “New frontiers of respeaking: live subtitles delivered
on smartphones, tablets and electronic glasses”
18.30
Chiusura Simposio
9
5th International Symposium Respeaking, Live Subtitling and Accessibility
Abstracts
10
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
SYMPOSIUM OPENING
ITU work on ICT Accessibility
Roxana Widmer-Iliescu - ITU
Abstract: The presentation will share the work of the International Telecommunication Union (ITU) in promoting ICT accessibility. ITU is the leading United Nations agency for Information and Communication Technologies (ICTs) which includes 193 Member States. The
presentation will highlight ITU’s work in promoting ICT accessibility in its technical standards,
its policy and regulatory advice to Member States, capacity building initiatives, such as developing accessible audio visual services, and will introduce ITU’s own accessibility policy which
covers accessible meeting services such as captioning for the deaf and hard of hearing.
ITU adopted the first international technical accessibility standard in 1994 and developed
the Telecommunications Accessibility Checklist (2006) to assist experts developing technical
standards to ensure that they take ICT accessibility for persons with disabilities into account
when developing technical standards. ITU Member States adopted the Connect 2020 targets
which call for all countries to have ICT accessibility policies in place by 2020. ITU is supporting
its Member States to achieve this goal through the work of its Study Groups and Regional
initiatives including the European Regional Initiative, Ensuring access to telecommunications/ICTs, in particular for persons with disabilities. Participants will learn about the range
of ITU resources to promote ICT accessibility including its reports Making Mobile Phones and
Services Accessible, Making Television Accessible and the ITU Model ICT Accessibility Policy
Report which regulators and policy makers can use in developing their own national ICT accessibility policies. These policies cover public access points, television/video programming,
mobile phones, web sites and public procurement. Multi-stakeholder cooperation, including
operators and organizations of persons with disabilities, is essential for the successful definition and implementation of ICT accessibility policies.
References
ITU http://www.itu.int/en/Pages/default.aspx; www.itu150.org
FSTP-TACL Telecommunications Accessibility Checklist (2006): http://www.itu.int/pub/T-TUT-FSTP-2006-TACL/en
European Regional Initiative, Ensuring access to telecommunications/ICTs, in particular for persons with
disabilities. http://www.itu.int/en/ITU-D/Regional-Presence/Europe/Pages/ActivitiesandProjects.aspx
ITU Report (2012) Making Mobile Phones and Services Accessible for Persons with Disabilities. http://www.
itu.int/ITU-D/sis/PwDs/Documents/Mobile_Report.pdf
ITU Report (2011) Making Television Accessible http://www.itu.int/ITU-D/sis/PwDs/Documents/Making_
TV_Accessible-FINAL-WithAltTextInserted.pdf
ITU Report (2014) Model ICT Accessibility Policy Report http://www.itu.int/en/ITU-D/Digital-Inclusion/Persons-with-Disabilities/Documents/ICT%20Accessibility%20Policy%20Report.pdf
Roxana Widmer-Iliescu, [email protected]
Senior Programme Officer and BDT( ITU Development Bureau) Focal Point for Study Group on Question 7 “Access to telecommunication /ICT services by Persons with Disabilities and with specific needs”
International Telecommunication Union - Switzerland
11
5th International Symposium Respeaking, Live Subtitling and Accessibility
THEME 1
LIVE SUBTITLING ON TELEVISION: FOCUS ON QUALITY
Live Subtitling in the UK: An Official Review
Pablo Romero-Fresco - University of Roehampton
Abstract: In May 2013, Ofcom, the governmental communications regulator in the UK,
published a consultation document in order to gather views from broadcasters, subtitle
providers and user associations as to how to improve the quality of live captioning on
UK TV to benefit deaf and hard-of-hearing viewers. Following this consultation, Ofcom
decided that broadcasters should be required to assess the speed, reduction rate, latency
and accuracy of their live captions using the NER model (Romero-Fresco and Martínez
2015). Ofcom also decided that the University of Roehampton (London) would review
the measurements provided by broadcasters from a third party standpoint in four stages
over two years between 2013 and 2015. So far, three of the four measurements have been
conducted, including the analysis of 216 programmes belonging to three genres (news,
entertainment and chat shows) and broadcast by the BBC, ITV, Channel 4, Five and Sky. The
review has covered 365.000 words, 55.000 subtitles and 50 hours of TV material.
The aim of this presentation is to discuss the results of this review, focusing on the different
issues involved in the quality of live subtitles, such as accuracy (including accuracy rates,
types of errors and reasons why they are occurring), delay, speed and format. It will also deal
with the challenges posed by the hybrid mode (the combination of live subtitles and prerecorded subtitles cued live) introduced in the UK as a result of the Ofcom review in order to
improve the quality of subtitling. Initially welcome as a means to introduce 100% accurate
and perfectly synchronised subtitles in live programmes, this hybrid mode is also bringing
about subtitles with speeds exceeding 400wpm and even 500wpm regularly. Despite not
being legible for the viewers, these subtitles contribute to increasing the overall accuracy
rate and to decreasing the average delay of the broadcasters’ output. This presentation will
offer some suggestions as to how this issue may be tackled by the subtitling companies and
how it can be factored in the accuracy analysis using the NER model.
References
Romero-Fresco, Pablo & Martínez Pérez, Juan (2015) “Accuracy in Live Subtitling- The NER model”,
in Jorge Díaz Cintas & Rocío Baños (eds) Audiovisual Translation in a Global Context: Mapping an Ever-changing Landscape, London: Palgrave.
Pablo Romero-Fresco, [email protected]
Reader in Audiovisual Translation and Filmmaking
University of Roehampton - Department of Media, Culture and Language - United Kingdom
12
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Live TV subtitling in Czech
Aleš Pražák - SpeechTech, s.r.o.
Abstract: We would like to introduce our solution for live TV subtitling that we provide
as a service for the Czech Television, the public service broadcaster in the Czech Republic. We operate a fully automated subtitling system (without a re-speaker) for the subtitling of meetings of the Parliament of the Czech Republic directly from the real acoustic
track, using specifically designed speaker-independent speech recognition.
At the same time, since we develop both the speech recognition and live subtitling
technologies, we implemented a novel approach to live subtitling through re-speaking.
We use only one re-speaker doing all the tasks with very close interaction with the recognition system, i.e. instant corrections, speaker colouring, online word additions, etc.
Moreover, our subtitling platform is completely distributed – our re-speakers operate
at home with bidirectional connection to the television company, so the subtitle latency is significantly reduced.
We also provide supplementary services, such as language model adaptation for
sport programmes, an automatic daily updating of speech recognizer vocabulary for
news subtitling, correction and automatic timing of subtitles for programme rerun or
re-speaker training using special four-phase training application. We have several years
of experience in live TV subtitling, including subtitling of sport events such as Winter
Olympic Games or FIFA World Cup. Now we have 8 skilled re-speakers who subtitled
almost 2.000 hours in 2014.
References
Pražák, A., Loose, Z., Trmal, J., Psutka, J. V. & Psutka, J. (2012) “Novel Approach to Live Captioning
Through Re-speaking: Tailoring Speech Recognition to Re-speaker’s Needs”. INTERSPEECH, ISCA.
Trmal, J.; Pražák, A.; Loose, Z. & Psutka, J. (2010) “Online TV Captioning of Czech Parliamentary
Sessions”. In: Petr Sojka, Ales Horák, Ivan Kopecek & Karel Pala (eds) LNCS (LNAI), vol. 6231, pp. 416422, Springer, Heidelberg.
Aleš Pražák, [email protected]
Project manager
SpeechTech, s.r.o. - Czech Republic
13
5th International Symposium Respeaking, Live Subtitling and Accessibility
The Language of Respeaking
Zoe Moores - University of Roehampton
Abstract: Subtitle providers in the UK are under particular scrutiny from the communications regulator Ofcom over the accuracy of their live subtitling. Since the very nature
of respeaking means it is done in an intense, high-pressure environment, any guidance
available on how to achieve more accurate subtitles will benefit trainers, individual subtitlers and deaf and hard of hearing viewers, for whom the degree of accessibility provided
depends on the quality of the subtitles they receive. In this paper I present the results of
a study into the errors that occur whilst respeaking in English and their causes.
The corpus consists of 47 weather forecasts, sampled from the major television channels
and totalling almost an hour of respeaking. The style and speed of this programme, along
with the variable content that appears within the chatty asides, makes it particularly
challenging for the respeaker and a relevant focus for a study on accuracy.
In the first half of the corpus, I respoke each clip into a new voice model and replicated
key features of in-house subtitling software. Errors were then analysed, potential
solutions tested and the clip respoken for a second time. From this detailed analysis, it
was possible to create a taxonomy of the English used in respeaking, which revealed the
lexical categories errors fell into, the parts of speech that were most affected and the
cause of each error. The key features of a respeaking style which maximizes accuracy
and reduces the likelihood of errors were revealed. In the second half, consideration
was given to whether these findings could be applied to a range of respeakers working
with professional software and to the balance between the degree of editing and the
impact on the content that the SDH viewer accesses.
This presentation concludes with some recommendations based on the findings obtained.
References
Greenberg, Steven & Fosler-Lussier, Eric (2000) “The Uninvited Guest: Information’s Role in Guiding
the Production of Spontaneous Speech.” In: Proceedings Of The CREST Workshop On Models Of Speech
Production: Motor Planning And Articulatory Modelling. Kloster Seeon, Germany, May 1-4.
Moores, Zoe (2014) Respeaking The Weather Accurately - Why The Nation’s Favourite Topic is a Respeaker’s Nightmare. UK, London: University of Roehampton (Unpublished dissertation).
Roach, Peter (2009) English Phonetics and Phonology A Practical Course Fourth Edition. UK: Cambridge
University Press.
Romero-Fresco, Pablo (2009) “More Haste Less Speed: Edited Versus Verbatim Respoken Subtitles.”
Vigo International Journal of Applied Linguistics, vol.6, pp. 109-133.
Romero-Fresco, Pablo & Martínez, Juan (2014). Accuracy Rate in Live Subtitling - The NER Model.
https://roehampton.openrepository.com/roehampton/bitstream/10142/141892/1/NER-English.pdf
Shockey, Linda (2003) Sound Patterns Of Spoken English. UK: Blackwell Publishing.
Zoe Moores, [email protected] - Graduate Student
University of Roehampton - Department of Media, Culture and Language - United Kingdom
14
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
New Approaches to Improve the Quality of Live Subtitling on TV
Juan Martínez Pérez - TransMedia Catalonia, UAB, CaiaC
Abstract: The aim of this presentation is twofold. First, to define certain quality-related
aspects of live subtitling on TV and how to benchmark them, and second to describe
the NERstar Editor. This tool is an advanced application for quality measurement and
data generation, which has been developed by the author on behalf of SWISS TXT in
close collaboration with the German company VerbaVoice.
Regarding the first point, two questions arise: what is meant by “quality-related aspects
of live subtitling on TV” and how to define quality benchmarking. Attention should also
be paid to the different preferences and expectations of the countries with respect to
quality-related criteria in live subtitling on TV, and the many differences in legislation,
funding and production models. Language always works within a culture, hence distinct cultural background will also have an impact in quality benchmarking. In other
words: quality is a highly complex interplay of very contrasting ideas and often controversial issues. As to the second point, the aim is to present NERstar Editor, a software
solution that has specifically been developed for assessing and improving the quality in
live subtitling on TV linked to accuracy, delay, speed and display issues.
References
Romero-Fresco, Pablo & Martínez, Juan (2014) Accuracy Rate in Live Subtitling - The NER Model.
https://roehampton.openrepository.com/roehampton/bitstream/10142/141892/1/NER-English.pdf
Juan Martínez Pérez, [email protected]
TransMedia Catalonia, CaiaC,
UAB, Barcelona - Spain
15
5th International Symposium Respeaking, Live Subtitling and Accessibility
THEME 2
RESPEAKING: EXPERIMENTAL STUDIES
More than meets the ear: Do subtitles matter
for the well-educated, speaking hearing-impaired?
Katarzyna Stachowiak - Adam Mickiewicz University
Abstract: The common belief is that the main target group of interlingual subtiling or
closed captions for the deaf and hard of hearing are individuals whose command of
spoken language is poor or who do not have that command (Neves 2008). On the other
hand, well-educated patients native in a spoken language or with excellent command
of it may benefit greatly from subtitles.
The aim of the study was to verify to what extent patients with hearing aids and cochlear implants benefit from subtitled media content, and more specifically: live debates.
There were three groups of participants with hearing loss: the ones with mild to moderate (up to 55 dB), moderately severe to severe (up to 90 dB) and profound (> 90 dB)
hearing loss. Additionally, there were groups of participants unilaterally and bilaterally
implanted (implanted for 2 years or more). Particular groups included 40 participants,
resulting in the total number of participants equal to 200. The subjects were aged 2443 and had higher education. There were 42-66% of females in the groups. The level
of understanding speech was above 80% in silence and above 50% in noise in all the
participants. The subjects’ native language and the language of daily use was spoken
Polish. The participants were sent two 5-minute films (live debates engaging 3 speakers), one subtitled in Polish. The participants were asked to fill in a questionnaire composed of 20 questions assessing the level of understanding the content with or without subtitling and the satisfaction from the subtitles. The results revealed a significant
need for and satisfaction from subtitling, and no correlation with age and the level of
understanding speech in silence, nor noise, in all the participants. The fact of having a
hearing loss or deafness correlated positively with the level of satisfaction, excluding
subjects with mild to moderate hearing impairment. The results shed new light on the
debate about potential beneficiaries of subtitling and the need to have media content
subtitled, as well as the needs of audience with hearing loss or deafness.
References
Neves Josélia (2008) “10 fallacies about Subtitling for the d/Deaf and the hard of hearing”, The Journal
of Specialised Translation, vol. 10: 128.
Katarzyna Stachowiak, [email protected]
Adam Mickiewicz University in Poznań
Department of Psycholinguistic Studies, Faculty of English - Poland
16
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Are interpreters better respeakers?
An exploratory study on respeaking competences
Agnieszka Szarkowska - University of Warsaw
Łukasz Dutka - University of Warsaw
Abstract: In this paper we present preliminary data from the research project “Respeaking competences, process, quality”. In the course of respeaking, a number of complex cognitive
processes take place, requiring the use of multiple cognitive resources. The main goal of
the project is to study these processes and to obtain insight into their cognitive intricacies.
In the project, three groups of participants (trainee interpreters, trainee translators
and bilinguals with no background in interpreting or translation) are subjected to a
number of psychophysiological tests: working memory capacity tests (see Baddeley
1986), paraphrasing tests (see Christoffels and de Groot 2004) and respeaking tests.
During the respeaking tests, participants’ brain activity is monitored by EEG (Emotiv),
their eye movements are recorded by an eye tracker, while their computer activity is
examined through screen recording software. These tools serve to understand the respeaking process by analysing cognitive resources used by participants while respeaking. EEG enables us to verify the level of the engagement of different parts of the brain
during respeaking and the lateralisation of brain functions (Fabbro et al. 1990). Eye
tracking helps identify prominent areas which attract the respeaker’s visual attention.
The respoken subtitles created by the study participants undergo a qualitative and
quantitative analysis using the NERstar software (Martínez Pérez 2013).
We hypothesise that interpreting trainees - as opposed to trainee translators - are better predisposed to respeaking owing to a number of similarities between interpreting
and respeaking.
References
Baddeley, Alan D. (1986) Working Memory. Oxford: Oxford University Press.
Christoffels, Ingrid K., Annette M. B. De Groot, Judith F. Kroll (2006) “Memory and language skills in
simultaneous interpreters: The role of expertise and language proficiency”. Journal of Memory and
Language 54 (3), 324-345.
Martínez Pérez, Juan (2013) “NERStar. Speech technology accuracy rate”. Paper presented at the Media for All 5 conference in Dubrovnik, September 2013.
Romero-Fresco, Pablo (2011) Subtitling through Speech Recognition: Respeaking. Manchester: St Jerome.
Romero-Fresco, Pablo (2012) “Respeaking in Translator Training Curricula. Present and Future Prospects”. The Interpreter and Translator Trainer, 6(1), 91-112.
Agnieszka Szarkowska, [email protected]
Assistant Professor - University of Warsaw - Institute of Applied Linguistics - Poland
Łukasz Dutka, [email protected] - University of Warsaw
Institute of Applied Linguistics - Poland
17
5th International Symposium Respeaking, Live Subtitling and Accessibility
Tapping the linguistic competence in the respeaking process:
comparing intralingual paraphrasing done by interpreting trainees,
translation trainees and bilinguals
Agnieszka Chmiel, Agnieszka Lijewska - Adam Mickiewicz University
Łukasz Dutka, Agnieszka Szarkowska - University of Warsaw
Izabela Krejtz, Krzysztof Krejtz - University of Social Sciences and Humanities
Abstract: Respeakers need to master a number of linguistic and technical competences
(Arumí Ribas & Romero Fresco 2008, Romero Fresco 2012, Eugeni 2008). Among them
is the ability to listen and speak at the same time in the same language or to translate the words spoken in a live programme into another language. Other necessary
linguistic skills are the ability to paraphrase and condense the text (Luycks et al. 2010,
Sandrelli 2013) as well as the ability to simultaneously listen to other speakers and to
one’s own voice in order to control the intonation and correctness of the respoken text
(Romero Fresco 2011). Thus, respeaking involves a re-expression skill similar to the one
applied in simultaneous interpreting (Romero-Fresco 2011: 45).
We have conducted a study to test whether due to their interpreting skills, interpreting trainees would perform significantly better in a paraphrasing task understood as a
subcomponent of respeaking. 15 interpreting trainees, 15 translation trainees and 6
non-interpreting and non-translating bilinguals participated in a two-day intensive respeaking course and then in an intralingual paraphrasing study. They were asked to paraphrase three types of sentences. Type 1 included semantic redundancies (which could
be replaced by hyperonyms), type 2 included oral discourse markers (which could be
omitted without sense loss) and type 3 included false starters (which could be normalized). The participants performed the paraphrasing in a delayed condition (respeaking
a given sentence after they have finished listening to it) or in a simultaneous condition
(respeaking a given sentence while listening to it).
In the analysis we looked at group differences in the ear-voice span and correlated the
quantitative data with the quality of the paraphrases and the participants’ memory
scores. We hypothesize that interpreting trainees – as opposed to trainee translators
and bilinguals – will be better predisposed to respeaking thanks a number of similarities between interpreting and respeaking, and also thanks to their well-trained working
memory.
18
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
References
Arumí Ribas, Marta & Romero-Fresco, Pablo (2008) “A Practical Proposal for the Training of Respeakers”. The Journal of Specialised Translation 10, pp. 106-127.
Eugeni, Carlo (2008) Le sous-titrage en direct: aspects théoretiques, professionels et didactiques. Macerata: Eum.
Luyckx, Bieke, Delbeke, Tijs, Van Waes, Luuk, Leijten, Marielle & Remael, Aline (2010) “Live subtitling
with speech recognition. Causes and consequences of text reduction”.
Romero-Fresco, Pablo (2011) Subtitling through Speech Recognition: Respeaking. Manchester: St Jerome.
Romero-Fresco, Pablo (2012) “Respeaking in Translator Training Curricula. Present and Future Prospects”. The Interpreter and Translator Trainer 6(1), pp. 91-112.
Romero-Fresco, Pablo (2013) “Quality and NER in the UK”. Paper presented at the 4th International
Symposium on Live Subtitling in Barcelona, March 2013.
Sandrelli, Annalisa (2013) “Reduction strategies and accuracy rate in live subtitling of weather forecasts: a case study”. Paper presented at the 4th International Symposium on Live Subtitling in Barcelona, March 2013.
Agnieszka Chmiel, [email protected]
Assistant Professor
Adam Mickiewicz University - Faculty of English - Department of Translation Studies - Poland
Agnieszka Lijewska, [email protected]
Assistant Professor
Adam Mickiewicz University - Faculty of English - Department of Psycholinguistic Studies - Poland
Łukasz Dutka, [email protected]
Consultant
University of Warsaw - Institute of Applied Linguistics - Poland
Agnieszka Szarkowska, [email protected]
Assistant Professor
University of Warsaw - Institute of Applied Linguistics - Poland
Izabela Krejtz, [email protected]
Assistant Professor
University of Social Sciences and Humanities
Interdisciplinary Center for Applied Cognitive Studies - Poland
Krzysztof Krejtz, [email protected]
Assistant Professor
University of Social Sciences and Humanities - Poland
19
5th International Symposium Respeaking, Live Subtitling and Accessibility
THEME 3
NEW APPLICATIONS OF TECHNOLOGY
Pushing forward the frontiers of access technology
Andrew Lambourne - Screen Systems
Abstract: Tools and techniques for production of access subtitles have been in a continuous state of development since the early days of such services in the 1980s. Respeaking is the latest in a series of techniques which have been used to solve the difficult
problem of providing subtitles for real-time broadcasts: other methods have included
Stenographic transcription systems, ordinary keyboards, multiplexed dual keyboards,
syllabic chord keyboards and automatic speech recognition. Fundamentally all these
systems face the same challenge - how to transcribe and present real-time unscripted
speech with minimum delay and maximum quality?
This paper describes new tools and techniques targeted at reducing the toughest challenges faced by a real-time subtitler – be they a trained respeaker or someone using
any other real-time transcription approach – and explores not only how these challenges can be minimised, but what further practical research could beneficially be conducted. Missing vocabulary, homophones, out-of-language words, number grammars,
corrections, and throughput delay are all examples of problems which can lead to a reduction in quality. We examine practical ways in which these problems can be reduced,
and suggest further research directions.
Additionally, we look at two opportunities relating to IP delivery. Firstly, the HbbTV system, which permits traditional broadcast material to be supplemented by IP-delivered
content. One application for this is in personalisation, and new proposals for delivering
audio could open up possibilities for assistance to hard-of-hearing viewers by enabling
them to select and personalise the mix of audio they receive. This would not only help
to solve the problem of intrusive music beneath dialogue, but would actually assist in
the production of access subtitles by making it easier to transcribe programmes for
which no script was available. Secondly, we look at new tools which have been developed to enable live webcasts to carry high quality properly synchronised live captions.
Andrew Lambourne, [email protected]
Business Development Director - Screen System - United Kingdom
20
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Evaluating performances of Automatic Speech Recognition
for Assisted Subtitling
Carlo Aliprandi - Synthema Srl
Abstract: The contribution will address recent advances in Automatic Speech Recognition (ASR), presenting and evaluating a new Speaker Independent ASR technology that
has been developed in the SAVAS project.
The demand for Access Services has quickly grown in the years, mainly due to National
and International laws. This trend is expected to consolidate for subtitling in particular.
However, the path of manual subtitling is no longer feasible, due to increased costs and
reduced production times.
In this contribution we will focus on an emerging technique, Assisted Subtitling. Assisted Subtitling consists on the application of Automatic Speech Recognition (ASR) to generate transcripts of programs and to use the transcripts as the basis for subtitles. We
will present recent advances, based on a new ASR technology that has been that has
been developed in the SAVAS project. We will describe three systems based on SAVAS:
• S .Scribe! is the system for pre-recorded content, capable of automatically transcribing audio and video content into time-aligned subtitles.
• S .Align! is the system for automatic synchronisation, capable of auto-timing audio
and video content to script and subtitles.
• S .Live! is the first-of-a-kind Speaker Independent Transcription System, with realtime
performances for online subtitling.
Eventually, we will give an overview of the methodology, quality features and metrics
employed to evaluate the SAVAS technology. The considered features include parameters linked to subtitle layout, duration and text editing such as speaker colours, delay,
persistence on screen, capitalization, punctuation, splitting or the use of acronyms,
apostrophes and numerals. An extended version of the NER model that also considers
errors related to the automatic allocation of speaker colours, time stamps and segmentations will be described. The quality performance achieved by the SAVAS applications
will be presented together with results of experiment testing the productivity gains
achieved in post-editing automatic subtitles. Finally, the main challenges that still remain to be solved will be highlighted.
21
5th International Symposium Respeaking, Live Subtitling and Accessibility
References
Aliprandi, Carlo & Verruso, Fabrizio (2006), “Tecnologie del Linguaggio Naturale e sottotitolazione multilingue diretta”, inTRAlinea, Special Issue: Respeaking, ISSN 1827-000X, 2006 http://www.intralinea.
it/specials/respeaking/ita_more.php?id=453_0_41_0_M
Aliprandi, Carlo (2013) “Speech Recognition with a focus for quality Subtitling of minor languages”, 4th
Symposium on Live Subtitling, March 2013, Barcelona.
Aliprandi, Carlo (2013) “The Savas project – Speech Recognition for Respeaking and Live Subtitling”.
In Carlo Eugeni & Luigi Zambelli (a cura di) Respeaking. On-line. Special Issue n. 1. June 2013. www.
accademia-aliprandi.it pp. 117-119.
Álvarez, Aitor, Aliprandi, Carlo, Gallucci, Isabella, Piccinini, Nicola, Raffaelli, Matteo, del Pozo, Arantza,
Cassaca, Renato, Neto, Joao, Mendes, Carlos, Viveiros, Marcio forthcoming, “Automating Live and
Batch Subtitling of Multimedia Contents for several European Languages”, Multimedia Tools and Applications (MTAP), An International Journal from Springer.
Álvarez, Aitor, Aliprandi, Carlo, Gallucci, Isabella, Piccinini, Nicola, Raffaelli, Matteo, del Pozo, Arantza,
Cassaca, Renato, Neto, Joao, Mendes, Carlos, Viveiros, Marcio (2014) “Assisted Subtitling: a new opportunity for Access Services”, Paper presented at the International Broadcasting Conference 2014
(IBC2014), September 10-15 2014, Amsterdam, The Netherlands.
Arantza del Pozo, Carlo Aliprandi, Aitor Álvarez, Carlos Mendes, Joao Neto, Sérgio Paulo, Nicola Piccinini, Matteo Raffaelli (2014) “SAVAS: Collecting, Annotating and Sharing Audiovisual Language Resources for Automatic Subtitling”, International Conference on Language Resources and Evaluation
(LREC2014), May 2014, Reykjavik.
C. Aliprandi, I. Gallucci, N. Piccinini, M. Raffaelli, A. del Pozo, A. Álvarez, R. Cassaca, J. Neto, C. Mendes,
M. Viveiros,
Carlo Aliprandi, Cristina Scudellari, Isabella Gallucci, Nicola Piccinini, Matteo Raffaelli, Arantza del Pozo,
Aitor Álvarez, Haritz Arzelus, Renato Cassaca, Tiago Luis, Joao Neto, Carlos Mendes, Sérgio Paulo, Marcio Viveiros (2104) “Automatic Live Subtitling: state of the art, expectations and current trends”, NAB
Broadcast Engineering Conference, April 5-10 2014, Las Vegas.
Eugeni, Carlo (2006). “Introduzione al rispeakeraggio televisivo”, inTRAlinea, Special Issue: Respeaking, ISSN 1827-000X, 2006, URL: http://www.intralinea.it/specials/respeaking/ita_more.
php?id=444_0_41_0_M
Lambourne, Andrew (2006) “Subtitle respeaking: A new skill for a new age”, inTRAlinea, Special Issue:
Respeaking, ISSN 1827-000X, 2006, URL: http://www.intralinea.it/specials/respeaking/eng_more.
php?id=447_0_41_0.
Martínez, Juan (2103) “Tools for assessing Accuracy Rate of Live Subtitles”, 4th Symposium on Live
Subtitling, March 2013, Barcelona.
Carlo Aliprandi, [email protected]
Speech Solutions Manager - SyNTHEMA Srl - Italy
22
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Beyond PC-based speaker-adapted speech recognition: Exploring the potential and
challenges of cloud-based speaker-independent speech recognition systems
Andreas Søeborg Kirkedal - Copenhagen Business School
Julián Zapata - University of Ottawa
Abstract: Automatic speech recognition (ASR) technology gives a computer system the
ability to recognize and process human speech. Developed and used for several decades, speech-enabled systems and devices are proven better than keyboard-based interfaces for tasks for which full natural language communication is useful, such as medical dictation, translation dictation, respeaking and (live) subtitling. Our presentation
will describe a pilot experiment aiming to assess and compare the performance of two
off-the-shelf ASR systems: a speaker-adapted (SA) system on a Windows laptop, and a
speaker-independent (SI) system on an Android tablet. The experiment was carried out
with speakers of three major languages and with different linguistic profiles: non-native English (Indian-accented and Spanish-accented) speakers, non-native French (Russian-accented and Spanish-accented) speakers, and native Spanish (Iberian-accented
and Latin-American-accented) speakers. Five tasks were involved in this experiment:
(1) typing, (2) reading aloud, (3) dictating with a commercial PC-based SA ASR system
on a laptop, (4) dictating with a commercial cloud-based SI ASR system on a tablet, and
(5) manually fixing ASR errors using a laptop keyboard. The focus of this experiment
was on task times and ASR performance (i.e., word accuracy rates) across various users, languages and devices. No translation or respeaking task was involved in this pilot
experiment. However, results highlight the potential advantages and drawbacks of a
particular ASR approach (i.e., PC-based SA vs. cloud-based SI) for specific users, who
have different accents and levels of proficiency in their working languages, and who
may have different levels of competence and experience in translation, respeaking and
(live) subtitling, and with ASR technology. Lastly, we enumerate several innovative areas for future work.
References
Garcia-Martínez, Mercedes, Singla, Karan, Tammewar, Aniruddha, Mesa-Lao, Bartolomé, Thakur, Ankita, Anusuya, M. A., Carl, Michael & Bangalore, Srinivas (2014) “SEECAT: ASR & Eye-tracking Enabled
Computer-Assisted Translation”. In Proceedings of the 17th Annual Conference of the European Association for Machine Translation , pp. 81-88.
Rodríguez, Luis, Reddy, Aarthy & Rose, Richard (2012) “Efficient Integration of Translation and Speech
Models in Dictation Based Machine Aided Human Translation” In Proceedings of the IEEE 2012 International Conference on Acoustics, Speech, and Signal Processing, pp. 4949-4952.
23
5th International Symposium Respeaking, Live Subtitling and Accessibility
Romero-Fresco, Pablo (2011) Subtitling Through Speech Recognition: Respeaking. Manchester: St. Jerome.
Zapata, Julián (2014) “Exploring Multimodality for Translator-Computer Interaction”. In Proceedings of
the 16th International Conference on Multimodal Interaction, pp. 339-343.
Andreas Søeborg Kirkedal, [email protected] - PhD Fellow - Copenhagen Business School
Spoken NLP Researcher - Mirsk Digital ApS - Denmark
Julián Zapata, [email protected] - PhD (candidate), Translation Studies, School of Translation and
Interpretation - University of Ottawa - Canada
24
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Transcribing documentaries: can respeaking be used efficiently?
Lukasz Daniluk - University of Roehampton
Anna Matamala - Universitat Autònoma de Barcelona
Pablo Romero-Fresco - University of Roehampton
Abstract: Respeaking is increasingly used to offer live intralingual subtitles for the deaf and
hard-of-hearing (Romero-Fresco 2011), hence guaranteeing media accessibility to a wider
section of the population. However, our hypothesis is that respeaking could be efficiently
used for other tasks, such as transcribing non-fictional products that do not have an already
created script. This could benefit transcribers in their jobs but would also allow to (semi-)automatically generate an input for machine translation processes, to name just one example.
This paper aims to present the results of an experiment in which three scenarios were
compared:
a) manual transcription
b) revision of an automatically generated transcription
c) respeaking
Three comparable 4-minute clips were selected, and 15 professional English transcribers were asked to do three different tasks: manually transcribe an excerpt, revise an
automatically generated transcription (ASR), and respeak another excerpt. The order
of the tasks and excerpt was randomized. The time spent in each of the tasks was monitorized, and data on the perceived effort and the transcriber experience were collected. This paper will describe the experimental set-up as well as the results of this test.
This experiment is part of the ALST project (2013-2015), funded by the Spanish Ministry of Economy (FFI2012-31024), which aims to investigate the application of speech
and translation technologies in the audio description of films and in the voice-over of
non-fictional genres. (Matamala et al. 2013)
References
Matamala, Anna, Fernández-Torné, Anna, Ortiz-Boix, Carla (2013) “Enhancing sensorial and linguistic
accessibility: further developments in the TECNACC and ALST projects”. 5th International Conference
Media for All. Audiovisual Translation: Expanding Borders. Dubrovnik, 25-27/09/13.
Romero-Fresco, Pablo (2011) Subtitling through speech recognition: Respeaking. Manchester: St. Jerome.
Lukasz Daniluk, [email protected] - Ph. D. Student
University of Roehampton - United Kingdom
Anna Matamala, [email protected] - Full-time Senior Lecturer
Universitat Autònoma de Barcelona - Spain
Pablo Romero-Fresco, P. [email protected] - Reader in Audiovisual Translation and
Filmmaking
University of Roehampton - Department of Media, Culture and Language - United Kingdom
25
5th International Symposium Respeaking, Live Subtitling and Accessibility
THEME 4
RESPEAKING AS A PROFESSIONAL OPPORTUNITY FOR THE BLIND
Teaching live subtitling for the deaf through respeaking the potential of blind respeakers and paraplegic live editors
Carlo Eugeni - Roma Tre University
Abstract: Live subtitling through respeaking is still one of the most sectorial professions in
Audiovisual Translation (Orero 2006). Teaching respeaking is all the more rare, especially at
academic level (Remael & van der Veer 2006, Romero Fresco 2011). Among the reasons for
this to happen, the organization of such courses in universities is hindered by a threefold
lack: of the minimum technological requirements; of teachers being able to teach it; and
of a real interest from both potential respeakers and the job market (Baaring 2006, Lambourne 2006). In this framework, disabled people are normally considered as the end users
of the live subtitling service and, in terms of accessibility, they just receive the message and
never give a real contribution in providing the service (Eugeni 2008).
To try and bridge some of these gaps, a course in live subtitling through respeaking and live
editing was organized by the association of disabled people “Cooperativa Progetto il Seme
Onlus” based in Milan. The project was aimed at introducing disabled people in the job
market, by giving them the possibility to exploit the skills they normally use as daily users of
the PC workstation. In particular, blind people have been trained to professional respeaking
and paraplegic people to professional live editing for three months by professional trainers.
The trainees were finally tested by the Italian Academy of Multimedia Writing and Communication “Giuseppe Aliprandi - Flaviano Rodriguez”.
In this speech, the didactic solutions that have been adopted during the course will be
analyzed on the basis of the needs and expectations of both blind people and paraplegic
people before the course itself. After that, their attitudes as trainees in respeaking and live
editing will be discussed in view of perfecting the didactics of live subtitling in general (Russello 2013). Finally, the result of the tests will be analyzed and overall considerations will be
made about the skills and competences of disabled people as professional live subtitlers.
26
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
References
Baaring, I. (2006) “Respeaking-based online subtitling in Denmark”. In Eugeni, C. & G. Mack (eds.)
InTRAlinea Special Issue: http://www.intralinea.org/specials/article/1685
Eugeni, C. (2008) Le sous-titrage en direct : aspects théoriques, professionnels et didactiques. Macerata: EUM.
Lambourne, A. (2006) “Subtitle respeaking: A new skill for a new age”. In Eugeni, C. & G. Mack (eds.)
InTRAlinea Special Issue: http://www.intralinea.org/specials/article/1686
Orero, P. (2006) “Real-time subtitling in Spain”. In Eugeni, C. & G. Mack (eds InTRAlinea Special Issue:
http://www.intralinea.org/specials/article/1689
Remael, A. & B. van der Veer (2006) “Real-Time Subtitling in Flanders: Needs and Teaching”. In Eugeni,
C. & G. Mack (eds.) InTRAlinea Special Issue: http://www.intralinea.org/specials/article/1702
Romero-Fresco, P. (2011) Subtitling through Speech Recognition: Respeaking. Manchester: St Jerome.
Russello, C. (2013) “Aspetti didattici”. In Eugeni, C. & L. Zambelli (eds.) Respeaking. Specializzazione
online. www.accademia-aliprandi.it pp. 44-53.
Carlo Eugeni, [email protected]
Università Roma Tre - Italy
27
5th International Symposium Respeaking, Live Subtitling and Accessibility
Vocational Training for Blind Speech-to-Text Reporters
Robin Ribback - VerbaVoice
Abstract: The project set up by a consortium of international research and technology
partners aims to set up vocational training for blind or partially sighted people in different
European countries. The training on the use of speech recognition technology is set to
provide live and remote speech-to-text services for hard of hearing, deaf and non-native
people in the fields of inclusive education, live events, political debates or broadcast media. It enables twofold inclusion for visual and hearing impaired people. A remote interpreting system permits blind interpreters to control their output using a braille display
connected to the cloud-based online platform both interpreters and clients are using. The
multinational project team consists of the social enterprise VerbaVoice GmbH (accessible
technologies for deaf and hard of hearing people, Germany), the Munich University of Applied Languages (SDI Munich, Germany), the BFW Würzburg (vocational training institute
for blind and partially sighted people, Germany), GIB-BLWG Nuremberg (institute for the
communication of hearing impaired people, Germany), the Universidad Autònoma Barcelona (UAB, Spain), the University of Roehampton (UoR, UK), subtitling company SUBTI (UK)
and multimedia service provider SWISS TXT (Switzerland). More partners are most likely to
join in. The project scope includes ICT-based teaching, training and assessment practices
particularly for audiences of disadvantaged backgrounds facilitating transitions from education to the world of work for disadvantaged learners. The training program will be based
on previous experiences made during training a training program for (non-blind) speechto-text-reporters held at the University of Applied Languages Munich in 2014 and one for
visually impaired speech-to-text reporters organized by VerbaVoice and BFW Würzburg in
2015. The methodology, curriculum and management of the project will be adapted to its
multinational character.
28
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
References
Linder, Gion & Martínez, Juan (2013) “Access the Accuracy for Live Subtitles in a Quick and Effective Way.” ITU Workshop on Making Media Accessible to all: the options and economics. Electronic
version available at: http://www.itu.int/en/ITU-T/Workshops-and-Seminars/mmaa/201310/Pages/
programme.aspx
Martínez Pérez, Juan & Romero-Fresco, Pablo (2011) “Apples and Oranges: A Comparative Study of
Accuracy in Respeaking”. 4th International Media for All Conference, Imperial College London.
Mikul, Chris (2014) “Caption Quality: Approaches to Standards and Measurement.” Electronic version
available at: http://www.mediaaccess.org.au/research-policy/white-papers/caption-quality-international-approaches-to-standards-and-measurement
Romero-Fresco, Pablo (2011) “The NER Model to Assess Accuracy in Respeaking”. ITU Workshop on
Telecommunications Relay Services for Persons with Disabilities. Electronic version available at: http://
www.itu.int/ITU-T/worksem/trs/programme.html
Robin Ribback, [email protected]
CIO VerbaVoice GmbH - Germany
29
5th International Symposium Respeaking, Live Subtitling and Accessibility
ROUND TABLE: FOCUS ON ITALY
Proposta di aggiornamento delle legislazioni italiane relative all’accessibilità
dei programmi delle emittenti televisive per i cittadini sordi
Giacomo Pirelli - On A.I.R. - Intersteno Italia
Luca Iacovino - Presidente ANS
Abstract: L’attuale Contratto di Servizio per la sottotitolazione televisiva lega la RAI alla
Commissione di Vigilanza Rai, ma tale vincolo non si estende alle altre emittenti televisive. Tale situazione non tiene conto della liberalizzazione delle bande televisive e del
conseguente numero maggiore di interlocutori.
Con l’occasione del rinnovo di tale contratto alla sua scadenza triennale, sarebbe opportuno aggiornarlo ed estenderlo alle altre emittenti televisive. Il nuovo Contratto di
Servizio e l’aggiornamento del Decreto Legislativo 31 luglio 2005, n. 177 Testo unico
della radiotelevisione, dovrebbero specificare le esigenze delle persone sorde riguardo
una fruizione accessibile dei programmi televisivi. Tale Contratto dovrebbe coinvolgere
non solo la RAI, ma anche le reti televisive private, come Mediaset, LA7, Sky, ecc., per
una piena applicabilità dei diritti fondamentali di accesso all’informazione da parte degli utenti sordi. Ciò sarebbe in conformità all’articolo 30 della Convenzione ONU sui
Diritti Umani delle Persone con Disabilità, ratificata in Italia nel 2009, ed alle normative
di diversi paesi europei, in particolare Regno Unito e Francia.
Le emittenti televisive sarebbero così obbligate a sottotitolare una percentuale dei
programmi, per esempio al minimo dal 60%. Esse dovrebbero inoltre definire con le
associazioni delle linee guida sulla qualità dei sottotitoli nei programmi televisivi, sia in
forma preregistrata che in tempo reale, per tutti i canali televisivi, seguendo gli esempi
delle migliori prassi (per esempio le linee guida approvate dall’OFCOM per la BBC e tutti
gli altri canali televisivi statali e privati nel Regno Unito e quelle approvate dal Conseil
Supérieur de l’Audiovisuel in Francia nel dicembre 2011 per tutti i canali).
30
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
References
ITALIA:
articoli 4 e 45 nel Decreto Legislativo 31 luglio 2005, n. 177 Testo unico della radiotelevisione.
Schema di contratto di servizio tra il Ministero dello sviluppo economico e la RAI-Radiotelevisione Italiana Spa, per il triennio 2013-2015
http://www.senato.it/leg/17/BGT/Schede/docnonleg/25703.htm
FRANCIA:
Article 74 - LOI n° 2005-102 du 11 février 2005 pour l’égalité des droits et des chances, la participation
et la citoyenneté des personnes handicapées
Charte relative a la qualité du sous-titrage á destination des personnes sourdes ou malentendantes
(Linee Guida di sottotitolazione televisiva), Conseil Supérieur de L’Audiovisuel 2011: http://www.
csa.fr/Espace-juridique/Chartes/Charte-relative-a-la-qualite-du-sous-titrage-a-destination-des-personnes-sourdes-ou-malentendantes-Decembre-2011
REGNO UNITO:
Section 303 - Communications Act 2003
Code on Television Access Standard: http://stakeholders.ofcom.org.uk/binaries/broadcast/other-codes/ctas.pdf
Linee Guida di sottotitolazione televisiva dell’OFCOM (Office of Communication)
http://www.ofcom.org.uk/static/archive/itc/itc_publications/codes_guidance/standards_for_subtitling/subtitling_1.asp.html
Aggiornamento linee guida OFCOM (2013):
http://consumers.ofcom.org.uk/2013/05/improving-the-quality-of-live-tv-subtitles/
Giacomo Pirelli, [email protected]
onA.I.R. - Intersteno Italia - Italy
Luca Iacovino, [email protected]
Presidente ANS - Italy
31
5th International Symposium Respeaking, Live Subtitling and Accessibility
La Meeting Industry in Italia oggi può offrire attraverso il respeaking nuove
opportunità sia per gli organizzatori di congressi ed eventi
che per i professionisti della documentazione scritta
Samuel Zambelli - Studio Acta
Abstract: Il V° Simposio Internazionale sul respeaking, sottotitolazione in diretta e accessibilità offre agli operatori della comunicazione/documentazione scritta un’importante occasione per riflettere sullo stato dell’arte della ricerca e, nel contempo, fornisce elementi
di analisi delle problematiche connesse con un ulteriore sviluppo dei software di riconoscimento del parlato e, conseguentemente, dei servizi di produzione di testi scritti.
Nel corso della mia attività professionale ho potuto seguire l’evolversi degli strumenti utilizzati dai professionisti nei diversi ambiti operativi: un repertorio di sussidi sempre più
sofisticati, che hanno comportato significative ricadute a livello operativo.
La rapidissima evoluzione tecnologica ha avuto anche forti implicazioni sul fronte della
formazione permettendo a soggetti dotati di specifiche competenze e riconosciute abilità di tipo tecnico di proporre servizi innovativi nell’area della documentazione scritta.
Una carente azione di informazione e sensibilizzazione di quanti, ai diversi livelli, possono
trarre vantaggi dall’introduzione delle nuove tecniche e software di produzione del testo
scritto, ha pesantemente condizionato il loro sviluppo mentre – anche grazie agli spunti
che emergeranno in questo convegno - si possono sicuramente prefigurare significativi
sviluppi nell’uso professionale dei software di riconoscimento del parlato e, in particolare,
del respeaking, ovvero della tecnica di produzione del testo scritto attraverso l’intermediazione di un professionista.
Mi limiterò a proporre qualche riflessione unicamente sulla tecnica del respeaking nella
sua applicazione nel mondo congressuale. Non mi occuperò dei software “speaker indipendent” che, a mio parere, se possono trovare applicazione in aree ben limitate della documentazione scritta (refertazione medica, trascrizione brevi messaggi, etc.), non hanno
ancora raggiunto un grado di precisione che permetta di fornire dei servizi di trascrizione,
né tanto meno di resocontazione, con un accettabile standard qualitativo.
Gli stimoli che questo Simposio internazionale ci propone necessitano quindi di un’attenta e non rituale analisi, sia sul piano accademico/didattico, sia sul piano operativo, dei
multiformi aspetti del respeaking come tecnica privilegiata per la resocontazione e per la
sottotitolatura. È questa un’importante occasione per inaugurare un rapporto di sinergica
collaborazione tra il mondo della ricerca e gli attori privilegiati dell’area congressuale e
convegnistica in senso lato.
Samuel Zambelli, [email protected]
Sales & Marketing Manager - Studio Acta - Italy
32
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
La telefonata sottotitolata
Lorenzo Di Ciaccio - Pedius
Abstract: La sottotitolazione ha da sempre avuto impiego nell’ambito dell’informazione e dell’intrattenimento. Partendo dal cinema fino alla televisione, il sottotitolo ha
agevolato la fruizione a tutta la fascia della popolazione con problemi uditivi. Oggi grazie alle nuove tecnologie è possibile introdurre la sottotitolazione in tempo reale anche
all’interno delle telefonate agevolando la comunicazione tra persone sorde e persone
udenti.
Pedius consente alle persone sorde di telefonare grazie alle le tecnologie di riconoscimento del parlato e di sintesi vocale.
Attraverso l’applicazione mobile gli utenti possono avviare una telefonata verso un numero telefonico. Durante la telefonata l’utente scrive come in una normale chat e una
voce artificiale pronuncerà i messaggi. In tempo reale la risposta vocale della persona
chiamata viene trasformata in testo ed inviata all’utente.
Lorenzo Di Ciaccio, [email protected]
CEO Pedius - Italy
33
5th International Symposium Respeaking, Live Subtitling and Accessibility
New frontiers of respeaking: live subtitles delivered
on smartphones, tablets and electronic glasses
Saveria Arma - Artis-Project / CulturAbile / MovieReading
Abstract: Subtitling through respeaking is a relatively well-known technique for live TV
programs and news. However, when it comes to conference respeaking, literature and
practices become much less abundant. This contribution represents an introduction
to respeaking-based live subtitling for the deaf and the hard of hearing in a conference setting. In particular, we will focus on new technical and technological solutions
for providing and reading respeaking-based subtitles: from classical “PC-to-beamer”
configuration with subtitles produced and projected by an operator onsite, to remote
respeaking operated by a team of geographically distant respeakers delivered on individual devices (smartphones, tablets and electronic glasses) in one or more languages.
The individual distribution of accessible content is becoming one of the major frontiers
in accessibility as a whole, not only intended as a tool for including people with sensory
impairment but - more and more often - as a cutting-edge cost-effective solution for
conferences and events attended by people who speak different languages or do not
have sufficient knowledge of the working language (this solution is also being considered as an alternative to traditional simultaneous interpreting). We will present a new
solution for producing real-time respeaking-based subtitles (onsite/remote captioning)
and to distribute them on multiple platforms, including electronic augmented-reality
glasses.
We will draw on preliminary evidence and results in terms of subtitles production and
delivery processes, as well as of feedback from the hearing and hearing-impaired audience as to the type of device, readability of subtitles and “social” function.
References
Eugeni, Carlo & Mack, Gabriele (eds) (2006) Intralinea, Respeaking Special Issue. Proceedings of the
First International Seminar on real-time intralingual subtitling, University of Bologna - Forlì Campus.
17th November 2006. Available at: http://www.intralinea.org/specials/respeaking
Romero-Fresco, Pablo (2011) Subtitling through Speech Recognition: Respeaking. Manchester: St.
Jerome Publishing.
Saveria Arma, [email protected]
Artis-Project / CulturAbile / MovieReading - Italy
34
5th In
lity
5th International Symposium Respeaking, Live Subtitling and Accessibility
Elenco degli oratori - List of speakers
Name
Page
Aliprandi, Carlo
18-19
Arma, Saveria
30
Chmiel, Agnieszka
15-16
Daniluk, Lukasz,
21-22
Di Ciaccio, Lorenzo
29
Dutka, Łukasz
14, 15-16
Eugeni, Carlo
23-24
Iacovino, Luca
26-27
Krejtz, Izabela
15-16
Krejtz, Krzysztof
15-16
Lambourne, Andrew
17
Lijewska, Agnieszka
15-16
Martínez Pérez, Juan
12
Matamala, Anna
21-22
Moores, Zoe
11
Pirelli, Giacomo
26-27
Pražák, Aleš
10
Ribback, Robin
25
Romero-Fresco, Pablo
9, 21-22
Søeborg Kirkedal, Andreas
20
Stachowiak, Katarzyna
13
Szarkowska, Agnieszka
14, 15-16
Widmer-Iliescu, Roxana
8
Zambelli, Samuel
28
Zapata, Julián
20
35
5th International Symposium Respeaking, Live Subtitling and Accessibility
36