FINISHED TRANSCRIPT ITU - ANDREA SAKS MARCH 12, 2012 11:00 P.M. CST 4TH FG AVA MEETING NEW DELHI, INDIA Services provided by: Caption First, Inc. P.O. Box 3066 Monument, CO 80132 1-877-825-5234 +001-719-481-9835 www.captionfirst.com *** This text is being provided in a rough draft format. Communication Access Realtime Translation (CART) is provided in order to facilitate communication accessibility and may not be a totally verbatim record of the proceedings. *** >> OPERATOR: The broadcast is now starting. All attendees are in listen only mode. >> PETER OLAF LOOMS: I'm testing to see if the captioners in the U.S. can hear us. If you can hear us we hope you can indicate through the captions. Great. Okay. I suggest we start in 60 seconds. Okay. Good. Good morning, everybody. Welcome to the Fourth Meeting of the ITU-T Focus Group on Audiovisual Media Accessibility. The time in India is 10:15. And the time in Europe is 5:45 in the morning. And I think we will have a number of participants not only in the room but also six so far online. So to both those of you in the room and those taking part through GoToWebinar meeting, very well, you're very welcome. I would like to start by thanking the hosts in India the Centre for Internet Society and ITU APT, the foundation of India. I would also like to ask first Mr. Satya Gupta to say a few words on behalf of his organisation. And Mr. Gupta has been kind enough to help us with the infrastructure for the wireless Internet. >> Thank you, good morning, everyone. And thanks a lot. We are really grateful to ITU to have their meeting on this very important subject in India on our invitation. So first of all, let me admit that I had come here to be on the other side of the table. I never expected that I would be called upon to say a few words actually. So I'm totally unprepared in fact. But it's a great honour and privilege to be a part of this meeting. And as you know, the -- in New Delhi things start a bit late actually. So maybe we'll get more participants here. But some of them are already logged in through the Internet. So I'm not actually planning to speak much on this occasion other than thanking the ITU to choose New Delhi for this meeting. And also the CIS, Centre for Internet Society who actually volunteered to host this on behalf of the Government of India and then my own organisation the ITU APT Foundation of India for whom the Secretary-General is also there Mr. Anil Prakash. And I would like Anil to say a few words actually than me. >> PETER OLAF LOOMS: I think we should start with the Executive Director of the Center for Internet and Society, Mr. Sunil Abraham and a personal thank you to you and your team for all the efforts you've put into arranging for us to hold this meeting in India. >> SUNIL ABRAHAM: Thank you. Very warm good morning to everybody gathered here in this room. And also those participating in this meeting online. It is indeed a privilege for a four-year-old research organisation to have opportunities like this to collaborate with a very august international body like the International Telecommunication Union. Accessibility is very important for us at the Center for Internet and Society. At the national level the recent developments that are very important when it comes to accessibility of the Internet, not broadcast media the Department of Information Technology is at the verge of sending to the committee of secretaries the National Electronic Accessibility Policy, which not only covers WCAG for the Internet but other accessibility standards for electronic infrastructure such as ATMs. In Indian I must also apologize is Mr. Gupta has just done that we seem incapable of starting things on time and also incapable of providing proper bandwidth and connectivity. This is the nation's capital. And this is one of the most premiere venues for holding international events members of international business and Government frequent this very often but inspite of our best efforts we haven't been able to provide broadband connectivity and have had to depend on Mr. Gupta's personal connectivity. Hopefully by tea time our friends here will do a better job and our colleagues across the world will have better reception just to highlight our own condition, 13 million people online only 100 million people who have ever been online are using cyber cafes et cetera in a country of 1.2 billion people with one-third of the population illiterate and 6% of the population at least disabled. Also an aging population so text-to-voice, voice-to-text accessibility of Internet media and accessibility of broadcast media is critical for this nation and therefore kind sirs and madams, we are very pleased you have come here and hopefully the expertise you share with us will be implemented through practical action in this country. Thank you so much. >> PETER OLAF LOOMS: Thank you very much for those very kind words. You mentioned the current initiatives on Internet accessibility. Ten days ago I was at a policy meeting on digital literacy and the whole business of making sure that people can become a part of the Information Society to look at the haves and the have nots. And it was interesting that digital television or television in general was seen as a means of outreach. A way of actually helping to reach out to those who are offliners, I think that's the term that was used. To give an example from the United Kingdom, 11% of the population are still offline in the UK. So half of those have serious disabilities. And at the meeting there was -- mention was made of recent research which looks at not just the ability to take part in the working society but also the economic impact it has for people who are offliners. And I found it interesting that those who don't have access to the Internet, for example, are paying more in their everyday lives than those who do. And the difference was far more than the cost of actually getting online. So accessibility can also be seen not just as a right, as a way of making sure that people can enjoy the same thing as the more fortunate fellow citizens. But also from an economic perspective, by making sure that we do not have a Digital Divide we do look towards digital media literacy accessibility can be seen as the first step the prerequisite in order to actually get started. We're looking in this particular Focus Group at how do we make things accessible. And the idea is not to repeat all of the things which we have done in the past. But to focus our attention on areas where we need to do something about legislation, about regulation, and about standards. Because afterall, that's what the ITU is about. It's a global entity to actually make sure that when we pick up a phone, that we can actually ring anywhere we like in the world. When we exchange television programmes, we can do so without running into difficulties. Or if we want to access a social media of some kind, that it works. And we also have the means to actually get online and make use of things like Facebook or Twitter or other kinds of social media. We have until the end of this year to prepare a roadmap of actions. To say what is actually needed in terms of legislation, regulation, and standards to make sure that that agenda becomes a reality. In order to do so, we have a very simple strategic model. We've started by looking at a vision. What do we mean by accessible media? In particular what do we mean by accessible digital media? So we need to know where it is we want to be. And what we have been looking at for the last three or four months is where are we at present? What is the current situation? What can we do? And in particular to identify barriers and obstacles. And these may be legal. They may have something to do with economics. We are, afterall, in the midst of a global economic well in some parts of the global economy a global recession or at least a period of lower economic growth. I noticed that the figures for India also indicated the economic growth isn't as high as it has been just a few years ago. So we need to establish what are the major obstacles or barriers to turning that vision into a reality. And that's why we've been doing careful analysis in terms of different platforms. Clearly we've been looking at digital television because that's one of the major platforms in terms of the number of users worldwide. But we've also been looking at IPTV. We've also been looking at mobile services. We've also been looking at social media. And in fact, we've been also looking at things like video games. Because for certain demographics in particular for young people, video games are part and parcel of their identity. And if young people are excluded from playing the same games as their peers, then we have a different kind of Accessibility Challenge. So on the one hand we're looking at some of the platforms on which we deliver content. And then we've got a number of groups, four main groups, looking at Access Services. The means by which we can make different kinds of audiovisual media more accessible. I'm sure those of you in the room from India are familiar with Hollywood movies, Hollywood films. And these are seen around the world. And outside of India frequently with subtitles or captions in English. So India has a long tradition of producing captioning from one language to another but there are other ways in which we can use technologies as we heard from Mr. Abraham about the use of text-to-speech. Being able to use speech synthesis to help certain groups with visual impairments but the other way around, too, speech recognition to drive various services to be able to produce subtitles or captioning. And over and above that, of course, things like sign language or visual signing. For those for whom sign language is their mother tongue. So what we are doing in this group is coming up with what is considered by the key stakeholders as a vision for accessible media. Looking at the present to look at where we are at. Identifying barriers. And then the third step is listing the kind of actions that will be necessary in order to break down those barriers, to overcome them. And many of the things that we have identified are not necessarily within the mandate of the ITU. One of the common challenges which we see in other fields is people's lack of awareness or unfamiliarity with what is already possible. What already can be done. And actually helping key stakeholders understand that when we talk about Access Services they are in some cases cheaper than people may have feared. But having an informed debate about what the nature of the challenge is, what the options are, what they cost. And finally, by the end of the year we hope to put together a roadmap of actions, specific recommendations which can reenforce what's already being done in the fields of legislation, regulation and standards when it comes to making audiovisual media accessible. So that's what we're looking at. And what we have arranged to do back in September was to have a total of 11 Working Groups to look at these various challenges. And one of these actually looks at a kind of metacommunication. The way in which we are communicating as a Focus Group. So probably one of the key groups we have is to look at the accessibility of this particular process and that is why we have made sure that wherever possible we will have captioning or subtitles for the meetings. We are working to make sure that virtual participants can take part also if they have different kinds of disability. So if we are working on audiovisual media accessibility I think it's crucial that we practice what we preach. So that's just the potted version, the short version, to explain what this Focus Group is about. Yes. And in the room we have a representative from IBM India. Anil, yes. >> (Off microphone). >> PETER OLAF LOOMS: Also, sorry, yes, but I know of his activities primarily from IBM India where his organisation is very active in the development of Access Services for different parties, yes. And also that's the gentleman behind you. And sorry; the -yes, sorry. Because I made a mistake because I've got two persons with the same name. I apologize. >> (Off microphone). >> PETER OLAF LOOMS: So I would like to invite the Secretary-General of ITU APT to say a few words in connection with this meeting. >> Thank you, Chairman. First of all, let me spread very warm welcome to all of you. Especially that you have come here as the chair and Working Group chair. First of all, let me accept my apology because it is my -- New Delhi is also the host city. And we were not able to start off quite on the connectivity and everything as Mr. Sunil mentioned. And I take all the responsibility because being the host of the meeting in Delhi. ITU APT Foundation is a member of ITU-T and ITU-D. It's a membership organisation. We have members more than 100. And we just concluded our 8th annual general meeting on the 13th of March. So same day. And Mr. Gupta is a very reverent member of APT Foundation of India and when this was discussed by CSI especially through Mr. Gupta, we have really welcomed this job and we said okay we'll do it in Delhi. And they were scouting that and we said yes because we also have a disability friendly infrastructure so we decided and we have wanted this venue. I hope that three days will be a fruitful discussion and those who are watching online they will also benefit with this and those that are present in the room and I extend a warm reception for all of you here. And you enjoy your stay. And last night the weather has really greatened with the thunderstorm and I hope the weather is good. Anybody caught in that? So I think the weather is very pleasant and very unlikely in the month of March. So I hope you enjoy. Thank you very much. >> PETER OLAF LOOMS: Thank you very much. It remind me of whether a glass is half empty or half full. You were talking about the thunderstorm. The advantage of the thunderstorm is the following morning you have clear skies, fresh streets and I found it a wonderful welcome even if I was delayed until nearly 2:00 o'clock in order to be able to land this morning. But there are some consolations immediately after a storm. And I apologize for making a mixup between the two gentleman with the same first name. So thank you for your words of welcome. And thank you for that. And I suggest we move on to the second point of the agenda, which is the approval of the draft points for the agenda. And the documentation allocation. For those of you taking part online, we will make sure it's visible. We are displaying the draft meeting agenda and the first two points so that you can see that. So for those of you in the room and those of you following virtually, you will need Document -- the Input Document 9.1 which is the agenda document being displayed if you don't have it you can also download it from the FG AVA site and those that have registered can access the document from there. So if we just ask those in the room or those taking part online, is there anything that we have omitted from the agenda? Anything that needs to be added? Nobody in the room. But anybody online? No? So we can take the Draft Agenda as -(Audio cutting in and out). >> PETER OLAF LOOMS: As the agenda for our meeting. Now we'll go to No. 3, which is approval of the meeting report for our Third Meeting. The meeting that was held in Barcelona on the 19th of January 2012. The meeting report is in output document 0003. If you need that, it can also be accessed through the FTP site. Are there any comments or observations in connection with that particular document? No. So we can take it as written approved. So I think we can then move on to Point 4 of the agenda. This is where we start the work itself. A review of progress on logistics and infrastructure for the work of Focus Group and the Working Groups since the third meeting in January in Barcelona. As I mentioned in my opening remarks, we want to practice what we preach. And, therefore, I would like to ask Mia Ahlgren in Sweden if she would like to introduce the three documents which she and Mark Magennis have been working on. We can see that Mia got up very early this morning. The time in Sweden is now 11 minutes past 6. And it would be good if we can arrange it so we can hear from Mia. Mia, can you hear us? No, perhaps -- well I can just briefly outline what Mia Ahlgren and Mark Magennis have been working on. We have a number of challenges to hold accessible virtual meetings of this kind. Mark and Mia have been working on making sure that our procedures are as accessible as possible. And that we use the tools at our disposal, fully and effectively and in a way which allows Persons with Disabilities to take part, too. So Input Document 104 reports on what they have been doing to date. They are conscious that they are the people who make sure that what we are doing actually works, actually allows Persons with Disabilities to take part in meetings. And the report outlines not only what has been done but also the specific recommendations about how we should organise things and how we should inform potential participants so that they can register for meetings, access documents and actually take part in the meetings themselves. So Document 104, that's the one which just explains what has been happening to date. Document 105 is the group's to do list. These are outstanding tasks. Things that still need to be addressed and resolved. And that's currently being displayed on the virtual meeting. We're just about to put it up so you can see it. There you go. So it outlines the procedures and tools. The proposed action and what our current status is for the work on those particular actions. So it's an ongoing to do list with things that we need to resolve. And as and when things have been resolved, these are translated into guidelines and into instructions which are posted on the FG AVA Web site. So that people can take part. The third document, Input Document 111, contains a number of Frequently Asked Questions. And this is -- we'll be pulling it up so you can see. After each meeting we have processed a questionnaire which all of those taking part in the virtual meeting have contributed to. But based on the observations we've had from the questionnaire but also from e-mails and other kinds of communication with potential participants but also actual participants, Mark Magennis and Mia Ahlgren have put together a document called Accessible Event: Frequently Asked Questions. And that's an attempt to build up in an incremental fashion our knowledge on the processes to make this kind of virtual meeting accessible. Are there any comments or observations to the input documents from that Working Group, from Working Group K? I would personally like to thank that Working Group for their incessant efforts to make sure that we actually practice what we preach. It's extremely important to have practitioners working with disability issues to make sure that we actually have procedures which are truly accessible so that we can actually practice accessibility in order to come up with media, audiovisual media which are truly accessible. Okay. Yes, so we're trying to use the chat function of the virtual meeting tool to see if we can get ahold of Mia Ahlgren to see if we can get her online either via the chat or the telephone connection. I think this is one of the areas where Murphy still ->> MIA AHLGREN: Okay. Now I think you can here me. >> PETER OLAF LOOMS: So Mia is online. I'll put my microphone on mute. Thank you for joining us, Mia. >> MIA AHLGREN: Thank you, actually I thought Mark would be here, as well, just to explain a little bit about the last document that you showed. The FAQ, it is not a document that we put together. It's actually a document showing another tool called Accessible Event. So just make that sure. And I think you've said it all, I will not talk more but I think to make this work better for us, that are online participants, I think it would be great if you could enable us to see the attendee list so that we can chat among ourselves and so on. So that would be a function to make this work better online. Otherwise we will keep on working and improve I think, I hope, for the next meeting in Tokyo. Thank you. (Unable to hear audio). >> PETER OLAF LOOMS: From the virtual meeting to the public -the PA system here in the room. We tried to arrange it so that your presentation could be heard by everybody else. But what I did was to recapitulate just to say briefly what you had said. And I pointed out that you mentioned the importance of the chat function, the ability of persons taking part virtually to be able to contribute in a chat. But also to chat with all of the others taking part online. Apart from that what were the other main conclusions you pointed out, be Mia? >> MIA AHLGREN: I think we can see the work of this group as to step-by-step experiences that we had from each meeting. So I think there are a lot of things to do but we take it step by step. And in the end we hope to have a list of functional requirements that we can use for other meetings after the Focus Group has ended. >> PETER OLAF LOOMS: It's still very much a case of work in progress. I think we need to turn that down a little bit while I'm speaking. Mia's point is that as we identify -- as we solve certain problems, new problems emerge. And it's clear that we will need to have a very detailed checklist for the technical setup to make sure that when we come to a new venue that those who are going to be handling the technical aspects are familiar with the kinds of challenges they need to face. Thank you very much, Mia, for getting up early to take part. And I hope you can stay online for a bit longer. >> MIA AHLGREN: I'll try to keep on. >> PETER OLAF LOOMS: Okay. Sorry, Mia. So you said I will try to keep on. Yes, thank you. So I suggest we move on to Point 5 of the agenda, which is an overview of inputs to the 11 Working Groups. So we start with those which address the whole of the Focus Group. I think we should -- it's listed as Point 5.1. But I would suggest we renumber it as 5.0. Under that point we have three documents. Input Document 101. Which is about an invitation to the conference DEEP Designing Enabling Economies and Policies to be held in Toronto, Canada from the 23rd to 25th of May. And a second document which goes with the first which is the agenda and programme of the conference of the same venue. And we will just to make sure that you can see the two documents on the webinar. So this is Document 102 which shows the agenda and programme of the conference. It's called DEEP. And that means Designing Enabling Economies and Policies. Okay. Then I think we should move on -- this is information for those who are interested in that particular event. Input Document 110 is just a background resource. And we'll have some more of them in the next couple of days. It's an enhanced podcast on accessibility. It's from a module on digital design and media at the ITU University of Copenhagen. They have the same acronym ITU, which sometimes leads to confusion. But the IT University of Copenhagen also known as ITU is a relatively small university with about 1200 post graduate students. So the document just provides a link to an enhanced podcast for those who have QuickTime, they can both hear the lectures and see the slides, which are synchronized. It's in an MPEG 4 format which allows the user to move forwards and backwards through the presentation using stills rather than a video format. So a one-hour lecture with slides typically between 30 and 60 megabytes. If you use a video format, the file size will typically be between 10 and 30 times as big. And clearly in countries which don't have much bandwidth, it makes a lot of sense to provide file formats which are not so bandwidth hungry. We are planning in connection with the tutorials tomorrow and Thursday to produce of using the -(Audio cutting in and out). >> PETER OLAF LOOMS: Format to those who take part. And in principle to those wishing to see what took place in the tutorial. (Audio cutting in and out). >> PETER OLAF LOOMS: Any comments or observations. >> ALEXANDRA GASPARI: Yes, one announcement the captioning text has also a chat box that everybody can access. It's on the right side of the feature. So we have two tools for the remote participation. We have the GoToWebinar and the captioning. Thank you. >> PETER OLAF LOOMS: We have just made another small change to the audio output and we need to check that the captioning is still working. That they can still hear us. Yes. Good. We better check that it's still working. Can you still hear us? It's still working inspite of the changes. So Murphy hasn't actually undermined our activities. Point 5.1. Since the meeting in January, the third meeting in January, we have asked Mr. Gion Linder from Swiss TxT which is one of the most innovative providers of Access Services for audiovisual media in Europe. They are responsible for providing Access Services for multiple the languages for Swiss broadcasts. And Gion Linder has graciously accepted to help Clyde Smith and the Working Group on captioning, which in our parts of the world is also known as sub -- in other parts of the world is also known as subtitling. And we have been working recently on this with him. So in the eminent future we should have some updated deliverables. The advantage of having Gion Linder on board is he is at the cutting edge sometimes we could say the bleeding edge of the use of these technologies. And therefore, he is well positioned to help us identify some of the major barriers and obstacles facing those who want to provide accessible audio media through the provision of captioning or subtitles. Okay. If we move on to Point 5.2, this Working Group on audio-description known in North America as spoken captions or spoken subtitles is headed by Pilar Orero and Aline Ramael. For those who are not familiar with audio-description and spoken subtitles, we will be giving some examples of this at the tutorials on Wednesday. Wednesday afternoon. And we will be providing some materials about it. I think that in particular that for countries already producing subtitles for foreign language or different language content and in a multi-lingual country like India which has Hindi and English as a supporting language and then at least 22 other languages mentioned in the Constitution let alone about 300 other languages also spoken in different parts of India. The challenge of helping the population to overcome illiteracy, we heard from -(Audio cutting in and out). >> PETER OLAF LOOMS: About a third have limited literacy and we know from research done in India that subtitles can be an important mechanism for promoting literacy skills. What we know now is that if one has produced subtitles for a programme, in particular if the subtitles are for a programme which is in a foreign language, we can do something in parallel to literacy. We can do something about people's ability in what for them is not their mother tongue. And this is the justification for things like spoken subtitles. Because they allow the provision of support to people who are not good readers but who need to follow a programme in a language which isn't their own. A quick example: I'm from Denmark. And about 40% of television programmes on the main public service broadcast are in languages other than Danish. Typically English. But it could well be other Scandinavian languages. What we have been doing for the last 50 years is subtitling these programmes. But about 10% of the population don't read well enough to be able to benefit from those subtitles so in a month's time we start a provision, a service, an automatic service which generates an additional sound track where the translated subtitles, the subtitles translating from a foreign language into Danish are simply read aloud and you'll be able to see an example. This is a fully automated service. And many of the listeners find it difficult to believe that this is speech synthesis. The interesting thing is that speech synthesis has been developing in leaps and bounds over the last five to ten years. Perhaps with the additional impetus of car navigation systems, this is driven -- this has driven costs down and made things possible in a way they weren't even five years ago. Both in terms of cost but also in terms of quality. So Working Group B deals both with audio-description which is primarily services for those who are blind or with serious visual impairments. And that's normally for same-language programme. And the companion service for that if it's a foreign language programming which needs to be translated into the mother tongue is spoken subtitles or spoken captioning. We have two working documents from Working Group B. One is a summary of three input documents we discussed back in January. This is Document 108. And Alexandra will put it up so that those following online can see it. So 108 should now be visible. It's now moved. Yeah, so 108 is now visible. So it's a document which starts to pull together some of the material which was identified in three different input documents in January. The aim being to highlight a number of audio-description and spoken captioning that is to say services that target primarily persons who are blind, persons of cognitive impairments. This is often overlooked. In a world where people are living longer, we are seeing an increased prevalence of things like cognitive impairments due to say brain hemorrhages, different kinds of events which have an impact on people's ability to concentrate, to follow the spoken language, to follow -- to read. It's often been an overlooked field. And some of those suffering from cognitive impairments may have these difficulties due to undernourishment or malnourishment. And in the Developing World, I think we need to look more carefully at avoiding undernourishment and malnourishment. But certainly for a number of decades until we have resolved these issues, once they have had an impact on people's ability to take part in education, to take part in the working society, we would have to look at the -- what it takes to make media more accessible to those who do in fact have cognitive impairments. Do we have any comments either in the room or from those online to Point 5.2? No, I don't think so. So I think we should thank Pilar Orero and Aline Ramael for putting together Document 108. Document 109 follows on from a remark I made in my opening presentation. I mentioned the fact that for children and teenagers, but also for other groups, video games whether these be computer games or console games or these handheld device games on phones or more conventional games on the Internet are increasingly part and parcel of people in urban areas in emerging economies. We heard some figures about the number of people who had ever been online in India. I think you mentioned 100 million. If you take your big neighbor to the east, China, China has more than 100 million people who are online gamers. Not only are they online but they are also actively -- playing online games. this is primarily in urban areas but not exclusively so. So in some cases you will find village which have the equivalent of an Internet cafe. And these are often socially mobile young adults. This represents an alternative for them in their leisure time. And if one is unable to take part in what one's peers are doing, this represents a serious issue, a social issue. And that's why we asked the Focus Group for contributions on the subject of video games and accessibility. And this is the reason for including Document 109. I would just like to point out that this is the -- a chapter from a forthcoming book. It has been provided on the explicit assumption that it does not go into general circulation. So it is for internal use until such time as the book is published. So those who have subscribed and are taking part in the Focus Group request -- yeah. So the time is now quarter to 7 in Europe. It's now quarter past 11 here in Delhi. I suggest we take a break for 15 minutes. And before we go to the break, I would just like to thank our colleagues who are doing a fantastic job yet again with the captioning. We are very pleased to have you on board. And this is a preliminary thank you for your work so far. It's a big help both to those in the room but also for those taking part online. So I suggest that we reconvene at 11:35 Delhi time which would be at about 5 past 7 Central European Time. See you in 15 minutes. (Break.) >> PETER OLAF LOOMS: We will be resuming in a couple of minutes. We're just checking the infrastructure here in Delhi. I plan to start in 2 minutes from now. Welcome back, everybody. I hope you can hear us. I hope very much that our captioners in the States can hear us, too. Yes. Good. We will resume with the agenda Point 5.3 from Working Group C on visual signing and sign language. And I would like to invite my colleague Dr. Ito to present the three or -- five documents we have received since our meeting in January. Dr. Ito, you have the floor. >> TAKAYUKI ITO: Thank you very much. My name is Ito from NHK in Japan Broadcasting Corporation. My Working Group C is discussing on visual signing and sign language for deaf people. We have five contributions for this meeting. I would like to show my thanks to you all who kindly contributed to the documents and evaluated this. These five documents can be categorized into three groups. The first one is research on automatic sign language interpretation. I got two contributions. One is No. 96 from Chulalongkorn University in Thailand. Describing development of Thai Sign Language and adjoining dictionary. They are conducting research on Thai Sign Language funded by the Thai Government. As you know, sign language is very localized in each country. And therefore it's very difficult to use the results from one sign language into other countries. In such a sense, it is very important for each country to start research on it's own sign language. The other one is No. 97 from Kogakuin University in Tokyo, Japan. And this document describes text-based notation system of Japanese Sign Language. I think this kind of notation system is very important to treat sign language interpreters such as e-dictionary or automatic translation. The feature of this notation system developed by Kogakuin University is it can treat no manual motion. It means facial expression or head motion. As well as manual motion. The second group is remote production of sign language interpretation programmes. I think this kind of a system is very important because it enables broadcasters to produce sign language interpretation programmes. I received the information from the Swedish Disability Federation and found it a good document describing remote production of sign language interpretation. We also have a document from Mr. Gatarski kindly contributed his Document No. 98 to this meeting. This document describes how to add sign language interpretation to your livecast produced remote location with less expensive facilities. I think this is a good example of production and can reduce the barrier of producing sign language interpretation programmes. The third one is guidelines to produce a sign interpretation programme. We better have operational guidelines or general instructions for production or for better understandable programmes such as size and positions of a signor in the display screen. And time synchronization, et cetera, et cetera. This kind of guideline can also contribute to reduce the barrier of producing sign language programmes. Moreover, it can contribute to provide better quality sign language interpretation programmes. Here two guidelines are contributed. One is from I TC, No. 99. And the other one is from Mark Magennis, national Council for the Blind in Ireland, No. 100. ITC guideline was originated in March 2002 and was the only guideline on sign language interpretation programme I think. Recently Mr. Mark Magennis and NCBI accomplished other guidelines which are rather extensive ones including not only sign language but also other accessibility media such as subtitles and audio-description. I would like to recommend members of our Working Groups look into that document. So I revise my report No. 95 based on these contributions. Which was the revised part was shown by yellow marker. So if you have any comment on my report, it is welcome. Through the last meeting and this meeting, I got contributions in three major fields mentioned above. These cover main issues which are felt important to provide a better environment to increasing -to increase accessibility for deaf people in terms of sign language. From now on I would like to focus on the business and regulatory related issues. After research it is also important to clarify which is language dependent and which is language independent. That's all. >> PETER OLAF LOOMS: Thank you, Mr. Ito, your presentation raises a number of very interesting points which may have broader relevance not just in your Working Group but across the various Working Groups. I think the first point I would like to highlight is this issue of remote production of Access Services. One of your documents talks about remote production of sign language. But this is also the case for other things like captioning and even for -- yes for live captioning. And it is I think important to have an understanding of the challenge of being synchronous with the programme or the content which is being offered. We have a body of research which looks into the importance of having subtitles or captions which are synchronous with what is being said or displayed on the screen. It will be useful to see if we can identify comparable research on the importance of synchronous sign language services. Traditionally we have been in the habit of looking at serial interpretations, somebody speaking and then translation taking place. We've seen that for interlingual translation at meetings when someone is speaking in one language and an interpreter sums up for the last minute or two. When we see sign language on live television programmes, there will naturally be a delay simply because the interpreter has to be able to hear and see in order to be able to provide the sign language interpretation. The question which we need perhaps to confirm is that the challenge is somewhat different. The viewers will almost certainly not be lipreading. But if they are, then it's going to be an issue of where the viewer is looking. Are they looking at the interpretation? Are they trying to decode the person on the screen? Or are they trying to do both? So if the user -- if the viewer is focusing on the sign language, then perhaps the demands for synchronous presentation are not exactly the same as they are for subtitles or captioning. We're also becoming aware of new demands when it comes to subtitling. Because in the past when we were just displaying subtitles for a preprepared -- preprepared subtitles for a programme, it was important that the subtitles started about the same time as the speaker. But when subtitles or captioning are not being used for speech synthesis, it's going to be also important for the in time and the out time to be broadly comparable with what is actually being said. So that was the first point I wanted to raise. The second one was that of guidelines. You present two guidelines, one from the ITC, which was one of the regulators before they were merged in the UK into Ofcom. And then you mentioned some updates, digital television access service guidelines produced in Ireland. I think in our earlier meetings we have identified the need to talk about examples of good practice. We're not saying best practice. Because I think there is agreement that there are probably different ways in which we can achieve excellence. But if we say examples of good practice, then I think not just for sign language but also for the other groups we need to revisit this issue of identifying examples of good practice and that also applies to guidelines themselves. So are there other questions in the room or online? If not, I think it remains for me to thank Dr. Ito and the contributors to Working Group C for their very interesting contributions, which have a broader perspective than may have been foreseen. Again, the areas of remote participation and the question of cost. And the question of guidelines for the preparation of Access Services. Okay. Then we can move on to Point 5.4 which is Working Group D on emerging access services. Just before giving the floor to Take, I would just like to explain what we mean by this for those who are attending this meeting for the first time. Accessibility through the provision of services like subtitles has been with us for probably 50 or 60 years. Some of the services are not so old, things like audio-description. And we have perhaps focused very much on what we can already do. The remit for this particular Working Group on emerging access services is to look at the areas where technologies and business processes are allowing us to do things in new ways. So if we look at the screen in this particular room on the left where you can see the live captioning, the captioning is being done by our colleagues in the States using stenography. But there are other ways which can complement that particular production method. So in addition to stenography, we also have respeaking, which is using speech recognition to generate live captioning. And we also will be seeing today and tomorrow examples of speech synthesis or text-to-speech to make certain things accessible in new ways by generating audio from text with the aid of synthetic voices. And indeed we have a number of examples from Dr. I to's group of the use of avatars that is to say not having physical humans doing the visual signing but actually putting text into a simulator which can simulate both hand movements but also facial expressions. And here it's not just a technical dimension but also one of acceptability and familiarity. Just because we can do something, doesn't necessarily mean that we should necessarily do so. And this is, therefore, important for us to look more carefully at what is the evidence in favor or -- the less favorable feedback on the use of human interpreters compared with avatars to deliver different kinds of service. We've talked at earlier meetings of situations where there may be no alternative to an avatar. For example in the case of people accessing mobile phones, if a person who doesn't understand sign language has to be able to communicate with the person who is only -- whose only language is sign language, there is the option already implemented in several countries of sending a text message to the mobile phone operator and receiving an MMS animation of that particular utterance so the use scenario would be a person who isn't able to communicate in sign language can do so by sending text and receiving an animation generated in almost real-time at the head end of a mobile phone service. Clearly this has implications for things like relay service the thing is when we do things like the news should we be using a human interpreter or should we be choosing an avatar? What are the pros and cons? And what is the general acceptability of these particular approaches? So the Working Group on emerging access services is looking at access services. But also on other technologies which can actually make access services more efficient or more effective. So this may be to do with the production of access services or it may be to do with the delivery of these services. And Dr. Ito's team in Japan has worked with clever manipulation of the audio of a television for example to stretch the duration of speech so that captioning or subtitles can be displayed for a longer period of time so that people can manage to read them. And what happens when people are not speaking is that the cleverness, the intelligence allows the programme to catch up, to actually speed up slightly and imperceptibly what's going on in between areas where there is dialogue. So there are other areas that we will be looking at to do with the use of signal processing of keeping audio and video uncluttered free from additions like music and sound effects or whatever. All the way through the production and transmission chains so that Persons with Disabilities can still use and enjoy the programmes on their own terms. And we'll be providing examples of that in the next couple of days. But I don't want to preempt the work of workgroup D so therefore, I would like to hand over to Take to talk about your work to date and in particular this Input Document 103. Take, the floor is yours. >> TAKEBUMI ITAGAKI: Thank you, Chairman this is Takebumi Itagaki with Brulen university with Input Document 103 this is submitted originally from the co-coordinator Mr. David Wood of EBU. He is also a member of the ITU-R Working Group party 6B which is also interested in those hybrid transmission of the material. That is the nature of this particular document. Of course in these days, most of the Dbb systems and other systems can hope with access service within the existing infrastructure. However, some of the accessing service could be better delivered with hybrid systems. That is one of the important elements he tried to raise. And also argue about compatibility with other systems. And also interoperability between those Connected TV or Smart TV. So basically we may have to define the word what is hybrid broadcasting, i.e., the TV with Internet connection displaying both broadcasting material and communication material and presented onto a TV screen. So the main problem is there are few synchronization mechanisms. Hence the main TV can be synchronized with additional materials such as the IP channel. One of those services could be used for the user mix sign language interpreter video or sign interpreter avatar generated at the set-top box level. However, those undefined synchronization mechanisms standards or interoperability issues has to be somehow documented and standardized. That is the -- Mr. Wood tried with this incompetent put document in relation to access services. Now this document has been submitted a few days ago. Now, this is targeted towards the next meeting of Working Party 6B. That will be the end of April. So now we are going to make a statement with this entity resource ITU-R Working Party. At the moment there is no other document being submitted by the members. So this will be at the moment the only one document from the Working Group D. That's all. Thank you. >> PETER OLAF LOOMS: Thank you very much, Take. I think we should just add a little bit of additional information for those in the room and those listening taking part online. When we talk about hybrid solutions, many of you may be a bit apprehensive about the technology. But essentially what we're looking at is combining a broadcast signal and making use of the Internet to provide a television-like service. So for the last four or five years we've seen TVs which already allow you to display the Internet on a TV screen. But this is not what we're talking about. It's not as an alternative to a computer. It's an attempt to look at the way in which you can provide some degree of unification. So that people who are familiar with a television set and a remote control device for example can actually access in a very easy-to-use way the information. And we're not talking about connecting a keyboard to the television sets. We're not thinking about using it as if it were a computer. We are looking at it very much as a kind of television-like experience. And to give you some other examples of things which are already being discussed for this integrated hybrid approach, take something like subtitles or captioning. At the moment, we can provide captioners so something like teletext or proper bitmap graphics at the bottom. They look nice. The difficulty is that there are many groups of users for whom the contrast is too low or too high or the colours are wrong. Or the size isn't big enough. Or the positioning isn't in accordance with their needs. So when it's in the -- in a broadcast delivery mode like Dbb we can do quite a lot but we don't have a great deal of flexibility to allow people to personalize their use of subtitles. A second example is to do with live subtitling or live captioning. In the same way as we were talking about delays before, when we produce subtitles, for example, for this meeting, you will see the subtitles appearing with a delay of three, four or five seconds. For many people, this isn't a problem. But there are at least two major studies indicating that those who are dependent on subtitles, captioning, for an understanding of a television programme, at least a quarter of the intended users don't benefit if that delay is more than three or four or five seconds, which is often the case. So we really need to find a solution to essentially resynchronizing things. And we can do that either by delaying the video and the audio at the broadcaster and there are some legal restrictions to being able to delay things. Or we can actually look at how we can actually delay the signal and the uses the receiver of the use's device to resynchronize things and again we have preliminary studies which indicate that the accuracy of resynchronization doesn't have to be spot on. If we can make sure that the captioning or subtitles are within a second, maybe even slightly ahead, this dramatically reduces the cognitive load. It dramatically improves the proportion of people who benefit from captioning. And one of the potential applications of a hybrid solution like HbbTV or another hybrid solution is that you can do things like that. It should be possible to provide personalisation of subtitles to allow for the resynchronization of subtitles for live subtitles. And that would then require some sort of flagging so that we know what was live and what was preprepared. So it's not just the area that Take mentioned. There are a number of different areas where being able to deliver television or television-like services are on a hybrid solution could be a benefit. For example, we already see in public service broadcasters examples of ondemand services or catchup TV services available through a hybrid solution. So an example, you get home ten minutes late because of traffic. You miss the main evening news. And you want to be able to see the stories you've missed. Well, you can record them or you may have a way of seeing it on the Internet. But if you're watching on the TV, why not be able to do catchup TV on the same device. This is the sort of thing that is simple. It's easy to understand. It's fairly easy to implement and it's certainly something which seems to be appreciated in the studies we have seen. So when we're talking about emerging access services, we are looking all the way through from the conception, the production, to the delivery and use of access services. And in the case in question, the availability of a hybrid platform to combine broadcast and broadband may dramatically improve flexibility and actually open up new areas in which we can make things like television truly accessible in way which hasn't been feasible in the past. Any comments or observations otherwise from the floor or from online? No? Well thank you very much, Take, for that. We should just make a note for the record we have a draft liaison document. I think it will be advisable to give Working Group D a mandate to provide a final draft. And make sure that is discussed broadly within the Focus Group. And once it's been through the necessary editorial processes, then we should have the basis of a document which can be submitted to ITU-R 6B if I remember correctly. So what I suggest is that when we have a draft, we will send it to the reflector. And those who have registered will then be able to have a look at it. And come -- make comments, recommendations, for changes or additions. Is that acceptable Take do you remember the deadline? It was towards the end of April so that needs to be submitted before Easter for us in Europe. That means before about the 10th of April or something of that kind. Thank you very much, Take. So should we take the next point, 5.5. That's me. I'm the chair of this particular group. And here I have a confession to make. I have not been able to upload the modified document. I will be doing so in the course of the next few days this is an important area. And not just for digital television but across the board when we're looking at things like IPTV. When we're looking at mobile services. In fact, anywhere where we need to communicate not just the existence of content but also the existence of the accessibility mechanisms to go with it. So it's interesting that we have examples again from the UK where we have high levels of subtitling or captioning. We have a number of broadcasters providing as much as 20% of that output as audio-description. On occasion the viewers may not be aware of the existence of audio-description simply because the information isn't mentioned in the Electronic Programming Guide. So essentially even if it's there, if it's not mentioned, essentially it doesn't exist. And this is the reason for wanting to flag the issues. The other reason that we will need to talk about on-air promotion is to do with the challenges when you go from one system to another. For example, in the U.S. and Japan, you've moved from analogue to digital. And you've switched off analogue transmission. For television. In Europe we've nearly finished the same process. But during the transition period, we have to communicate information to users about which channels can be found where. Because unlike their old analogue receivers, the devices have to be retuned to be able to show where those channels are now to be found. So they may at one point be in one particular multiplex and suddenly disappear and reappear in another. And that particular Working Group is also identifying how do you communicate with audiences in particular more senior audiences who are not particularly technology savvy that they need to do something or their children or grandchildren may need to help them to retune their televisions or to install a new release of a browser or to install a new plug-in or to put a new app on their mobile phone. So we have a number of challenges which are not to do with the content per se. Nor the information about the content. But in fact about the delivery of content. And the things that need to be done in order to continue to enjoy something you've had in the past. So these are some of the things which are now emerging from studies which have been conducted in the UK in the digital Switchover Help Scheme which to date has been one of the best funded schemes of it's kind. It had originally a budget of 400 million pound to handle the switchover process. It seems likely that there will have been an underspend of nearly 40%. And this money is now actually being redeployed in connection with helping offliners, persons in the UK who are not digitally literate in the sense of being able to use the Internet to go online because television may be the most appropriate outreach medium to actually help them understand to help them to get started along with campaigns like Give Me an Hour which is a volunteer programme which started last year. So we have some inputs from several different projects. And in the course of the next few days we will be updating the existing deliverable to include these most recent findings. The time is now 5 minutes to -- no it's 12:25. We need to get a message to Pradipta to get him to present. So maybe I or somebody should send him an SMS or give him a call. The time is now 6:54 in the UK. I do have a telephone number let me just find it. Country code plus 44-7769-349437. So we'll give Pradipta a ring. He's in India working at the University of Cambridge. And he will be awake. But make sure that he's ready to be able to make his contribution online. So we are in the process of communicating with Pradipta Biswas at the University of Cambridge who is part of a fairly big Accessibility Team. His research concerns the production of simulators, which allow for the testing of web or mobile or TV services without having to ask human subjects to do so. And the aim is not to avoid testing things on the person for whom they intended but to eliminate fairly basic design flaws at an early stage with the aid of simulators. So that when we get onto testing things with the audiences for whom they are intended, we can actually make full and effective use of their time. And tomorrow afternoon we will be showing examples of that -- the work they are doing at Cambridge with some India examples, how you can use the simulator to analyze the design of a booking system for the Indian railways. And that's one of the examples we will be demonstrating tomorrow. How we can simulate persons with different kinds of disabilities and explain to the designers where their particular design interface design may have weaknesses for different kinds of users. So Pradipta is in the process of getting on. It's going to take a few more moments. The business of simulation is an important one. The research I was responsible for doing three years ago on subtitling for the news in Denmark involved testing on 30 people who were between the ages of 40 and 96 with a wide range of disabilities. It was interesting that the lady who was 96 turned up at the broadcasting centre on her bicycle. So she was still very, very fit. But even so, admitted that there were certain difficulties watching the news. In the panel there was also a lady who had been head of a nursing department working in different parts of Scandinavia but had suffered a brain hemorrhage. And until she had had that, she was running five or six marathons a year in her 60s. But some of these events come unexpectedly even to people who are in good physical -- in a good physical state. And she had to relearn to speak and it had taken her about 18 months after having had this brain hemorrhage. And she was reporting on the difficulties of reading subtitles. So the idea of using a simulation is not to avoid things of this kind. But the practicalities, the logistics of inviting people with different kinds of disabilities, getting them to a test site and involving them means that we have to respect their time, respect their particular situation. And if there are things we can do with the aid of a simulator, I think there's a good case for doing so. So that when we actually use live participants, we can really benefit from their contributions and their insights in a way -- to help us with things which we may not otherwise have been able to foresee. I think it is now 12:30. Shall we go to the next point and then come back to Pradipta when we have him online? So that would mean to go onto Working Group G. That's the same situation, Peter Molsted from DR is in the process of terminating some studies on synchronization issues. And he has been looking at the synchronization or the resynchronization of subtitles in connection with both preprepared broadcasts, preprepared subtitling and live subtitling. They have been conducting a very detailed series of detective activities to actually identify at what point delays occur. When you produce subtitles if we take the situation in the UK, they still haven't switched over analogue broadcasting. And the people producing subtitles have access to the programme in it's analogue form the interesting thing about that is that arrives at the TV-set two seconds before the digital version. And if it's in high definition it arrives at the television set four to five seconds before the high definition digital version. Simply because encoding the signal takes time. And that means if you're doing subtitling for a high definition channel and you have access to the original signal, this may be analogue or it may be the uncompressed can digital signal. Then you can, in fact, reduce the delay by in some cases two and five seconds. I think we should go back to Working Group F and now we have Pradipta on the line from Cambridge. >> PRADIPTA BISWAS: Hello, can you hear me. >> PETER OLAF LOOMS: Good morning, Pradipta, we can hear you clearly. >> PRADIPTA BISWAS: Thank you. Good afternoon, everyone. So this Working Group is about participation and media and the Input Document what we try to do is first define or scoping the terms what does it mean by participation and digital media. And then we try to identify the existing problem. And we try to figure out the -things for -- recommendations for the future and we also try to find out some recommendations by which we can make more inclusive future in case of participation. So if I now go through the document quickly. So consists of both devices and services. Device means the different types of connective device like television, computers, gaming consoles, SmartPhones, et cetera and services, television broadcast or software and so on. And that participation will be on the virtual media Internet connection because it consists of both audio media and also viewing the media and transmission of media. And here you can see that on the second page we try to find out all of the existing stakeholders. Here we also took an end-to-end production from production to -- considering users with different ideas which considers both situations and -(Audio cutting in and out). >> PRADIPTA BISWAS: Second is a participation which starts at a very simple use case ->> PETER OLAF LOOMS: Pradipta, just a practical word. We are experiencing some audio cutout. So may I suggest you speak slightly more slowly so that the captioners in the States can follow what we are saying. So give yourself slightly more time and if we have any technical problems we'll just tell you and ask you to repeat what you have said. (Audio cutting in and out). >> PETER OLAF LOOMS: Right now we can't hear you, Pradipta. We can hear there is some activity. But the audio is cutting in and out. We still can't hear you, Pradipta. While we are waiting perhaps I can contribute with something having been discussing the matter with Pradipta. One of the overriding concerns as we heard from Pradipta is what do we mean by participation? In the battle days of television, we used to talk about couch potatoes. To say people who were sitting back in a fairly passive physical mode watching things. That may have applied predominantly to things like television. But clearly does not apply when it comes to social media and things like games where people are interacting there, they are contributing, they are doing things. Pradipta, we still have difficulties. So it seems to do with your side, Pradipta, it has to do with the buffering. We still can't hear you. I will continue to explain some of the things that have been happening. So if we take television as I said the predominant mode in the past was the couch potato mode. Sitting back watching things on your own or with somebody else and there would have been some sort of social interaction with the other people in the room. Well, there are already significant changes in which we would view things even like television. So in some industrialized countries as much as 10% of television viewing is now asynchronous. As we say people are watching it when it suits them and not when it's broadcast so in the UK as much as early as six months ago they passed the 10% mark in which television itself was being viewed minutes or hours or even days after the programme was shown. So participation clearly has to take into consideration whether we're talking about real-time events or asynchronous events. And whether we are looking at communication in for example a social viewing in the same room or people actually discussing with each other while watching a television programme. And that's just television. So if we take something like social media, something like Facebook, which is now quite widely used in India and many other places, too. But it's -- Facebook is a good example of a social medium which only exists because people are not just consuming, they are not just looking in, they are actually contributing to begin with. They are choosing things. They are assessing things. They are talking about like or dislike. A thumbs up, thumbs down. They may be commenting on things. They may be adding messages. They may be posting pictures. They may be posting other kinds of things. They may be setting up groups on specific interests. This particular Focus Group has a Facebook group. But for certain commercial entities like Coca-Cola they have nearly 60 million Facebook user on their page. So we have a continuum from being relatively passive from sharing media to commenting on media to mashing up to creating new things based on what is already out there and ultimately actually creating things from scratch. And therefore, in order to be able to work with digital audiovisual media we need to have a clear and coherent taxonomy that's the way in which we can describe different kinds of participation which apply not just to old media like television but also to social media, to games and to many of the other things which people are using as part and parcel of their everyday lives and that's why Pradipta and his group are beginning the work to actually provide us with the intellectual framework to talk about participation which will then feed back into the work of the other Working Groups. So I'm afraid we'll have to send our apologies to Pradipta. We think that the difficulty lies at your end rather than this end. And it seems to be something to do with the buffering, something which is involving a limiter or compressor which is simply cutting off the audio so that we can't hear you in a clear and consistent fashion. So what I can do is recommend that those who are interested in seeing what is going on is to read Input Document 106, which is the updated version from Working Group F. And this mentioned work in progress, the new areas they are trying to map out in particular how we can actually incorporate into a taxonomy something on participation which can make sense for both old media and new. Some television, social media games. Other kinds of media. The second document, that's Document 100, sorry; I was misreading it. It's Document 107. This is from the ongoing work of both of that centre in Cambridge and a multi country research project on virtual user modeling and simulation standardisation. Service provider how do we put together simulators? How do we actually standardize simulators so they actually provide us with appropriate answers to our design questions. So for the first step is to actually have simulators that reflect different groups or clusters of disabilities. And secondly to have some sort of consensus as to the kinds of things that the simulators have to be able to do before we start applying them. And just to reiterate the point I made earlier, this is not to avoid doing things for Persons with Disabilities with their direct participation. But to make better use of their time by avoiding all of the basic design flaws with the aid of a simulator so that we can make full and effective use of tests with live participants. Okay. That was Point 5.6. And I think we should then perhaps move on. We should perhaps it's time to finish off what I had started on on Group G. The challenges for Group G are to do with some of the issues we've heard about from the access service groups. Essentially what constitutes quality. And what interpretations different stakeholders have when they talk about the quality of subtitles. The quality of audio-description or the quality of visual signing. It would seem to be a bit surprising that we can already establish from the preliminary work in the various groups that there is not necessarily a clear consensus about what constitutes quality even in something as familiar as subtitling or captioning. The second point to make is that many of these services require a degree of synchronisticity in order to work appropriately they need to be in sync with the content or programme to which they relate. And the realities of moving from the analogue to digital world are that any kind of digital encoding involves delays. And therefore we need to understand the whole of the supply chain from -- from the source to the user's device and ultimately to the viewer or the user himself or herself in order to establish not just Quality of Service but also quality of experience. The updated document which I have seen in Danish looks at some recent studies to be able to fiat what point in this end-to-end supply chain delays occur and what can be done about it or in some cases how can we make full use of the fact that the delay in encoding high definition video is four to five seconds whereas the delay in the encoding of live bitmap subtitles or captioning is only a few milliseconds. If we know what these delays are, in some cases we can actually take corrective action and actually reduce the delays which would otherwise be present. But that assumes that we are able to identify where these delays are. And we also need to be able to flag. So it's different kinds of subtitles or captions. So that preprepared subtitles are provided with a delay where -- whereas live subtitles aren't in order to benefit from the encoding delays for the video and the audio. So this is still work in progress. But it continues from what we have heard at meetings 2 and 3 to look at some of the issues facing one particular platform that of dish to a television. And the general lessons learned should be broadly applicable in the first instance almost directly to IPTV that is afterall essentially just a different way of distributing audiovisual content. But it will also contain lessons learned that in some cases will be applicable to things like social media, to services on mobile phones and possibly to things like video games. Yeah. I think it might -- we could change the agenda and get a presentation if we move -- say we finish Point 5.7. If we jump to Point 5.9, that's the Working Group I on mobile and handheld devices which is -- you can take mine. We have John Lee from Canada as the co-coordinator and I'm very interested to hear on the state of play from your particular Working Group. So the floor is yours, John. >> JOHN LEE: Thank you, Peter. So the -- for this meeting we had planned on having two different contributions. Unfortunately they were not in on time to be submitted but I can give verbal demonstrations as to what they are. The first is a demonstration related to automatic captioning that occurs on an offsite formed of devices. The way the system works is you install a programme on your mobile device and speak into it and while the person speaking it the person is from an offsite server which uploads the automatic transcription onto a Web site that people can access and read. Very much like the captioning that we have currently available in the room right now. The accuracy is a little bit of an issue of course because it's automatic. But it's a very effective system that can be used in situations where a user with a disability issues can use in regular social situations. And it is a portable system that does not require prior setup. So it's a very interesting system that's currently being developed. It's not quite a product yet that's available widely or broadly but it is a demo and prototype for people to try out. Secondly the contribution that unfortunately I was not able to bring here was related to the use of mobile devices and accessory to main media as a way of providing accessibility information. So in a basic environment such as watching TV where there's a lot of discussion of where do you place the captioning? At what rate? How are things presented on the screen? What we're trying to look at is how do you use a mobile device to offset the needs on the display and take the accessible information and put it on a secondary screen which would be an accessory device so the captioning you would be carrying on your mobile device while you're watching the TV and it creates a personal experience that is very useful for individuals. And not only that there are other aspects to this where if you have a hearing aid and your hearing aid is already associated to your normal device you don't need to go and reassociate to other radio frequencies that may be available for hearing aid information to come through. You can just have an IP connection from your mobile phone to the server through an IP connection. And all the data is pushed down so you can use it without requiring a new connection. Because that connection is permanent within a personal environment by making use of the mobile phone. So these are a few of the use cases that we're exploring and trying to provide uses of everyday devices to enhance accessibility for everyday users can. >> PETER OLAF LOOMS: Thank you, John. I have a couple of questions myself. The first one, in terms of accuracy, that was back to the whole business of what constitutes quality, it would be interesting to compare what happens in that setup with the quality of this uploaded to Google because they now have an automatic captioning service. Which is quite interesting. Because it's -- they are also on YouTube examples of where it really goes quite badly wrong. But that would be an interesting yardstick to compare the two. >> JOHN LEE: I agree. And I think part of what we need to do is we need to understand what are the quality measures that we need to look at. I believe that that's an area that we really should be spending some time to -- so when these new system come up we can effectively compare them to existing yardsticks as you mentioned. >> PETER OLAF LOOMS: Yeah so that's something we have already identified in at least two groups. And that's beginning to become one of the major themes. How can we contribute while clarifying how people see quality and what measures or yardsticks or metrics do they have for talking about the quality of captioning. I assume that the service you talked about was in English to begin with. >> JOHN LEE: Currently I'm not very familiar with other languages but I believe it's currently the available in English and it is built for download -- I don't have the site offhand but I can provide it at a later time and it's currently available for the Android platform and iOS platforms and it's a downloadable app that you can load and run and it runs very much like a dictation application. So you just speak into it and then it's connected to an IP connection to a server that does the processing and then displays it online. It's a very effective way of getting automatic captioning. >> PETER OLAF LOOMS: The second point was to do with the participation back to Pradipta's presentation where we're looking at participation or interaction and there it becomes quite critical. We had one scenario where we're talking about handheld devices in their own right. So people accessing content on a SmartPhone or on a computer tablet as the primary platform. So I was in the airport yesterday watching people who were looking at films on their tablets. I didn't see anybody using their SmartPhones because nobody had one of the really big Samsungs but suddenly on tablets just looking around I saw three people watching films on tablets and I saw one person watching films on computers. But the other scenario that you're talking about is one that's becoming more and more widespread. SmartPhones and tablets are probably along with small laptops the first true examples of personal computing. Desktop computers were never really personal because they were there and we couldn't walk around with them. But if something is in your pocket or in your bag, it suddenly becomes personal in the true sense. And that means even in social contexts like say watching TV with your family, you may in fact be using a second screen approach or if you go to the cinema or theater we held a meeting in Barcelona and we saw examples of people in Barcelona being taught if they had an Android or an Apple SmartPhone to turn on their phones when they went into the theater so that they could follow captioning all audio-description on their SmartPhones because it was being offered directly to their SmartPhones in the theater itself. And they found a way of muting phone traffic, too. But that seems to be quite important. Also if we take the example of people sitting in their living rooms with family members, those who may need more personalisation can do so on their SmartPhone or increasingly on their tablet which they can have right in front of them. And this is a new set of user scenarios which we need to take into consideration. And that's -- requires us to look fairly carefully at what we mean by media and the context in which media are being used. >> JOHN LEE: Definitely, Peter. And with that respect, there's actually an area that we really should look at very carefully. And that's what are the available protocols to connect the devices from one device to the other. Right now there's a lot of different organisations within even the ITU and Study Groups that are studying this very question and the an area for disability purposes we should look at what is the thing that will enhance the experience for all possible users. And it's an area of study that we should probably open. And just to address something else you mentioned, which is the use of mobile devices as the primary device for media, I didn't really touch on that previously. But that is a very big area of worry within the mobile industries. When you talk about things like captioning, when you talk about things like automatic -- automatic -- sorry; automatic descriptions and automatic captioning because what happens is the mobile device in and of it itself is not a primary media device. It was never intended to be a primary media device when it was first started. So adding a lot of information to it, there are certain things that were never considered. So for example, when you have captioning running on a TV screen, the service area that the caption covers on a TV screen relative to the image is not very high. But on a mobile screen where you're talking about a three and a half inch screen for any type of captioning to be visible and reading it also has to take up a much bigger portion of the screen otherwise you actually lose content data so there are things we are exploring to see what is available be to do this. And I know that there's a lot of work going on in the U.S. It's probably something we should bring into the Working Group, as well to see if we can explore that very issue. And adding to that because we typically use IP connections on mobile devices, what happens right now is when the broadcast information contains that -- contains that accessibility information it is broadcast simultaneously however when that same information media content is moved into an IP repository typically those extra layers of accessibility functions are stripped out from the bare content which is the audio and visual so adding them at a later time or trying to reprocess the information becomes very difficult to do and that's something from a broadcasting perspective it's something that needs to be talked about, as well. >> PETER OLAF LOOMS: Thank you, John, we should just do a brief promo for Take's presentation because on Thursday morning if not Wednesday afternoon, he will be talking briefly about experimental services which have been following this second screen approach which has been focused HbbTV and also others done providing second screen services and some of the issues that emerged from synchronizing things so we've got a history of this that dates back about 30 years. It sounds surprising. But audio-description in Portugal was first done broadcasting the audio-description on the medium wave, on radio. The reason for doing so was that most of the people needing audio-description were elderly and in their homes they had a medium wave radio in their same living room as their television set therefore it made a lot of sense for them to turn it on and tune it to a radio for a frequency just to listen to the audio-description and they have been doing that for 30 years. So sometimes it's just a question of using our brains and thinking about what technologies are there that the second point to make in many cases perhaps we have already done it in a different way and perhaps we could learn by following an evidence-based approach from what's already being done. So what I would suggest to John is that as a huge collaborative project which is now in it's ninth year -- eighth year called Beyond 30. And it looks at the whole business of advertising commercials on a screen basis. And this is coordinated from an Australian university, Murdoch University. No relationship to Rubert it's just his grandfather who was a priest and who founded Murdoch University. But the research institute called institute -- the Interactive Television Research Institute does this kind of multiple screen research with a very specific remit to see how things like advertising work and they have done studies at least four years ago on the usability aspects of delivering services on small screens at a time when we were still using 2.5 and 3G so I think there's probably something that can be learned if not from the results, certainly from a methodological perspective we can build on and I would be happy to provide you with the context there. So the time is now 7 minutes past 1 here in Delhi. Or 8:37 in Europe, west of Europe. And that's 7:37 in Ireland and the UK. And Portugal. So I would recognize that we adjourn for -- until 2:30 India time. In an hour and a half from now. And again, many thanks to our colleagues who provided such a good service with the live captioning. I think before we adjourn for lunch we could show our appreciation in the conventional fashion. (Applause) (Thank you!). >> PETER OLAF LOOMS: Thank you very much. (Session ended at 2:38 a.m. CST March 13, 2012) *** This text is being provided in a rough draft format. Communication Access Realtime Translation (CART) is provided in order to facilitate communication accessibility and may not be a totally verbatim record of the proceedings. ***