Contents BSA Annual Conference Programme .................................................................................................. 2 Oral presentations ................................................................................................................................ 5 Symposium abstracts ......................................................................................................................... 12 From hair cells to hearing abstracts ................................................................................................... 12 PPC symposium Abstract .................................................................................................................. 14 Special Interest Group Session Abstracts .......................................................................................... 16 Sponsor workshops ............................................................................................................................ 18 Oral Presentation Abstracts: Tuesday afternoon, 1400-1545 ............................................................ 20 Oral Presentation Abstracts: Tuesday afternoon, 1615-1700 ............................................................ 26 Oral Presentation Abstracts: Wednesday afternoon 1430-1600 (parallel session 1) ......................... 28 Oral Presentation Abstracts : Wednesday afternoon 1430-1600 (parallel session 2) ........................ 33 Poster Abstracts: General sensation and perception .......................................................................... 40 Poster Abstracts: Hearing loss and hearing aids ................................................................................ 70 Poster Abstracts: Cochlear implants .................................................................................................. 96 Poster Abstracts: Balance and vEMPs ............................................................................................. 105 1 BSA Annual Conference Programme Keele University 1st – 3rd September 2014 Quick guide: Keynotes Orals Exhibition Posters Clinical workshops Special events MONDAY 1st September Time WESTMINSTER THEATRE MEETING ROOM 1.098 10.00–11.00 11.00–12.00 BSA Council meeting (90 min) 12.00 – 1.00 13.00– 14.00 14.00– 15.45 EXHIBITION SUITE Exhibitor set up POSTERS Poster set up Exhibitor set up Lunch Introduction to conference Keynote I: Prof Anne Schilder Symposium From hair cells to hearing Sponsor Workshop Oticon 15.45–16.15 Exhibition Poster viewing Refreshments 16.15–17.15 Symposium continued 17.15–18.00 Twilight lecture : Prof Trevor Cox (Author of Sonic Wonderland) Sponsor wokshop Interacoustics Social event: wine/nibbles + dinner 2 TUESDAY 2nd September TIME 08.30–9.30 09.30–10.30 10.30–11.00 11.00-12.00 WESTMINSTER THEATRE Keynote II: Dr Stefan Launer BSA AGM (45 mins) Professional Practice Committee symposium 12.00– 13.00 13.00–14.00 14.00–15.45 Keynote 3: Prof T Ricketts Oral presentations MEETING ROOM 1.098 MEETING ROOM 1.099 POSTERS Exhibition refreshments Poster viewing Sponsor workshop Tinnitus Clinic New members workshop Meeting with sponsors Lunch Sponsor workshop GN Resound Discussion Forum ARIG “Who defines Rehabilitation?” Exhibition 15.45-16.15 Poster viewing Refreshments 16.15– 17.00 Oral presentations 17.00–17.30 Action on Hearing Loss presentation BSA Awards Ceremony Poster prizes Conference dinner in Keele Hall 19.30–23.00 EXHIBITION SUITE Forum continued Exhibition 3 WEDNESDAY 3rd September Session 4 TIME 08.30– 9..30 9.30 – 10.30 WESTMINSTER THEATRE UK Biobank Update Innovation forum 10.30–11.15 11.15- 12.30 12.30– 13.30 13.30–14.30 14.30– 16.00 LECTURE ROOM ONE (0.061) APD Update BIG Update Future plans for the Annual Conference Ted Evans Lecture Prof Corne Kros Oral presentations (basic science) MEETING ROOM 1.098 Journal club Adult Hearing Screening Electrophysiology SIG EXHIBITION SUITE Exhibition Refreshments POSTERS Poster viewing Exhibition closes Lunch Oral presentations (clinical/translational) End of meeting 4 BSA Annual Conference 2014 Oral presentations Monday 1st September Westminster Theatre 13.00. Opening Keynote I: Scientific evidence within the clinical context Prof. Anne Schilder, University College London. 14. 00. Symposium: From Hair Cells to Hearing 14.00. The sensory hair cell in normal hearing and disease Prof. Dave Furness, Keele University 14.35. Coding of acoustic information in the auditory pathway Prof. Alan Palmer, MRC Institute of Hearing Research, Nottingham. 15.10. Integrating acoustic and electric information following cochlear implantation Dr. Padraig Kitterick, NIHR Nottingham Hearing Biomedical Research Unit. 15.45 Break 16.15. Listening difficulties and auditory processing disorder (APD) in children: proposed mechanisms Prof. David Moore, Cincinnati Children’s Hospital, Cincinnati, USA. 17.15. The Twilight Lecture: The Acoustics Behind Sonic Wonderland Prof. Trevor Cox Salford University, Manchester. 18.00 Social event – wine and nibbles Room 1.098 14.00. Sponsor Workshop – Oticon (Alison Stone) 14:00 – 14:30 Evidence for a new compression strategy for children 14:45 – 15:15 New clinical tools for paediatric fittings 15:25 – 15:45 Connectivity for Kids 16.15 PM: Sponsor Workshop Interacoustics 5 Tuesday 2nd September, morning Westminster Theatre 8.30. Keynote II: Hearing Systems in a connected world. Dr Stefan Launer, Phonak, Zurich, Switzerland 9.30. BSA AGM 10.30. Professional Practice Committee Symposium – How you can make audiology better 10.30. Introduction Graham Frost, PPC Vice Chair 10.35. What PPC is doing to make audiology better Dr Daniel Rowan, ISVR, Southampton University. 11.00. The impact of NICE accreditation Deborah Collis, Associate Director, Accreditation and Quality Assurance, NICE 11.30. PPC docuemts: What, How, When? It’s your turn. Dr Sebastian Hendricks, Barnet & Chase Farm Hospitals and the RNTNEH 11.55 Summary Graham Frost. 12.00 – 13.00. Lunch Room 1.098 11.00. Spomsor Workshop – Timmitus Clinic Tinnitus pathophysiology and evolving clinical options for therapy. Mark Williams: 12.00 – 13.00. New/young members workshop Tuesday 2nd September, afternooon Westminster Theatre 13.00. Keynote III: Speech recognition and spatialisation in complex listening environments: effects of hearing aids and processing Prof. Todd Rickets Vanderbilt Bill Wilkerson Center for Otolaryngology, Nashville, USA Oral Presentations 14.00 – 14.15. L.F. Halliday, O. Tuomainen, and S. Rosen. Auditory processing deficits in children with mild to moderate sensorineural hearing loss (Listening problems with hearing loss) 6 14.15 – 14.30. J.G. Barry, D. Tomlin, D.R. Moore and H. Dillon. Assessing the role of parental report (ECLiPS) in supporting assessment of children referred for auditory processing disorder 14.30 – 14.45. E. Holmes, P.T. Kitterick and A.Q. Summerfield. Do children with hearing loss show atypical attention during ‘cocktail party’ listening? 14.45 – 15.00. S.A. Flanagan and U. Goswami. Temporal asymmetry and loudness : amplitude envelope sensitivity in developmental dyslexia 15.00 – 15.15. L.Zhang, P.Jennings and F. Schlaghecken.Learning to ignore background noise in VCV test 15.15 – 15.30. A. Sarampalis, D. Alfandari, J.H. Gmelin, M.F. Lohden, A.F. Pietrus-Rajman and D. Başkent. Cognitive listening fatigue with degraded speech 15.30 - 15.45. D. Kunke, M. J. Henry, R. Hannemann, M. Serman and J. Obleser. A link between auditory attention and self perceived communication success in older listeners Break 16.15 – 16.30. Brian C.J. Moore and Sara M.K. Madsen. Preliminary results of a survey of experiences in listening to music via hearing aids 16.30 – 16.45. listeners J. L. Verhe* and J. Hots. Mid-bandwidth loudness depression in hearing-impaired 16.45 – 17.00. A. King, K. Hopkins and C. J. Plack. The effects of age and hearing loss on neural synchrony and interaural phase difference discrimination 17.00 - 17.10. Action on Hearing Loss presentation. BSA Awards Ceremony and Poster Prizes Meeting Room 1.098 14.00. Sponsor Workshop – GN Resound. Paul Leeming, Support Audiologist Meeting Room 1.099 14.00. BSA ARIG Discussion forum on patient-centred care: “Who defines rehabilitation?” (i) shared decision making (ii) facilitating change through developing self-efficacy (iii) outcome measures for clinicians and researchers Led by Amanda Casey, Dr. Helen Pryce and Dr. Mel Ferguson 19.30-23.00. Conference dinner in Keele Hall 7 Wednesday 3rd September, morning Westminster Theatre 8.30-9.30. Large scale hearing studies using the UK Biobank resource Introduction to UK Biobank session Prof. Kevin Munro, University of Manchester. Hearing in middle age: a population snapshot Prof. Kevin J Munro. Cigarette smoking, alcohol consumption and hearing Dr. Piers Dawes, University of Manchester. Speech in noise hearing, pure tone threshold and cognition Prof. David Moore, Cincinnati Children’s Hospital, Cincinnati, USA. Hearing loss and cognitive decline: the role of hearing aids, social isolation and depression Dr. Piers Dawes Concluding comments Dr. Piers Dawes 9.30 – 10.30. Innovation forum 12.30 Discussion on the future plans for the Annual Conference (open to all) Lecture theatre 01 10.30-11.15. BSA Auditory Processing Disorder SIG update: Onwards and Upwards Dr Nicci Campbell, ISVR, University of Southampton, and Prof Dave Moore, Cincinnati Children’s Hospital, Cincinnati, USA. 11.15-12.00. BSA Balance Interest Group update The Video Head Impulse Test and its relationship to caloric testing Dr Steven Bell, Hearing and Balance Centre, University of Southampton. Motivational approach to behaviour change in vestibular rehabilitation to attendance Dr Nicola Topass, Audiology Department, Royal Surrey County Hospital. improve clinic Room 1.098 10.30-12.30 BSA Journal Club: Adult Hearing Screening Led by Dr Cherilee Rutherford, Dr Lorraine Gailley, John Day 12.30 – 13.20. Electrophysiology SIG meeting 12.30 Lunch 8 Wednesday 3rd September, afternoon Westminster Theatre 13.30. Keynote IV: Ted Evans Lecture. Adventures in mammalian mechanotransduction: adaptation, aminoglycosides and anomalous currents. Prof. Corné Kros, University of Sussex. Oral Presentations Parallel session 1 14.30 – 14.45. B. Coomber, V.L. Kowalkowski, J.I. Berger, A.R. Palmer and M.N. Wallace. Changes in nitric oxide synthase in the cochlear nucleus following unilateral acoustic trauma 14.45 – 15.00. M. Sayles and M. G. Heinz. Amplitude-modulation detection and discrimination in the chinchilla ventral cochlear nucleus following noise-induced hearing loss 15.00 – 15.15. L.D. Orton and A. Rees. Commissural improvement of sound level sensitivity and discriminability in the inferior colliculi 15.15 – 15.30. S.M. Town, K. C. Wood, H. Atilgan, G. P. Jones and J. K. Bizley. Probing the physiology of perception: Invariant neural responses in ferret auditory cortex during vowel discrimination 15.30 – 15.45. R.P, Carlyon, J.M. Deeks, A.J. Billig and J.A. Bierer. Across-electrode variation in gap-detection and stimulus-detection thresholds for cochlear implant users , § MRC Cognition and Brain Sciences Unit, Cambridge, UK ,‡ University of Washington, Seattle, W.A, U.S.A. 15.45 – 16.00. N.P.Todd and C.S. Lee. Source analyses of vestibular evoked potentials (VsEPs) activated during rhythm perception Lecture theatre 1 Parallel session 2 14.30 – 14.45. E. Heffernan, N. Coulson, H. Henshaw and M. Ferguson. The psychosocial experiences of individuals with hearing loss 14.45 – 15.00. S.E.I. Aboagye*, M.A. Ferguson§, N. Coulson‡, J. Birchall§‡ and J.G. Barry*, “You’re not going to make me wear hearing aids are you?” Factors affecting adolescent’s uptake and adherence to aural habilitation 15.00 – 15.15. S. Dasgupta, S.Raghavan, M.O’Hare and L.Marl. X linked gusher syndrome – a rare cause for hearing loss in children 15.15 – 15.30. S.Wadsworth, H. Fortnum,, A. McCormack and I. Mulla.. What encourages or discourages hearing-impaired children to take part in sport? 15.30 15.45. G. Al-Malky, M. De Jongh, M. Kikic, S.J. Dawson and R. Suri. Assessment of current provision for auditory ototoxicity monitoring in the UK 15.45 – 16.00. D. Hewitt. Everything an Audiologist needs to know about speech in noise 16.00. Conference Close 9 Keynote speakers Biographies and talk information Professor Anne Schilder, ENT surgeon and NIHR Research Professor, leads the evidENT team at UCL dedicated to developing high quality clinical research and promote evidence based management of ENT, Hearing and Balance conditions,. She is professor of Paediatric ENT at UCL's Ear Institute, Royal National Throat, Nose and Ear Hospital and University Medical Centre Utrecht, as well as Visiting Professor at Oxford University. She is an expert in clinical trials; her RCTs in the field of upper airway infections in children have influenced how global health-care systems think about the management of these conditions and have been translated in evidence-based guidelines and health policies. Her evidENT team brings together clinicians and researchers across the UK. evidENt works with patients, charities, industry and nother stakeholders to ensure patients benefit from new and better treatments and to learn how to improve ENT, Hearing and Balance services for the future. Her talk is entitled: Scientific evidence within the clinical context. Trevor Cox is Professor of Acoustic Engineering at Salford University where he is a researcher and teacher in acoustics, signal processing and perception, and renowned for his media work, documentaries and communicating to the general population as well as to students. He has won prestigious awards from the Institute of Acoustics of which he is a former President. He has written many books, including Sonic Wonderland: A Scientific Odyssey of Sound and a substantial body of scientific literature and research articles. His talk is entitled: The Acoustics Behind Sonic Wonderland Sonic Wonderland is a popular science book about the most remarkable sounds in the world. This talk will look at some of the detailed acoustic science behind a few of the wonders, picking examples that required first hand research in preparing the book. It will include solving the mystery of Echo Bridge, something that first appeared in the Journal of Acoustical Society of America in the 1940s. The talk will include measurements on the badly tuned musical road in California. To finish, the talk will outline a search for places with excessive reverberation time. This will include the vast oil tank in Scotland where Trevor broke the Guinness World Record for the longest 'longest echo'. The author returned to the tank this spring to play the saxophone, and he will reveal what it is like to play in a space with a reverberation time of 75 seconds. Stefan Launer has been Vice President Science & Technology of Sonova since April 2008 and joined the Management Board in April 2013. He started his professional career at Phonak in 1995 in the Research and Development department where he held various functions. Today he is in charge of managing the basic science and technology programs in various fields of hearing health care, the development of core technologies and intellectual property rights. Stefan Launer studied Physics at the University of Würzburg and in 1995 was awarded a PhD from the University of Oldenburg on modeling auditory perception in hearing impaired subjects. His talk is entitled: Hearing Systems in a connected world. Over the last two decades hearing instruments have turned into intelligent systems offering a range of 10 different algorithms for addressing listening needs in specific acoustic environments. More recently modern hearing systems are becoming wirelessly connected to each other to form body area networks. These ear to ear connections allow applying new features such as binaural signal processing techniques to improve communication in complex listening conditions. A second class of applications allows connecting hearing systems to external audio sources such as phones, remote microphones, TV, audio players etc. Today ear to ear wireless links allow to apply binaural signal processing mimicking the way the auditory system processes sounds binaurally. These algorithms can offer significant benefit in various difficult listening conditions beyond the performance of classical noise reduction and directional microphone systems. The second class of systems is particularly designed to improve communication over larger distances eg school environments. A third set of applications of wireless connectivity is to offer new approaches to service delivery and new fitting approaches to hearing impaired people. In this talk I want to present the state of the art of hearing instrument technology and discuss future perspectives of the new technology trends. Todd A. Ricketts, Ph.D, CCC-A, FAAA, is an associate professor and the Director of Graduate Studies, Department of Hearing and Speech Sciences at the Vanderbilt Bill Wilkerson center for Otolaryngology and Communication Sciences; and, Director of the Dan Maddox Hearing Aid Research Laboratory. Todd has published more than 100 scholarly articles and book chapters. To date he has presented over 300 scholarly papers, poster presentations, short courses, mini-seminars, and workshops both nationally and internationally. He continues to pursue a federally and industry funded program studying various aspects of hearing, hearing aids and cochlear implants. He was named a Fellow of the American Speech Language and Hearing Association in 2006 and his article "Directional benefit in simulated classroom environments" received the Editors award from the American Journal of Audiology at the 2008 AHSA convention. He also is a past editor-in-chief of the quarterly journal Trends in Amplification, a current associate editor for the Journal of Speech, Language and Hearing Research and the past chair of the Vanderbilt Institutional Review Board. His talk is entitled: Speech recognition and spatialisation in complex listening environments: effects of hearing aids and processing. Corné Kros is Professor of Neuroscience at the University of Sussex. Noted for his seminal studies of cochlear hair cell physiology, in particular the properties of spontaneous activity in pre-hearing inner hair cells and the process of mechanoelectrical transduction by which these cells detect sound, Professor Kros's research is at the forefront of basic science in hearing. His current interests lie in the effects of aminoglycoside antibiotics on hearing, and the development of blocking agents that might prevent aminoglycoside-induced damage to the hair cells. His talk is entitled: Adventures in mammalian mechanotransduction: adaptation, aminoglycosides and anomalous currents C.J. Kros*§, *Sussex Neuroscience, School of Life Sciences, University of Sussex, Brighton, UK, § ENT Department, University Medical Center Groningen, Groningen, The Netherlands Hearing requires sound being transduced into electrical signals in the brain. The key step in this mechano-electrical transduction (MET) occurs in about a hundred ion channels atop each of the auditory hair cells in the cochlea (Kros et al, 1992). Gated by tip links between adjacent stereocilia in the hair bundles these MET channels, when open, allow current flow from the endolymph into the 11 hair cells, depolarizing them. Just two degrees of hair bundle deflection open 90% of the channels (Géléoc et al, 1997) to encode the loudest sounds. An adaptation process dependent on influx of Ca2+ ions through the MET channels keeps the hair cells optimally sensitive to small changes in sound intensity, and has been extensively studied in hair cells from lower vertebrates. Recent evidence suggests that in mammalian auditory hair cells adaptation is similarly Ca2+ dependent and rapid, with Ca2+ acting at or near the MET channel itself (Corns, Johnson, Kros & Marcotti, in preparation). The unusual characteristics of ion permeation through the MET channels, favouring the entry of Ca2+ ions by a vestibule at the extracellular side of the pore, also makes the cells vulnerable to ototoxic damage by aminoglycoside antibiotics such as gentamicin and dihydrostreptomycin. The aminoglycosides enter hair cells through the MET channels and are effectively trapped once inside (Marcotti et al, 2005). We are currently screening for compounds that compete with the aminoglycosides for entry into the hair cells, as a potential means to reduce the ototoxicity of these otherwise clinically useful drugs. Heretically, MET currents with unusual properties can be found under a variety of conditions in which tip links are absent (Marcotti et al, 2014). These large currents occur predominantly in response to stimuli of the opposite polarity to those that generate normal MET currents and have slower kinetics. The underlying ion channels appear to lack the vestibule of the normal MET channels and may be MET channel precursors situated at the base of the stereocilia. Acknowledgements Work in the Kros lab is supported by the MRC. References Géléoc G.S.G., Lennan G.W., Richardson G.P. & Kros C.J. 1997. A quantitative comparison of mechanoelectrical transduction in vestibular and auditory hair cells of neonatal mice. Proc Biol Sci, 264, 611-621. Kros C.J., Rüsch A. & Richardson G.P. 1992. Mechano-electrical transducer currents in hair cells of the cultured neonatal mouse cochlea. Proc Biol Sci, 249, 185-193. Marcotti W., van Netten S.M. & Kros C.J. 2005. The aminoglycoside antibiotic dihydrostreptomycin rapidly enters hair cells through the mechano-electrical transducer channels. J Physiol, 567, 505-521. Marcotti W., Corns L.F., Desmonds T., Kirkwood N.K., Richardson G.P. & Kros C.J. 2014. Transduction without tip links in cochlear hair cells is mediated by ion channels with permeation properties distinct from those of the mechano-electrical transducer channel. J Neurosci, 34, 5505-5514. Symposium abstracts From hair cells to hearing abstracts The sensory hair cell in normal hearing and disease D.N. Furness. Institute for science and technology in hearing, Keele University. Hair cells are the sensory receptor of our hearing and balance systems. Their ability to detect mechanical stimuli caused by sound and head movements is based on a structure at their tops called the hair bundle, composed of a precisely organised array of minute hairs (called stereocilia). This bunde, as well as being physically highly sensitive, for istanve to high impact sounds or certin drugs, is the target of a number of genetic abnormalities which underlie conditions such as Usher syndrome and other hearing impairments. One of the prinicpal targets in Usher syndrome is the tip 12 link, a structure found within the hair bundle that is composed of dimers of two molecules, cadherin 23 and protocadherin 15. Another common hearing impairment is associated with one of the candidate proteins for the hair-cell transducer channels, TMC1. Mouse mutations in these proteins have revealed much about how defects in them cause hearing loss by targeting the hair bundle and affecting both its structure and its ability to detect mechanical stimuli. Coding of acoustic information in the auditory pathway Professor Alan R. Palmer, Director, MRC Institute of Hearing Research, University Park, Nottingham, NG7 2RD, UK. As a result of the mechanical action of the basilar membrane and transduction in the cochlear hair cells, responses of auditory nerve fibres are tuned like a series of overlapping band pass filters allowing a good representation of the frequency content of any sound, which becomes less clear at high levels as the filters broaden. Activity in the auditory nerve signals the frequency content, the timing and the sound level of the sounds. Pathways from the first brainstem nucleus (the cochlear nucleus) converge in the brainstem to allow combination of information from the two ears for analysis of the location of the sound source, which is then sent to the auditory midbrain. Pathways from the cochlear nucleus also send information about the sound spectrum and its pitch directly to the auditory midbrain where it is integrated with inputs from all lower brainstem nuclei, before sending the information on to the auditory cortex via the thalamus. The auditory cortex has several frequency mapped areas which process sounds in parallel. There is some evidence for processing of different aspects of sounds in different cortical areas, giving rise to suggestions of different anatomical and functional processing streams for different aspects of sound perception. The deeper layers of the cortex send projections back down the auditory pathway enabling the cortex to modulate the ascending flow of auditory information. Integrating acoustic and electric information following cochlear implantation P.T. Kitterick. NIHR Nottingham Hearing Biomedical Research Unit, Nottingham, UK A cochlear implant evokes neural activity in the auditory nerve by delivering a train of electric pulses into the cochlea via a micro-electrode array. The amplitude of the pulses at each electrode is modulated by an externally-worn speech processor to encode the temporal envelope of sound energy within a particular frequency band. A unilateral cochlear implant provides sufficient temporal and spectoral information to support speech understanding at favourable signal-to-noise ratios, and bilateral cochlear implants provide access to inter-aural level cues which can support sound source localisation. Traditional candidates for cochlear implantation in the UK have a bilateral profound hearing loss. Thus, their listening abilities following implantation primarily reflect access to the electric signals from the implant. However, some UK recipients may have access to residual acoustic hearing at low frequencies, and recipients in other healthcare systems are being implanted with increasing levels of residual acoustic hearing in the non-implanted ear. While patients with well-preserved access to residual acoustic hearing report subjective differences in quality between the electric and acoustic signals, benefits arising from simultaneous access to both signal modalities have been demonstrated using tests of speech perception and localisation. Vocoder studies with normal-hearing listeners suggest that additional benefit may be obtained if electric and acoustic information is delivered to neural populations with similar characteristic frequencies in each ear. Acknowledgements Supported by infrastructure funding from the National Institute for Health Research. 13 Listening difficulties and auditory processing disorder (APD) in children: proposed mechanisms D.R. Moore* * Communication Sciences Research Center, Cincinnati Children’s Hospital and Department of Otolaryngology, University of Cincinnati, Ohio, USA; § School of Psychological Sciences, University of Manchester. Developmental APD is sometimes diagnosed for children presenting with listening difficulties (LiD), but normal audiograms, who score poorly on tests mostly of speech perception. However, most children with LiD, whether or not diagnosed APD, have difficulties with cognitive functions, also associated with other learning disorders. Here, I discuss two possible mechanisms underlying LiD/APD in children (see also Moore, 2014), first, that some children have ‘suprathreshold’ hearing loss, and second, that cognitive insufficiency may be modelled by dichotic listening. A link with cochlear deficiencies is suggested by (i) impaired temporal perception (e.g. frequency discrimination), which may indicate impaired phase locking in the brainstem, (ii) peripheral processing abnormalities (e.g. acoustic reflexes, MOC inhibition), found in some children with LiD, (iii) children with auditory neuropathy, without hearing loss, can also have LiD/APD. Dichotic listening, used frequently to diagnose APD, is also used in cognitive neuroscience to study executive function, strongly and specifically implicated in LiD/APD. By comparing the ‘free recall’ of dichotic CV syllables with ‘forced attention’ to right or left ear, Hugdahl (2009) has discovered a fronto-parietal network involved in top-down attention regulation of bottom-up auditory processing in individuals with learning disorders. We are currently examining this network in children with LiD/APD. References Hugdahl K, Westerhausen R, Alho K, Medvedev S, Laine M, & Hämäläinen H. (2009). Attention and cognitive control: unfolding the dichotic listening story. Scand J Psychol, 50,11-22. Moore, D.R. 2014. Sources of pathology underlying listening disorders in children. Int J Psychophysiol. in press. PPC symposium Abstract HOW YOU CAN MAKE AUDIOLOGY BETTER. A symposium by the BSA Professional Practice Committee. You, the audiological professional, are fundamental to the development of best clinical practice within the UK and you have an obligation to ensure that others in both your and allied professions benefit from your knowledge, experience and ideas. Whether a student, recently qualified or are following a career in audiology, you can make a significant contribution to making audiology better. The principle aim of the BSA is the advancement of Audiology. This not only includes the scientific study of hearing and balance function and other related sciences, but also, as importantly, the advancement of best clinical practice in the diagnosis, management and rehabilitation of hearing and balance and allied disorders. The development and promotion of good clinical Audiological practice at all stages of care by the BSA is facilitated by its Professional Practice Committee which provides and disseminates guidance on good practice which is demonstrably high on relevance, quality and impact. This guidance is achieved through the provision of recommended procedures, guidelines, minimum training standards and training days. 14 As part of the ongoing process to optimise the impact of the BSA, the PPC is in the process of submitting an application to NICE for accreditation of its guidance. This symposium will include a presentation by Deborah Collis, Associate Director of Accreditation and Quality Assurance for NICE, who will be speaking about the accreditation process and its potential benefits. This symposium will demonstrate the key role that you should be playing in the development and promotion of good Audiological practice and discuss how this may best be achieved in order to have greatest impact. This will include exploring ways by which you can effectively engage in and contribute to the process of making audiology better. UK Biobank symposium abstract Large scale hearing studies using the UK Biobank resource P. Dawes1, H. Fortnum2, D.R. Moore2,3, R. Emsley4, P. Norman5, K. Cruickshanks6, A.C. Davis7, M. Edmondson-Jones2, A. McCormack2, Robert Pierzicki2 and K.J. Munro1,9 ,1 School of Psychological Sciences, University of Manchester, UK, 2 NIHR Nottingham Hearing Biomedical Research Unit, Nottingham, UK, 3Cincinnati Children’s Hospital Medical Center, Cincinnati, USA, 4Centre for Biostatistics, Institute of Population Health, University of Manchester, UK, 5 School of Geography, University of Leeds, UK, 6 Population Health Sciences and Ophthalmology and Visual Sciences, School of Medicine and Public Health, University of Wisconsin, USA, 9Central Manchester Universities Hospitals NHS Foundation Trust, UK The UK Biobank is a large data set established for investigations of the genetic, environmental and lifestyle causes of diseases of middle and older age. Over the course of 2006-2010, 503,325 UK adults between the ages of 40 to 69 years were recruited. Participants responded to questionnaire measures of lifestyle and demographic factors, performed a range of physical measures and donated biological samples. A subset of 164,770 participants completed a hearing test (the Digit Triplet Test, a measure of speech recognition in noise). During 2012 to 2013, 17,819 participants completed a repeat assessment, including 4,425 participants who completed the hearing test at both time points. A multi-disciplinary team including researchers from Manchester, Nottingham, Leeds, Cincinnati and Wisconsin have been collaborating to analyse hearing and tinnitus data from the UK Biobank. In this symposium, we report the first analyses, including i) an overview of patterns of hearing impairment and hearing aid use, ii) cigarette smoking and alcohol consumption as risks for hearing loss, iii) relationships between speech recognition in noise, audiometric thresholds and cognitive performance and iv) how associations between cognitive performance and hearing loss may be mediated by hearing aid use and social isolation. Acknowledgements This research was based on data provided by UK Biobank. Additional support was provided by the National Institute of Health Research (Nottingham Hearing BRU and Manchester BRC), the Medical Research Council, and Cincinnati Children’s Hospital. DRM was supported by the Intramural Programme of the Medical Research Council [Grant U135097130]. KJC was supported by R37AG11099, R01AG021917 and an Unrestricted Grant from Research to Prevent Blindness. The Nottingham Hearing Biomedical Research Unit is funded by the National Institute for Health Research. This paper presents independent research funded in part by the National Institute for Health Research (NIHR). The views expressed are those of the author(s) and not necessarily those of the NHS, the NIHR or the Department of Health. This research was facilitated by Manchester Biomedical Research Centre. This research was conducted using the UK Biobank resource. 15 Special Interest Group Session Abstracts APD BSA Auditory Processing Disorder SIG update: Onwards and Upwards N.G. Campbell* and D.R. Moore§, *Institute of Sound and Vibration Research, University of Southampton, UK, §Communication Sciences Research Center, Cincinnati Children's Hospital, Cincinnati, USA Over the past 3 years the SIG has shaped international thinking. We published a peer-reviewed Position Statement and Practice Guidance Document (2011), and an ‘APD White Paper’ (2013) in the International Journal of Audiology. We initiated and collaborated with the American Academy of Audiology (AAA) to present very successful global APD days as part of the AAA conferences in Boston (2012) and in Orlando (2014). Since 2011, and our call for evidence based practice, there has been a surge in the number of randomised control studies. Finally, we hosted an APD Satellite Day at the 2013 BSA Annual Conference. This presentation reviews those past achievements and new and future developments in tests, resources and technology, and our new APD Position Statement and Practice Guidance Document. In our revised approach, the parcellation into Developmental, Acquired and Secondary APD, which has proven useful and popular, is retained. For Developmental APD, we emphasize the goal of management of children, with a more hierarchical approach to testing and diagnosis. Assessment starts with questionnaires, followed by testing; where-after individuals are guided towards various interventions that suit their specific needs. Theoretic underpinnings become more biological and, therefore, testable, preventable or treatable. Acknowledgements We acknowledge the anonymous reviewers, members of the BSA Professional Practice Committee and all that took part in the public consultation for the Position Statement and Practice Guidance Document. We further acknowledge the commentaries of the leading international researchers that formed part of the BSA ‘APD White Paper’. References Moore D.R., Rosen S., Bamiou D., Campbell N.G. & Sirimanna T. (2013). Evolving Concepts of Developmental Auditory Processing Disorder (APD): A British Society of Audiology APD Special Interest Group ‘White Paper’. IJA, 1-11. BSA APD SIG Steering Committee: Moore D.R., Alles R., Bamiou D., Batchelor L., Campbell N.G., Canning D., Grant P., Luxon L., Murray P., Nairn S., Rosen S., Sirimanna T., Treharne D. and Wakeham, K. (2011). BSA Position Statement: APD. http://www.thebsa.org.uk/resources/apd-position-statement/ BSA APD SIG Steering Committee: Campbell N.G., Alles R., Bamiou D., Batchelor L., Canning D., Grant P., Luxon L., Moore D.R., Murray P., Nairn S., Rosen S., Sirimanna T., Treharne D. and Wakeham, K. (2011). BSA Practice Guidance: An Overview of Current Management of APD.http://www.thebsa.org.uk/resources/overview-current-management-auditory-processingdisorder-apd/ BIG The Video Head Impulse Test and its relationship to caloric testing SL. Bell*, F. Barker, §, E.Mackenzie*, H.Hesleton*., C.Parker*, A.Sanderson*, *Hearing and Balance Centre, University of Southampton UK, § Windsor ENT, Berkshire 16 The subjective Head Impulse Test (HIT) (Halmagyi and Curthoys, 1988) was proposed to indicate the status of the vestibular occular reflex (VOR) at high frequencies. It relies on direct observation of the eyes whilst rapid short duration impulses are applied to the subject’s head. The presence of overt saccades are an indirect indication of peripheral abnormality in the canal being stimulated. The Head Impulse Test tests the VOR at higher frequencies than the caloric test (Jorns-Haderli et al., 2007). However it relies on the detection of overt saccades by the observer. The video Head Impulse Test (vHIT) MacDougall et al. (2009) has been developed as an alternative system suitable for routine clinical use to make head impulse testing more objective. The system consists of software and lightweight goggles containing both an eye position camera that can track the subject's pupils, and a gyroscope that can track the angular movement of the head. The system records head impulse velocity, together with eye movement velocity in response to high velocity impulses that are applied to the head by the clinician. Impulses can be applied in the three planes of the semicircular canals. During this test the head is moved with acceleration that should be sufficient to cause the afferents of the semicircular canal on one side to be completely inhibited. Hence a semicircular canal on one side can be tested in effective isolation from the same canal on the other side This study has two aims: To explore normative ranges for vHIT gain and to compare the results of vHIT testing with calorics in a sample of patients attending a clinic for balance disorder. Two normative studies were conducted to explore the normal ranges of vHIT gain in different semicircular canals. A clinical sample of 51 patients (20 male, 31 female) were tested with both lateral canal vHIT and air calorics. Normative studies found that vHIT gain is near unity for lateral canals, but is significantly raised in the vertical canals. Care must be taken to avoid artefacts when recording from the vertical canals. vHIT gain appears relatively insensitive to peripheral vestibular disorder as indicated by caloric testing, with low sensisivity but fair specificity. vHIT gain is abnormal in canals with no measurable function. The vHIT does not appear well suited to screen and identify patients who require caloric testing although the test may give complementary information to caloric testing. Declaration of interest The author reports no conflicts of interest. The authors alone are responsible for the content and writing of the presentation. References Halmagyi GM, Curthoys IS. (1988) A clinical sign of canal paresis. Arch Neurol. ; 45(7): 737-9. Jorns-Haderli, M., Straumann, D. and Palla, A. (2007) Accuracy of the bedside head impulse test in detecting vestibular hypofunction, J Neurol Neurosurg Psychiatry, 78, 1113-1118. MacDougall, H. G., Weber, K. P., McGarvie, L. A., Halmagyi, G. M. and Curthoys, I. S. (2009) The video head impulse test- diagnostic accuracy in peripheral vestibulopathy, Neurology, 73, 1134-1141.. Motivational approach to behaviour change in vestibular rehabilitation to improve clinic attendance N. Topass, Audiology Department, Royal Surrey County Hospital, UK This study aims to compare how the use of a motivational approach to behaviour change would improve patient clinic attendance and thus therapy adherence. Vestibular rehabilitation is the recommend primary treatment for stable vestibular lesions (Shepard et al, 1995). The prognosis for uncompensated peripheral vestibular lesions is generally very good with the percentage of patients 17 who dramatically or completely improve set at 90% (Shepard et al, 1995). Adherence to vestibular rehabilitation programs can however prove to be difficult as is the case in many chronic health conditions. An important part of learning a new behaviour is for the patient to identify and acknowledge the value of the new behaviour (Konle-Parker, 2001). Motivational interviewing is a counselling technique which has been developed to help a patient to explore and resolve ambivalence related to behaviour change (Emmons & Rollnick, 2001). A review of our clinic attendance was compiled to determine the efficacy of the customised vestibular rehabilitation program. The data was reviewed in terms of: attendance, did not attends(DNAs); and cancellation history for the six month period from September 2012 to February 2013. The percentage of DNA’s was 12.6% and the late cancellation rate was 15.4% (A late cancellation is made within 1 week of the appointment). A review of current literature revealed that patient motivation may be a key element to the high DNA rate for this particular speciality, thus a motivational approach to the patient pathway was introduced. The patient would then ‘opt in’ or ‘opt out’ of the therapy. The therapy program also included elements referred to as the ‘box’, the ‘line’ and the ‘circle’. The results of the change in protocol was assess for the period from September 2013 to February 2014. The percentage of DNA’s was then 4.7% and the patient late appointment cancellation rate was 4.7%. The conclusion is thus that a motivational approach to vestibular rehabilitation delivery can improve clinic adherence and thus patient adherence to therapy. Acknowledgements Supported by Community Health Psychology, Farnham Road Hospital, UK. References Emmons K. & Rollnick S. 2001. Motivational interviewing in health care settings opportunities and limitations. Am J Prevent Med, 20(1), 68-74. Konkle-Parker D.J. 2001. A motivational intervention to improve adherence to treatment of chronic disease, J Am Acad Nurse Pract, 13(2), 61-68. Shepard N., Telian S., & Michigan A. A. (1995). Programmatic vestibular rehabilitation. Otolaryngol-Head Neck Surg , 112, 173-182. Tønnesen H. 2012. Engage in the process of change: Facts and methods, WHO-CC Clinical Health Promotion Center Bispebjerg University Hospital Copenhagen, Denmark Health Sciences Lund University, Sweden Sponsor workshops Monday 1st September at 2 PM: Oticon (Alison Stone) 14:00 – 14:30 Evidence for a new compression strategy for children 14:45 – 15:15 New clinical tools for paediatric fittings 15:25 – 15:45 Connectivity for Kids Monday 1st September at 4.15 PM: Interacoustics Title TBA Tuesday 2nd September at 11 AM: The Tinnitus Clinic 18 Mark Williams: Tinnitus Pathophysiology and Evolving Clinical Options for Therapy. The investigation of tinnitus neural correlates within mammals has enabled the development of novel therapeutic approaches that strive to reduce or reverse pathological central changes in humans afflicted with bothersome auditory hallucinations. This talk will review the pathophysiology of subjective tinnitus and introduce evolving therapeutic intervention options. Tuesday 2nd September at 2 PM : GN Resound Paul Leeming, Support Audiologist 19 Oral Presentation Abstracts: Tuesday afternoon, 1400-1545 Auditory processing deficits in children with mild to moderate sensorineural hearing loss (Listening problems with hearing loss) L. F. Halliday, O. Tuomainen, and S. Rosen, Division of Psychology and Language Sciences, University College London, UK Impaired auditory processing has been proposed as a primary deficit in developmental disorders of language, including dyslexia and specific language impairment (Goswami, 2011; Tallal, 2004). However, it remains uncertain whether deficits in auditory processing contribute to the language difficulties experienced by many children with mild to moderate sensorineural hearing loss (MMSNHL). We assessed the auditory processing and language skills of 49, 8-16 year old children with MMSNHL and 41 age-matched typically developing controls. Auditory processing abilities were assessed using child-friendly psychophysical techniques. Discrimination thresholds were obtained for stimuli spanning three different timescales (μs, ms, seconds), and across three different levels of complexity, from simple nonspeech tones (frequency discrimination, frequency modulation detection, rise-time discrimination), to complex nonspeech sounds (assessing discrimination of modulations in formant frequency, fundamental frequency, and amplitude), and speech sounds (/ba/-/da/ discrimination). Thresholds were obtained both when children with MMSNHL were wearing hearing aids and when they were not. Language abilities were assessed using a battery of standardised assessments of phonological processing, reading, vocabulary, and grammar. Principal components analysis identified three components underlying the auditory processing test battery. The first, auditory processing (‘AP’) component appeared to reflect discrimination of temporal fine structure and spectral shape, and was significantly impaired in children with MMSNHL. Deficits in this component were almost always present in children with MMSNHL who had difficulties in language (i.e. they were necessary to cause language difficulties) but many children with deficits in AP had normal language (i.e. they were not sufficient). The second, ‘AMD’ component reflected discrimination of slow-rate amplitude modulation. Deficits in this component were rare in children with MMSNHL, including those who had poor language (they were not necessary) but when they were present they were almost always associated with poor language abilities (they were sufficient). Finally, deficits in the third, speech processing (‘SP’) component appeared to be neither necessary nor sufficient to cause language difficulties in children with MMSNHL. Our findings challenge the proposal for a one-to-one causal relationship between deficits in auditory processing and language difficulties in children with MMSNHL. Rather, they suggest that deficits in the discrimination of temporal fine structure and spectral shape are likely to comprise one of a number of risk factors for impaired language development in this group. Acknowledgements Supported by the ESRC. References Goswami, U. 2011. A temporal sampling framework for developmental dyslexia. Trends Cog Sci, 15, 3-10. Tallal, P. 2004. Improving language and literacy is a matter of time. Nat Rev Neurosci, 5, 721-728. Assessing the role of parental report (ECLiPS) in supporting assessment of children referred for auditory processing disorder 20 J.G. Barry*, D. Tomlin, §,#, D.R. Moore*,$, H. Dillon‡,# *MRC Institute of Hearing Research, Nottingham, UK; § University of Melbourne, Department of Audiology and Speech Pathology, Melbourne, Australia; ‡National Acoustics Laboratory, Sydney, Australia; #The Hearing Cooperative Research Centre, Melbourne, Australia; $Communication Sciences Research Center, Cincinnati Children’s Hospital and Department of Otolaryngology, University of Cincinnati, Ohio, USA Children referred for auditory processing disorder (APD) present with a complex array of symptoms, in addition to listening difficulties. This has led to a debate about the nature of the APD, and the assessment and management of children affected by it. To progress understanding of APD, it is necessary to reliably capture the reason(s) why children present for assessment, and then relate those reasons both to the disorder itself, as well as to the broader range of listening-based learning difficulties. The Evaluation of Children’s Listening and Processing Skills (ECLiPS) (Barry & Moore, 2014) was developed as a first step towards achieving these goals. The ECLiPS is a reportbased measure comprising a five factor structure which is designed to assess everyday listening in the context of cognitive abilities commonly affected in APD. Here, we performed a series of correlational and discriminant analyses to compare parental responses on the ECLiPS for 50 children (35 referred for suspected APD) with their performance on 5 tests used to diagnose APD, 4 tests of cognitive ability (nonverbal IQ, sustained attention and auditory serial and working memory, and 2 measures of academic ability. Few correlations were observed between the ECLiPS and the diagnostic tests of APD, confirming previous conclusions of a mismatch between abilities probed by clinical tests of APD and report-based measures of listening difficulty. Correlations with all ECLiPS factors were observed with academic abilities (rs = 0.52 – 0.40). Finally, auditory working memory correlated with all ECLiPS factors (rs = 0.30– 0.45), while attention was associated with language abilities (rs = 0.57). Discriminant analysis suggested the ECLiPS and a combination of auditory and cognitive tests was able to discriminate between the groups included in the stud with 83% accuracy. Overall, our findings suggest cognitive difficulties associate with many of the symptoms of APD that parents are sensitive to, and they further underline the importance of a broad-based assessment test battery incorporating parental report and cognitive testing. Acknowledgements Supported by the MRC, and MRC-T References Barry, J.G. & Moore, D.R. (2014). Evaluation of Children’s Listening and Processing (ECLiPS) Manual, Edition 1. Do children with hearing loss show atypical attention during ‘cocktail party’ listening? E. Holmes*, P.T. Kitterick§ and A.Q. Summerfield*, *Department of Psychology, University of York, York, UK, §NIHR Nottingham Biomedical Research Unit, Nottingham, UK Individual differences in cocktail-party listening could arise from differences in either central attention or peripheral transduction. We aimed to isolate differences in attentional processing between normally-hearing and hearing-impaired children, based on a procedure devised by Hill and Miller (2010). During the presentation of acoustical stimuli, we expected to see differences in brain activity between normally-hearing and hearing-impaired children as a result of differences in peripheral processing of speech. Differences before the onset of acoustical stimuli were expected to 21 show differences in the control of attention without being confounded by differences in transduction. Participants were 24 normally-hearing children and 11 children with moderate sensorineural hearing loss, all aged 7-16 years. We recorded brain activity using 64-channel electroencephalography. On each trial, two sentences spoken by adult talkers (one male and one female) were presented simultaneously from loudspeakers at two spatial locations (one left and one right of fixation). A third ‘child’ talker was presented from straight ahead. Participants were cued, in advance of the acoustic stimuli, to either the location (left/right) or the gender (male/female) of the target talker. The task was to report key words spoken by that talker. A control condition, in which the visual cues did not have implications for attention, allowed cortical activity evoked by the configurational properties of the cues to be distinguished from activity evoked by attentional processes. Children with normal hearing displayed differences in event-related potentials (ERPs) between trials that cued the location compared to the gender of the target talker. Evidence of cue-specific preparatory attention, manifest as more positive ERPs evoked by the cue to location than gender in posterior scalp locations, started approximately 950 ms after the onset of the visual cue. Sustained cue-specific selective attention started approximately 350 ms after the onset of acoustical stimuli. During both of these time windows, children with hearing loss showed smaller differences between location and gender trials than normally-hearing children. We expected to find atypical activity during the acoustical stimuli, which could be explained by peripheral transduction. Importantly, we also found atypical brain activity during the preparatory phase, which cannot be attributed to peripheral transduction. The results suggest that children with hearing loss do not utilise location or gender information to prepare for an upcoming talker. This deficit may contribute to their difficulty understanding speech in noisy environments. Acknowledgements EH is supported by the Goodricke Appeal Fund. References Hill K.T. & Miller L.M. 2010. Auditory attentional control and selection during cocktail party listening. Cerebral Cortex, 20, 583-590. Temporal asymmetry and loudness : amplitude envelope sensitivity in developmental dyslexia S.A. Flanagan and U. Goswami, Centre for Neuroscience in Education, Department of Psychology, University of Cambridge, UK. One auditory difficulty found in children with developmental dyslexia is impaired processing of the amplitude envelope of a sound (Goswami et al, 2002). Children with dyslexia also show impaired ‘phonological awareness’, the ability to reflect on the constituent sound elements in words. Deficits in amplitude envelope processing are related to phonological deficits in dyslexia, across languages. Sensitivity to envelope structure and dynamics is critical for speech perception as it signals speech rate, stress, and tonal contrasts, and reflects prosodic and intonational information. Here we replicated the experiment of Stecker and Hafter (2000) to investigate temporal asymmetry of the envelope; the difference between sounds that increase in level slowly and have fast offsets (i.e., S-F) and those that have fast onsets and slow, damped offsets (i.e., F-S) in a paired comparison. These pairs of sounds have identical overall spectral contents, durations, and level. We report on the results of 14 subjects who were in the range 8 – 12 years, 7 of whom were dyslexic and 7 were typically developing readers. The results for the typical readers was similar to those found by Stecker and Hafter with an overall loudness advantage for S-F test tones, with the 22 effect attenuated by the context of a damped, S-F, standard. However, for the dyslexic subjects the results were consistently different from the typical readers, and those of Stecker and Hafter. For the dyslexic subjects the loudness advantage for S-F test tones was greatly reduced and in some instances reversed for both conditions of the standard. This is opposite to the general findings for the perception of ramped and damped sounds (Ries et al, 2008). Impaired neural entrainment to the onset envelope in developmental dyslexia as proposed in the temporal sampling framework (Goswami, 2011) could explain the reduced or absent loudness advantage for S-F tones. Acknowledgements Supported by the Daphne Jackson Trust and the University of Cambridge. References Goswami U. et al, 2002. Amplitude envelope onsets and developmental dyslexia: a new hypothesis. Proceedings of the National Academy of Sciences U.S.A. 99, 10911- 10916 Stecker G. C. & Hafter E. R. 2000. Temporal asymmetry and loudness. J Acoust Soc Am, 107, 3359 -3362 Ries D.T., Schlauch R.S. & DiGiovanni J.J. 2008 The role of temporal-masking patterns in the determination of subjective duration and loudness for ramped and damped sounds. J Acoust Soc Am, 124, 3772-3783 Goswami U., 2011 A temporal sampling framework for developmental dyslexia. Trends in Cognitive Sciences, 15(1), 3-10. Learning to ignore background noise in VCV test L.Zhang*, P.Jennings* and F. Schlaghecken§,*WMG, University of Warwick, UK, §Department of Psychology, University of Warwick, UK Experiments on perceptual learning in the visual domain show that people can improve their detection performance by learning to ignore (visual) 'noise'. Once participants have learned to ignore the constant visual noise and can successfully detect targets, this skill then transfers to new, random visual noise (Schubo, et al., 2001). This study explores whether this also follows for perceptual learning in the hearing domain. It aims to investigate if it is possible to improve our brains’ ability to process auditory stimuli by training a listener to ignore background noise and recognise the sounds of consonants across time. In addition, Felty et al (2009) demonstrated that listeners achieved better word recognition performance with fixed babble noise than with random babble noise. This research also investigates if the learning effect is generalised from training normal hearing listeners under fixed babble noise to random background environments. Twenty normal-hearing English native speakers (aged between18 to 40) participated in an experiment. They were randomly assigned to a fixed or random babble noise training goup. Both groups were required to do a pre and post test with vowel consonant vowel (VCV) tasks (including eight consonants /b/, /d/, /f/, /g/, /k/, /m/, /n/, /p/ with male and female voices) in random babble noise. The background noise for the three days training session was different for each group (the fixed group was trained with constant babble noise and the other one was trained with random babble noise). The results from this study show how people’s listening performance can be improved with training to ignore fixed and random babble noise. Acknowledgements We would like to thank Dr. James Harte, Dr. Katherine Roberts and Professor Christopher James for their useful suggestions for the experiment set up. 23 References Felty RA, Buchwald A, Pisoni D.B. 2009. Adaptation to frozen babble in spoken word recognition. Journal of the Acoustical Society of America. 125, 93-97. Schubo, A., Schlaghecken, F., & Meinecke, C. (2001). Learning to ignorethe mask in texture segmentation tasks. Journal of Experimental Psychology: Human Perception and Performance, 27, 919–931. Cognitive listening fatigue with degraded speech A. Sarampalis*, D. Alfandari*, J.H. Gmelin*, M.F. Lohden*, A.F. Pietrus-Rajman* and D. Başkent§, *Department of Psychology, University of Groningen, Groningen, The Netherlands, § Department of Otorhinolaryngology / Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands Degradations in the speech signal, either because of external factors (such as background noise, reverberation, or poor signal transmission) or internal factors (such as hearing loss or ageing) often lead to reduced intelligibility and increases in listening effort. We use the term listening effort to refer to the proportion of cognitive resources allocated to the task of understanding speech. There is now ample evidence that even when intelligibility remains unchanged, reductions in the quality of the speech signal result in increases of listening effort in a variety of tasks and paradigms, suggesting that basic intelligibility measures are insufficient in fully evaluating some important aspects of speech communication. One frequently-reported consequence of increased listening effort is mental fatigue. Elderly individuals and those with hearing problems often complain of fatigue in difficult listening settings. Nevertheless, there have been few attempts to investigate listening fatigue empirically. We use the term cognitive listening fatigue (CLF) to describe performance decreases in an auditory task as a result of sustained effort. The aim of the present study is two-fold: first to identify the most suitable tasks and methods for measuring CLF and second to investigate the hypothesis that mildly degraded speech leads to increased fatigue when compared to undegraded speech. We used four cognitively-demanding auditory tasks, some adapted in the auditory domain for the first time, each imposing different cognitive demands on the listener, such as selective attention, inhibition, and working memory in a procedure that lasted 120 minutes of uninterrupted testing. Specifically, an auditory Stroop task, a dichotic-listening task, a flanker task, and a syllogisms task were each repeated three times in 10-minute blocks of trials in a counterbalanced order. Participants’ reaction times and accuracy were recorded, as well as a subjective measure of their willingness to continue after each block. Half of the participants listened to undegraded speech, the others to spectrally degraded signals. Reaction times decreased throughout the entire testing session, for all four tasks. Participants became progressively faster in their responses, even when their accuracy did not markedly decrease. The greatest decrease in accuracy was seen in the dichotic listening task where, after an initial period of improvement (presumably due to practice), performance decreased in the latter part of testing. The subjective measure showed marked decreases in the willingness to continue with the experiment after approximately 70 minutes. There was little difference in these effects between the degraded and undegraded condition, which may have been due to the mild nature of the degradation. Overall, these results indicate that there was transfer of fatigue from one task to the next, but a single-task procedure may be more effective in measuring fatigue. Nevertheless, fatigue was evident in at least one of these tasks. 24 Declaration of interest: The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. A link between auditory attention and self perceived communication success in older listeners D. Kunke*, M. J. Henry*, R. Hannemann§, M. Serman§ and J. Obleser*, *Max Planck Research Group, „Auditory Cognition“, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany, §Siemens Audiologische Technik GmbH, Erlangen, Germany. Many studies show that current standard practice in audiology including pure tone audiometry and speech audiometry have limited application in gaging the every day communication success of individuals. The aim of this study was to investigate the relationship between auditory attention and every day listening skills in the older population. Auditory rhythmic sequences of low frequency tones (a same–different judgement of a standard and a temporally jittered target tone, separated by a series of 6 intervening distracting tones; similar to the Jones, et al, 2002 design) were used to test the performance of 38 older listeners between the ages of 50 and 80 years who presented with hearing thresholds of 20 dB HL or better in the task-critical frequency range (0.25 – 1 kHz). We employed this judgement task in three different conditions (discriminating the acoustic dimensions pitch, loudness and interaural time differences i.e. direction). Test-retest reliability was calculated and results show that the pitch discrimination task was the most reliable test (r > 0.85). Performance was assessed for each condition as perceptual sensitivity (d´), response bias (decision criterion, c), and reaction time. Participants were also asked to complete a short version of the (SSQ) Speech, Spatial and Qualities of Hearing Scale (Gatehouse & Noble, 2004). Sensitivity in the pitch task predicted significantly the self-assessed speech scale of the SSQ (r = 0.496, p = 0.002), while auditory threshold and speech audiometry results (Oldenburger sentence test and the Freiburger word test) did not (all p > 0.15). Not only do these findings imply a link between auditory attention and everyday communication success, but this can also be an important step towards a valid and reliable screening tool. Acknowledgements Supported by Siemens Audiologische Technik and the Max Planck Institute for Human Cognitive and Brain Sciences References Gatehouse S. & Noble W. 2004. The speech, spatial and qualities of hearing scale (SSQ). Int J Audiol, 43(2), 85-99. Jones M.R., Moznihan H., Mackenzie N. & Puente J. 2002. Temporal aspects of stimulus-driven attending in dynamic arrays. Psychological Science, 13(4), 313-319. 25 Oral Presentation Abstracts: Tuesday afternoon, 1615-1700 Preliminary results of a survey of experiences in listening to music via hearing aids Brian C.J. Moore and Sara M.K. Madsen, Department of Experimental Psychology, University of Cambridge, UK An internet-based survey was conducted of experiences in listening to music via hearing aids. Some key results are presented here. Valid responses were obtained from over 500 participants. Responses were analysed according to the self-reported degree of hearing loss: Mild, Moderate, Severe, and Profound. 36% of participants reported problems with acoustic feedback when listening to music, and the prevalence of such problems did not vary significantly with degree of hearing loss. A substantial number of participants reported problems such as “warbling effects” or “when (the organ) music STOPS, aids howl”, that are probably associated with feedback cancellation. In response to the question “When listening to music via radio, TV or stereo system, do you find your hearing aids to be helpful”, 64% of participants responded “Yes, a lot” or “Yes, a little”, but 13% responded “No, a bit worse” or “No, a lot worse”. In response to the question “When listening to live music, do you find your hearing aids to be helpful”, 53% of participants responded “Yes, a lot” or “Yes, a little”, but 17% responded “No, a bit worse” or “No, a lot worse”. The difference in responses for reproduced and live music probably reflects the higher sound levels that are typical of the latter. For live music, hearing aids tended to be more helpful for those with severe or profound hearing loss. While 61% of participants agreed with the statement that hearing aids “Make the music louder”, only 28% agreed with the statement that hearing aids “Help hear soft passages without the louder parts being too loud”. This indicates that the automatic gain control systems in the hearing aids were not performing as well as would be desired. 32% of participants reported that hearing aids “sometimes make the music seem distorted”, perhaps reflecting clipping or overload of the circuitry resulting from the high levels of live music. Only 21% of participants agreed with the statement that hearing aids “improve the tone quality of music”, while 29% agreed with the statement that hearing aids “worsen the tone quality of music”. Judgements of tone quality tended to be more positive for participants with mild hearing loss. Overall, the results point to a need to improve feedback cancellation systems and automatic gain control systems in hearing aids, and to increase the dynamic range of hearing aids to avoid distortion for high input levels. Acknowledgements Supported by the Starkey (USA). Mid-bandwidth loudness depression in hearing-impaired listeners J. L. Verhey* and J. Hots* *Department of Experimental Audiology, Otto von Guericke University Magdeburg, Magdeburg, Germany Beyond a certain critical bandwidth, loudness increases with bandwidth, an effect commonly referred to as spectral loudness summation. For bandwidths smaller than the critical bandwidth (i.e., subcritical bandwidths), it is commonly assumed that the loudness does not depend on bandwidth. This view was recently challenged by Hots et al. (2013, 2014). They observed in normal-hearing listeners a mid-bandwidth loudness depression, i.e., a decrease in loudness when the bandwidth was increased up to about the critical bandwidth. The level difference of the noise with a subcritical bandwidth and an equally-loud sinusoid could be as large as 8 dB. The effect was not restricted to a 26 certain frequency and increased as the level decreased. The present study investigated if such an effect was also observed in hearing-impaired listeners with a hearing loss of cochlear origin. The effect was studied at a centre frequency of 1500 Hz. In the frequency range around this frequency, the listeners had a flat hearing loss. The individual hearing loss ranged from 40 to 75 dB HL. Listeners were asked to adjust the level of the noise test signal to be perceived as equally loud as a 1500-Hz reference sinusoid. The test bandwidth ranged from 15 to 1620 Hz. The level of the reference signal was chosen individually based on the threshold at this frequency. It was set to be roughly equal to the loudness of a 30-dB sinusoid for normal –hearing listeners. In general, listeners showed a higher level of the noise than the sinusoid at equal loudness. For some listeners, the largest difference was found close to the normal critical bandwidth as in the normal-hearing listeners. For most hearing-impaired listeners, this maximum was shifted towards larger bandwidths, presumably reflecting the reduced frequency selectivity in these listeners. The effect was smaller than in normal-hearing listeners when compared at about the same reference loudness but was similar when compared at about the same reference level (70 dB SPL). The results of the hearing-impaired listeners argue against peripheral suppression as the underlying mechanism. The data are consistent with the hypothesis that the reduction in pitch strength with bandwidth affects loudness. Acknowledgements: This work was supported by the Deutsche Forschungsgesellschaft (SFB/TRR31). References Hots, J., Rennies, J., & Verhey, J. L. 2013. Loudness of sounds with a subcritical bandwidth: A challenge to current loudness models? J Acoust Soc Am, 134, EL334–EL339. Hots, J., Rennies, J., & Verhey, J.L. 2014. Loudness of subcritical sounds as a function of bandwidth, center frequency, and level,” J Acoust Soc Am, 135, 1313–1320. The effects of age and hearing loss on neural synchrony and interaural phase difference discrimination A. King, K. Hopkins and C. J. Plack,Department of Computer Science, University of Manchester, Manchester, UK Hearing difficulties occuring in middle-age and later may be partly due to reduced neural synchrony to the temporal characteristics of sounds (e.g. Ruggles et al, 2012). Poor neural synchrony may affect sensitivity to interaural phase difference (IPD), which may be used to locate and separate sounds from different angles around the listener. Neural synchrony can be recorded from the auditory brainstem as the electrophysiological frequency-following response (FFR; Worden & Marsh, 1968). This study aimed to determine whether changes in IPD sensitivity associated with hearing loss and age (as published in King et al, 2014) are underpinned by poor phase locking as defined by FFR strength. Listeners (N=36) varied in age (19-83 yr) and absolute threshold (−1-59 dB, SPL at 250 and 500 Hz). They discriminated—separately—IPDs in temporal fine structure (TFS) and envelope of 20 Hz amplitude-modulated (AM) tones carried by either 250 or 500 Hz tones. With absolute threshold partialled out, increasing age was correlated with poorer envelope-IPD discrimination, and to a lesser extent, TFS-IPD discrimination. With age partialed out, increasing absolute threshold was correlated with poorer TFS-IPD discrimination, but not envelope-IPD discrimination. FFR to four different AM tones was measured simultaneously with electroencephalography (tones presented dichotically, two to each ear). AM rates of 16, 27, 115 and 145 Hz were used, with carrier tones of 307, 537, 357 and 578 Hz respectively. Neither FFR at the AM rates, nor FFR at the 27 component frequencies, was correlated with absolute threshold. As such, poorer TFS-IPD discrimination with hearing loss could not be explained by neural desynchrony. FFR to the components, and to the 145 Hz AM rate, deteriorated with age irrespective of absolute threshold. Correlations between TFS-IPD thresholds and FFR at the component frequencies, and between envelope-IPD thresholds and FFR at the AM rates, were almost entirely explained by age. On the other hand, IPD thresholds increased with increasing age even when FFR strength was partialed out, suggesting other age-related factors also affect IPD sensitivity. Acknowledgements Supported by the UK Medical Research Council (grant ref: G1001609) and Oticon A/S. References King, A., Hopkins, K. & Plack C. J. 2014. The effects of age and hearing loss on interaural phase difference discrimination. J Acoust Soc Am 134, 342-351. Ruggles, D., Bharadwaj, H. & Shinn-Cunningham, B. G. 2012. Why middle-aged listeners have trouble hearing in everyday settings. Curr Biol, 22, 1417-1422. Worden, F.G. & Marsh, J.T. 1968. Frequency-following (microphonic-like) neural responses evoked by sound. Electroenceph Clin Neurophysiol, 25, 42-52. Oral Presentation Abstracts Wednesday afternoon 1430-1600 (parallel session 1) Changes in nitric oxide synthase in the cochlear nucleus following unilateral acoustic trauma B. Coomber, V.L. Kowalkowski, J.I. Berger, A.R. Palmer and M.N. Wallace, MRC Institute of Hearing Research, Nottingham, UK Nitric oxide (NO) is a potent neuromodulator in the mammalian brain. A free radical produced by the enzyme nitric oxide synthase (NOS), NO regulates plasticity via presynaptic and postsynaptic mechanisms. Neuronal NO production, and consequently neuronal NOS (nNOS) distribution, is widespread throughout the brain, including the auditory system (Guix et al., 2005). We have previously demonstrated that unilateral acoustic over-exposure (AOE) altered NOS expression in the cochlear nucleus (CN) of guinea pigs, in a manner that correlated with behavioural evidence of tinnitus (Coomber et al., 2014). Here, we further examined NOS expression in the ventral CN (VCN): specifically, the types of neurons that express NOS, the distribution across the rostral-caudal axis of the VCN, and the time course of AOE-induced NOS change, using immunohistochemical methods. NOS was present in a variety of principal cells – including spherical and globular bushy, multipolar, and octopus cells – spanning the entire rostral-caudal extent of the VCN. Unilateral AOE, and subsequent tinnitus development, resulted in strong asymmetrical differences, with both a larger number of NOS-positive cells and a greater amount of NOS on the ipsilateral side, occurring across the whole extent of the VCN. Notably, the mean soma size of NOS-positive neurons was significantly smaller in the ipsilateral VCN, compared with the contralateral VCN. Moreover, we found that changes in NOS expression – similar to that seen in tinnitus animals – were apparent as early as 24 h post-AOE, although asymmetries were highly variable between animals at all short-term intervals examined (up to 21 d). Asymmetrical NOS expression at these shorter-term intervals was not correlated with hearing level threshold shifts. Changes in excitability in both the VCN and dorsal CN (DCN) have been implicated in tinnitus generation (Kaltenbach & Afman, 2000; Kraus et al., 2011; Vogler et al., 2011). NOS, and 28 consequently NO, may be involved in pathological changes in the VCN that subsequently lead to tinnitus. Acknowledgements BC was supported by Action on Hearing Loss (International Project Grant G62). References Coomber B., Berger J.I., Kowalkowski V.L., Shackleton T.M., Palmer A.R. & Wallace M.W. 2014. Neural changes accompanying tinnitus following unilateral acoustic trauma in the guinea pig. Eur J Neurosci, in press. Guix F.X., Uribesalgo I., Coma M. & Muñoz F.J. 2005. The physiology and pathophysiology of nitric oxide in the brain. Prog Neurobiol, 76, 126-52. Kaltenbach J.A. & Afman C.E. 2000. Hyperactivity in the dorsal cochlear nucleus after intense sound exposure and its resemblance to tone-evoked activity: a physiological model for tinnitus. Hear Res, 140, 165-72. Kraus K.S., Ding D., Jiang H., Lobarinas E., Sun W. & Salvi R.J. 2011. Relationship between noise-induced hearing-loss, persistent tinnitus and growth-associated protein-43 expression in the rat cochlear nucleus: does synaptic plasticity in ventral cochlear nucleus suppress tinnitus? Neuroscience, 194, 309-25. Vogler D.P., Robertson D. & Mulders W.H. 2011. Hyperactivity in the ventral cochlear nucleus after cochlear trauma. J Neurosci, 31, 6639-45. Amplitude-modulation detection and discrimination in the chinchilla ventral cochlear nucleus following noise-induced hearing loss M. Sayles* and M. G. Heinz§*,*Department of Speech, Language and Hearing Sciences, Purdue University, West Lafayette, IN, USA, §Weldon School of Biomedical Engineering, Purdue University, West Lafayette, IN, USA Amplitude modulation (AM) is a common feature of natural sounds. Modulation supports perceptual segregation of "objects" in complex acoustic scenes, and provides information for speech understanding and pitch perception. Previous work in our laboratory demonstrated increased modulation gain, without change in modulation-transfer-function bandwidth for auditory-nerve fibers (ANFs) in hearing-impaired (HI) chinchillas, compared to normal-hearing (NH) controls, when evaluated in quiet conditions using high modulation-depth narrowband signals (Kale & Heinz, 2010; Kale & Heinz, 2012). However, it is unclear how (or if) this increased modulation gain relates to psychophysically measured improved modulation detection thresholds in HI vs NH human listeners (Moore & Glasberg, 2001; Füllgrabe et al, 2003). The ventral cochlear nucleus (VCN) is the first brainstem processing station of the ascending auditory pathway and an obligatory synapse for all ANFs. VCN neurons typically show enhanced spike synchrony to the amplitude envelope relative to their ANF inputs, when measured with high modulation-depth signals (e.g., Frisina et al, 1990). Using an approach based on signal-detection theory, the present study characterized the AM-detection and discrimination sensitivity of all VCN unit types in NH and HI chinchillas under quiet conditions and in background noise. The HI animals were exposed to 116dBSPL 500Hz-centered octave-band Gaussian noise for 2 hours under anaesthesia, and then allowed to recover for at least two weeks. Auditory brainstem response (ABR) thresholds were increased by ~25dB across the audiogram following this intervention. Distortion-product otoacoustic emissions (DPOAEs) were decreased by ~15dB, indicating some degree of outer hair-cell loss. We recorded spike times in response to sinusoidal amplitude-modulated (SAM) tones with modulation depths (defined as 20·log10(m), where m is the modulation index between 0 and 1) 29 between -30 and 0dB in 1-dB steps from all major VCN unit types (primary-like, primary-withnotch, transient chopper, sustained chopper, onset chopper, onset-L). Stimuli were presented in quiet, and in three levels of background noise (10, 15, 20dB r.m.s. relative to the SAM tone, 16.5kHz bandwidth). A statistical bootstrap method was used to express spike synchrony as d’ for AM detection and discrimination. Under quiet conditions, primary-like units showed worse AM detection thresholds in HI animals compared to NH controls. In contrast, chopper units showed improved AM detection thresholds (by ~5dB) in HI animals. This may be a neurophysiological correlate of the improved AM detection observed psychophysically in humans. However, in background noise AM detection thresholds increased in all unit types. Background noise had a greater effect in HI animals than in NH controls. These results have implications for models of temporal-information processing in the normal and impaired auditory system, especially in complex listening environments. Acknowledgements Supported by an Action on Hearing Loss Fulbright Commission scholarship (M.S.), and NIHNIDCD grant R01-DC009838. References Kale S. & Heinz M.G. 2010. Envelope coding in auditory nerve fibers following noise-induced hearing loss. JARO, 11, 657-673. Kale S. & Heinz M.G. 2012. Temporal modulation transfer functions measured from auditory nerve responses following sensorineural hearing loss. Hear Res, 286, 64-75. Moore B.C.J. & Glasberg B. R. 2001. Temporal modulation transfer functions obtained using sinusoidal carriers with normally hearing and hearing-impaired listeners. J Acoust Soc Am, 110, 1067-1073. Füllgrabe C., Meyer B. & Lorenzi C. 2003. Effect of cochlear damage on the detection of complex temporal envelopes. Hear Res, 178, 35-43. Frisina R.D., Smith R.L. & Chamberlain S.C. 1990. Encoding of amplitude modulation in the gerbil cochlear nucleus: I. A hierarchy of enhancement. Hear Res, 44, 99-122. Commissural improvement of sound level sensitivity and discriminability in the inferior colliculi L.D. Orton and A. Rees, Institute of Neuroscience, Newcastle University, Newcastle upon Tyne, UK The inferior colliculi (ICs) are the foremost nuclei of the auditory midbrain and the final subcortical locus in the auditory pathway where bilateral interactions occur. Direct processing between the ICs is mediated by the commissure of the inferior colliculi (CoIC) - connections that are reciprocal and tonotopic. We investigated if the CoIC has an influence on the representation of sound level in the spike output of single neurons in the ICs. Experiments were conducted in urethane anaesthetised, adult guinea pigs. Neural activity in the left IC was rapidly and reversibly deactivated by one of two distinct, yet complimentary techniques: cryoloop cooling (CC) or microdialysis of procaine (MDP), a short acting sodium channel blocker. We have previously verified the use of CC for this purpose (Orton et al, 2012). To confirm the validity of MDP in these experiments, we recorded multi-unit responses adjacent to the microdialysis probe in the left IC in response to pure tone stimuli. Neural activity was suppressed on commencement of MDP. Washout of procaine with artificial cerebrospinal fluid (aCSF) produced recovery of driven responses to control levels within 30 minutes. Histological analysis of the lesion produced by the microdialysis probe found a loss of just 1.49% (±0.31) of the total volume of the left IC. We generated rate-intensity-functions (RIFs) from the responses of 38 well discriminated single units in the right IC before, during and following 30 recovery from reversible deactivation of neural activity in the left IC in response to CC (n=22) or MDP (n=16). The slopes of RIFs were observed to shift to the right on deactivation so that a higher sound level was required to produce the same firing rate. The median level required to elicit halfmaximal firing rate in the population increased from 51 to 59 dB before recovering back to 51 dB. We computed receiver operating characteristics for each RIF using each position in the RIF as a target relative to all other responses. We then summed the absolute deviation from 0.5 in all responses in a RIF to produce a discriminability index (DI) to quantify how well changes in sound level could be discriminated from changes in firing rate along each RIF. The population of DIs reduced during contralateral IC deactivation. Similar effects were observed in response to CC and MDP. These findings demonstrate that: 1) MDP is an effective and reliable means of rapidly and reversibly abolishing spiking in guinea pig IC; 2) the CoIC has a significant influence on RIFs in IC; and 3) the ability of IC neurons to signal changes in sound level were decremented by removing intercollicular processing, suggesting commissurally mediated gain control enhances sound level encoding in the ICs. Acknowledgements This research was supported by BBSRC grant BB/J008680/1 References Orton L.D., Poon P.W.F. & Rees A. 2012. Deactivation of the inferior colliculus by cooling demonstrates intercollicular modulation of neuronal activity. Front Neural Circuits, 6, 100. Probing the physiology of perception: Invariant neural responses in ferret auditory cortex during vowel discrimination S.M. Town, K. C. Wood, H. Atilgan, G. P. Jones and J. K. Bizley, Ear Institute, UCL, London, UK Perceptual invariance is the ability to recognize an object despite variation in sensory input. For example, when listening to a violin solo the instrument has a constant identity, despite variations in sound waveform as different notes are played. Likewise in speech, we can recognize phonetic components, such as the vowel “u”, across talkers with different voice pitches. However, it is unclear how the brain supports perceptual invariance and specifically whether neurons in auditory cortex extract invariant representations of vowel identity across pitch. Here we study an animal model of perceptual invariance in which ferrets (n=4) were trained in a two-alternative forced choice task to discriminate synthetic vowel sounds. On each trial of the task, the subject was presented with two vowel sounds (250 ms in duration) and required to respond at a particular location depending on vowel identity. Across trials vowel pitch was varied and ferrets were required to generalize discrimination across pitch variation. During task performance, multiunit activity was recorded from microelectrodes positioned in auditory cortex. We identified sound-responsive units, selected as those whose firing rate during vowel presentation differed by ≥ 3 standard deviations from mean spontaneous activity. For each responsive unit, we asked if it was possible to decode vowel identity from the spiking responses observed across all pitches. Using a permutation test to assess significance, the majority of units (60%) were found to encode information about vowel identity across pitches. We also decoded vowel pitch across all vowel identities in a smaller proportion of units (20%). Of these units, approximately half (11% of the entire neural population) provided information about both vowel pitch and identity. Finally, neural responses recorded from approximately one third of units (38%) were informative about the behavioural choice of the animal. By using a classification procedure in which temporal parameters were free to vary, we were also able to estimate when in the neural response information was encoded. We found that information about vowel timbre arose early in the 31 response (0-100 ms after stimulus onset) whereas information about pitch and choice arose in the sustained and offset periods of the stimulus (200-400 ms). Our results show that auditory cortical neurons may offer a physiological substrate for invariant perceptual representations of sound. Furthermore, during behaviour information about both acoustic and perceptual parameters is represented in auditory cortex, potentially during different time windows of the neural response. Acknowledgements Supported by the BBSRC, Royal Society and Wellcome Trust Across-electrode variation in gap-detection and stimulus-detection thresholds for cochlear implant users R.P, Carlyon§,, J.M. Deeks§,, A.J. Billig§,, and J.A. Bierer‡, § MRC Cognition and Brain Sciences Unit, Cambridge, UK ,‡ University of Washington, Seattle, W.A, U.S.A. Cochlear implant (CI) users usually exhibit marked across-electrode differences in detection thresholds with "focussed" modes of stimulation, such as tripolar (TP) or partial-tripolar (pTP) mode. This may reflect differences either in local neural survival (e.g. "dead regions ") or in the distance of the electrodes from the modiolus (" electrode distance "). There is consideral interest in differentiating between these two possible sources of variability, because the former may have greater implications for speech perception and for possible re-programming strategies. However, detection thresholds typically vary much less across electrodes in the monopolar mode of stimulation implemented in contemporary speech-processing strategies. We reasoned that across-electrode variations in a supra-threshold task, when stimulation level was adjusted to produce a comfortably loud sensation across electrodes, would be more dependent on neural health than on electrode distance. We therefore measured stimulus-detection and gapdetection thresholds for at least four electrodes in each of ten Advanced Bionics CI users, using 1000-pps pulse trains. The electrodes were selected for each user to have a wide range of stimulusdetection thresholds. We also reasoned that listeners might exhibit substantial across-electrode variatioins in the gap detection task even in monopolar mode, as such variations have been observed in other supra-threshold tasks. Therefore all measures were obtained both in pTP and monopolar mode. Univariate ANCOVAs, with subject entered as a fixed factor, assessed the correlations between different measures with overall subject differences removed. Both stimulus-detection and gapdetection thresholds correlated highly significantly across modes (r=0.65 and 0.66, p<0.001 in both cases). However, there was no significant correlation between stimulus-detection and gap-detection thresholds in either mode. Hence gap detection thresholds likely tap a source of across-electrode variation additional to, or different from, that revealed by stimulus-detection thresholds. Stimulusdetection thresholds were significantly lower for apical than for basal electrodes in both modes ; this was only true for gap detection in pTP mode. Finally, although the across-electrode standard deviation in stimulus-detection thresholds was greater in pTP than in monopolar mode, the reliability of these differences – assessed by dividing the across-electrode standard deviation by the standard deviation across adaptive runs for each electrode – was similar for the two modes ; this metric was also similar across modes for gap detection. Hence reliable across-electrode differences can be revealed even using clinically available monopolar stimulation. Declaration of interest: The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 32 Source analyses of vestibular evoked potentials (VsEPs) activated during rhythm perception N.P.Todd*,C.S. Lee§, *Facuty of Life Science, University of Manchester, Manchester, UK, § Department of Psychology, Goldsmiths College, London, UK. Vestibular evoked potentials (VsEPs) of cortical, including frontal and temporal lobe, origin produced by short tone pips have recently been described (Todd at al 2014ab). These results suggest that vestibular inputs could potentially influence the perception of sound sequences, including the perception of rhythm (Trainor et al 2009). In the light of these developments, and a recently described new synthesis (Todd and Lee 2014), we report a reanalysis of prior data from an experiment to look for evidence of movement-relatedness in auditory evoked potentials (AEPs) (Todd 2003; Todd and Seiss 2004). In this experiment EEG was recorded while subjects listened to irregular, regular and syncopated rhythms, and then actively synchronised with the regular rhythm. The reanalysis makes use of a source model which allows a specification of the current activity over time in temporal and frontal lobe areas during passive listening to a rhythm, or active synchronization where it dissociates pre-motor and motor contributions. The model successfully captures the main features of the rhythm in showing that: (1) the metrical structure is manifest in greater current activity for strong compared to weak beats; (2) the introduction of syncopation produces greater activity than for a regular rhythm. In addition the outcomes of modeling suggest that: (1) activity in both temporal and frontal areas contribute to the metrical percept and that this activity is distributed over time, i.e. not localized to beat onsets; (2) it is confirmation of a beat following syncopation that produces the greatest activity associated with syncopation; (3) two distinct processes are involved in auditory cortex, corresponding to tangential and radial (possibly vestibular dependent) components of the temporal lobe current sources. Acknowledgements Partially supported by the Wellcome Trust. References Todd NPM (2003) The N2 is movement related. Proceedings of the British Society of Audiology Short Papers Meeting on Experimental Studies of Hearing and Deafness, September 2003. Todd NPM and Lee CS (2014) The sensory-motor theory of rhythm and beat induction 20 years on: A new synthesis and future directions. Front Neurosci, under review. Todd NPM, Paillard AC, Kluk K, Whittle E and Colebatch JG (2014a) Vestibular receptors contribute to cortical auditory evoked potentials. Hear Res, 309, 63-74. Todd NPM, Paillard AC, Kluk K, Whittle E and Colebatch JG (2014b). Source analysis of short and long latency vestibular-evoked potentials (VsEPs) produced by left versus right ear airconducted 500 Hz tone pips. Hear Res, 312, 91 – 102. Todd NPM and Seiss E (2004) Electrophysiological correlates of beat induction as internally- and externally-guided action. In SD Lipscomb, R Ashley, RO Gjerdingen and P Webster (Eds) Proc 8th Int Conf Music Percept Cog, Evanston, IL. (Causal Productions: Adelaide) pp 212-218. Trainor LJ, Gao X, Lei J, Lehtovaara K and Harris LR (2009) The primal role of the vestibular system in determining rhythm. Cortex 45, 35-43. Oral Presentation Abstracts Wednesday afternoon, 1430-1600 (parallel session 2) The psychosocial experiences of individuals with hearing loss E. Heffernan*, N. Coulson§, H. Henshaw* and M. Ferguson*, *NIHR Nottingham Hearing Biomedical Research Unit, University of Nottingham, UK, §Division of Rehabilitation and Ageing, University of Nottingham, UK 33 Hearing loss is not simply a physical condition; it also substantially affects psychosocial functioning (Espmark et al, 2002). Specifically, people with hearing loss (PHL) can experience difficulties with participation in social roles and activities, which may place them at risk of isolation and poor wellbeing (Arlinger, 2003). Theoretical models from health psychology have much potential to improve our understanding of the psychosocial experiences of PHL. One such model, the Common Sense Model (CSM) by Leventhal et al (1980), has been successfully applied to numerous chronic health conditions but not to hearing loss. The primary tenet of the CSM is that individuals have parallel cognitive and emotional responses to their health condition and that they select and evaluate coping strategies to manage these responses (Hagger & Orbell, 2003). The aim of this qualitative study was to explore the psychosocial impact of hearing loss using the CSM as a theoretical framework. Semi-structured interviews were conducted with 33 participants (eight hearing healthcare professionals and 25 adults with mild-to-moderate hearing loss). Maximum variation sampling was used to recruit diverse participants (Patton, 1990). Therefore, PHL who varied in terms of age, gender, occupation, hearing loss duration and hearing aid use were recruited. Also, various hearing healthcare professionals, including audiologists, hearing therapists and international opinion leaders on hearing loss, were recruited. Data collection was concluded when theoretical saturation was reached. The data are currently undergoing the thematic analysis process outlined by Braun and Clarke (2006). Preliminary results indicate that the CSM can provide valuable and unique insights into the psychosocial impact of hearing loss. The results of this study will help to inform the development of a measure that assesses the impact of hearing loss on social participation. Acknowledgements This work is part of an NIHR funded PhD studentship. References Arlinger S. 2003. Negative consequences of uncorrected hearing loss: A review. Int J Aud, 42, 2S17-2S20. Braun V. & Clarke V. 2006. Using thematic analysis in psychology. Qual Res Psych, 3, 77-101. Espmark, A-K.K., Rosenhall U., Erlandsson S. & Steen B. 2002. The two faces of presbyacusis: Hearing impairment and psychosocial consequences. Int J Aud, 41, 125-135. Hagger M.S. & Orbell S. 2003. A meta-analytic review of the common-sense model of illness representations. Psych Health, 18, 141-184. Leventhal, H., Meyer, D. & Nerenz, D. 1980. The common sense model of illness danger. In: S. Rachman (ed). Medical Psychology (Vol. 2). New York: Pergamon, pp. 7–30. Patton, M. Q. 1990. Qualitative Evaluation and Research Methods. New York: Sage Publications Inc. “You’re not going to make me wear hearing aids are you?” Factors affecting adolescent’s uptake and adherence to aural habilitation S.E.I. Aboagye*, M.A. Ferguson§, N. Coulson‡, J. Birchall§‡ and J.G. Barry*, *MRC Institute of Hearing Research, Nottingham, UK, §Nottingham Hearing Biomedical Research Unit, ‡Faculty of Medicine & Health Sciences, University of Nottingham, UK It has been suggested that uptake and adherence to aural habilitation among adults with hearing loss is determined by the perceived susceptibility and seriousness of the condition and the perceived benefits and barriers to intervention (Laplante-Lévesque et al, 2010). However factors affecting 34 uptake and adherence to habilitation among adolescents (11-18 years) with mild-to-moderate sensorineural hearing loss (20-70 dB HL) are complex and not well understood. Reports from teachers and clinicians suggest that this population, despite experiencing poorer speech perception and more effortful listening than their normally hearing peers, frequently reject hearing aids and other assistive technology. However, without appropriate support they are at risk of having difficulties with language development, academic achievement and psychosocial functioning (Bess et al, 1998). In the current study, 11-18 year olds with mild to moderate sensorineural hearing loss (n=20) participated in individual semi-structured interviews to identify situations and tasks at school or at home which caused them most difficulty and which could be targeted by habilitation. Cognitive and affective factors (Steinberg, 2005) impacting on uptake and adherence to previous aural habilitation were identified and interpreted using a self-regulatory model of health behaviour (Leventhal et al, 1980). Findings from this study will help delineate a framework for how psychosocial factors interact to affect behavioural outcomes among this population. The study represents a first step towards developing new habilitation programmes for their support. Acknowledgements Supported by the Medical Research Council. References Bess F.H., Dodd-Murphy J. & Parker R.A. 1998. Children with minimal sensorineural hearing loss: prevalence, educational performance, and functional status. Ear Hear, 19, 339-354. Laplante-Lévesque A., Hickson L. & Worrall L. 2010. Factors influencing rehabilitation decisions of adults with acquired hearing impairment. Int J Audiol, 49, 497-507. Leventhal H., Meyer D. & Nerenz D. 1980. The common sense representation of illness danger. Contr Med Psychol, 2, 7-30. Steinberg, L. 2005. Cognitive and affective development in adolescence. Trends Cogn Sci, 9, 69-74. Co-occurrence of cognition, communication and listening difficulties in children with APD, SLI or ASD M.A. Ferguson*, D.R. Moore§, *NIHR Nottingham Hearing Biomedical Research Unit, UK, § Cincinnati Children’s Hospital Medical Center, USA It has been suggested that children with developmental disorders, such as auditory processing disorder (APD), specific language impairment (SLI) and autistic spectrum disorder (ASD), have similar characteristics. Diagnosing children with APD can be problematic due to the lack of a ‘gold standard’ diagnostic test and suggestions that diagnosis is based on referral route (Ferguson et al., 2011). A multicentre population study (IMAP) tested 1476 normally-hearing children, aged 6-11 years (Moore et al., 2010) . The children were categorised according to the (i) Children’s Communication Checklist (CCC-2), and (ii) the Childrens Auditory Processing Performance Scale (CHAPPS). The CCC-2 is designed to screen for different types of communication impairments, and categorised children into groups with subprofiles of structural and pragmatic language difficulties, as well as autistic behaviours. The children were identified as being aligned to diagnostic categories of either language impairment (general, LI, or specific, SLI) or autistic spectrum disorder (ASD, or Asperger’s syndrome). The CHAPPS was used to categorise children as having either listening difficulties (overall score ≤ 5th percentile), or not. Cognition (nonverbal IQ, memory), language and reading were significantly poorer (p<.05) for the ‘clinical’ groups who had communication or listening difficulties compared to the typically 35 developing (TD) group. The exception was the Asperger’s group whose performance was similar to the TD group. Listening abilities were significantly poorer for all the ‘clinical’ groups compared to the TD children (p<.05). Furthermore, a significant proportion of children who had listening difficulties, were shown to have poorer structural (56%) and pragmatic (35%) language abilities, as well as demonstrating autistic behaviours (35%). The clinical implications are that APD co-occurs with LI and ASD, is likely to sit along the LIASD spectrum, and may not be a unique, discrete disorder. Children with listening difficulties who attend audiology clinics should be screened for functional everyday communication difficulties (e.g. CCC-2) to ensure appropriate onward referrals. A newly developed validated questionnaire, the ECLIPS (Evaluation of Children’s Listening and Processing Skills) could be a useful clinical tool to identify functional difficulties. Acknowledgements This research was funded by the National Institute for Health Research and Medical Research Council. References Ferguson M.A., Hall R.L., Riley A. & Moore D.R. 2011. Communication, listening, speech and cognition in children diagnosed with auditory processing disorder (APD) or specific language impairment (SLI). J Sp Lang Hear Res, 54, 211-227. Moore D.R., Ferguson M.A., Edmondson-Jones A.M., Ratib S. & Riley A. 2010. Nature of auditory processing disorder in children. Pediatrics, 126, e382-e390. X linked gusher syndrome – a rare cause for hearing loss in children S. Dasgupta, S.Raghavan, M.O’Hare and L.Marl, Department of Audiovestibular Medicine,and Audiology, Alder Hey Children’s NHS Foundation Trust, Liverpool A congenital mixed hearing loss with stapes fixation and a perilymph gusher at stapes surgery was first described by Olson and Lehman (1968) which was later established to have an X linked trait by Nance et al (1973). This X linked gusher syndrome is a very rare condition with only circa 50 cases reported in the English language. The aim of this study is to review the condition with our series of 2 children presenting with it. Patient 1, born in 2004 of African Liberian parentage was picked up to have a mixed hearing loss at the age of 2 years. The right side showed an air conduction threshold of 80 dBHL across the frequency range with a bone conduction threshold at an average of 55 dBHL while the left side indicated a profound hearing loss. He was aided early and then had a left cochlear implant. Patient 2, born in 2006 is patient 1’s half brother born of a Caucasian father. Picked up by the newborn hearing screening programme he had early amplification. His air conduction thresholds were in the 80 dBHL threshold range with bone conduction thresholds at the same level as his brother’s. However his hearing losses were symmetrical. In both brothers, tympanometry was normal whilst transient otoacoustic emissions were not recordable. MRI scan was normal in both. A high resolution CT scan(HRCT) with temporal bone windows demonstrated almost identical inner ear dysmorphic features including a loss of basal partition of the cochlea communicating with the internal auditory meatus(IAM) which were entirely consistent with X linked gusher syndrome. X linked gusher syndrome is characterised by a communication between the subarachnoid space and the perilymphatic space of the IAM and the cochlear basal turn due to an absent lamina cribrosa. Phelps et al (1991) in the largest series to date observed the features in HRCT appearing consistently which was later observed by Saylisoy et al (2014). The diagnosis of the condition rests on radiological findings and the characteristic mixed hearing losses. The hearing loss reported is 36 usually bilaterally symmetrical severe and can be progressive and present in late childhood or early adult life. Our cases are unique from several aspects reported for the first time here: their African race; a proven early congenital onset and non progression of the loss and the presence of a significant asymmetry. These different phenotypic presentations indicate the variable expressivity of the syndrome and must be considered when evaluating congenital mixed losses in children. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Olson N.R. & Lehman R.H. Cerebrospinal fluid otorrhea and the congenitally fixed stapes. 1968. Laryngoscope, 78, 352–359. Nance W.E., Setleff R., McLead A. et al. X-linked mixed deafness with congenital fixation of the stapedial footplate and perilymphatic gusher. 1973. Birth Defects Orig Artic Ser, 4, 64–69. Phelps P.D., Reardon W., Pembrey M. et al. X-linked deafness, stapes gushers and a distinctive defect of the inner ear. 1991. Neuroradiology, 33, 326–330. Saylisoy S., Incesulu A., Gurbuz M.K. & Adapinar B. Computed tomographic findings of X-linked deafness: a spectrum from child to mother, from young to old, from boy to girl, from mixed to sudden hearing loss. 2014. J Comput Assist Tomogr, 38, 20-4 What encourages or discourages hearing-impaired children to take part in sport? S.Wadsworth*,§, H. Fortnum*, §, A. McCormack*,§,‡ and I. Mulla#, *NIHR Nottingham Hearing Biomedical Research Unit, Nottingham, UK, § Hearing and Otology Group,Division of Clinical Neuroscience, School of Medicine, University of Nottingham, ‡ MRC Institute of Hearing Research, University Park, Nottingham, NG7 2RD, UK, #The Ear Foundation, Nottingham, UK Limited physical activity including a lack of participation in sport has been linked to poorer physical and mental health in adults and children (Bailey, 2006). Children are leading increasingly sedentary lifestyles and are at greater risk of chronic disease such as diabetes and obesity. Children who participate in sport benefit from improved social skills, better communication and selfconfidence and an improved quality of life (Kurkova et al, 2010). Hearing-impaired children should have the same opportunities to participate in sport and receive the benefits. Therefore the aim was to understand what encourages and discourages these children to take part in sport. This qualitative study uses focus groups with hearing-impaired children conducted in schools or through social clubs. Information was collected on the children’s thoughts and feelings about taking part in sport. Including their involvement in and access to sport, what they do and don’t like about sport, how they are involved in sport and who with. Thematic analysis of the focus group transcripts identified several key themes including friendships, family influence and communication. Children with a hearing impairment are more likely to take part in sport if they have adequate support for example children were more likely to try a sport if their friends and family played it and encouraged them to try it. Hearing impaired children with active siblings had tried more sports. Communication was a common theme throughout regardless of the type of technology a child used. Children found communication easier during sport if they were allowed to use their technology and had difficulty during sports such as swimming. Children expressed a desire not to have to identify themselves at a club as having a hearing loss. This affected their confidence when participating in sport. 37 Acknowledgements This research is supported by the National Institute for Health Research. References Bailey R. 2006. Physical education and sport in schools: a review of benefits and outcomes. J School Health, 76, 8, 397–401. Kurkova P., Scheetz N & Stelzer J. 2010. Health and physical education as an important part of school curricula: a comparison of schools for the deaf in the Czech Republic and United States. Am Ann Deaf, 155, 1, 78–95. Action on Hearing Loss. 2011. Taking action on hearing loss in the 21st Century. Assessment of current provision for auditory ototoxicity monitoring in the UK G. Al-Malky 1, M. De Jongh. 1, M. Kikic 1, S.J. Dawson 1, R. Suri 2,, 1UCL Ear Institute, London, UK, 2 Department of Respiratory Medicine, CF Unit, Great Ormond Street Hospital, London, UK Aminoglycoside antibiotics and chemotherapeutic agents such as cisplatin are two of the most established drugs associated with permanent ototoxicity. This side effect and others, such as nephrotoxicity, have limited the use of these treatments to patient groups that have to use them. These groups include oncology and cystic fibrosis (CF) patients. Repeated exposure to these medications needs continuous auditory monitoring for early identification of ototoxicity. This is currently more relevant because survival rates from these disorders have significantly improved, making maintenance of good quality of life essential. The American Speech-Language-Hearing Association (ASHA) and American Academy of Audiology (AAA) have produced the most comprehensive guidelines for ototoxicity monitoring to date (ASHA, 1994, AAA, 2009), however similar guidelines are not available in the UK. The aim of this study was to assess the current practice in UK services regarding ototoxicity monitoring using online surveys specifically designed for each of three targeted professions using the UCL Opinio survey tool. These three professions included audiology, oncology and CF unit clinicians. Hyperlinks to the surveys were mailed through respective professional bodies such as the British Academy of Audiology, Children’s Cancer and Leukaemia Group and UK CF Trust. Responses were obtained from 56 oncology, 133 audiology and 33 CF clinicians representing all regions of the UK. Encouragingly 69.6%, 63.9% and 79.3% of the three respective professions indicated that they do perform auditory monitoring. However there were substantial variations in the reported referral criteria, audiological tests used for monitoring, criteria for identification of ototoxicity, and indications for change in management. Oncologists confirmed that Cisplatin and Carboplatin were considered first line treatments by 94.6% and 78.6% of the respondents, whereas extended interval Tobramycin was the first line treatment by 97.0% of CF clinicians. More than 70% of the audiologist didn’t know whether patients were counseled about ototoxicity prior to treatment or not and 49.4% of them stated that baseline testing wasn’t routinely performed. Regarding the location where monitoring was performed, it was shown that more than 90% of testing was done in audiology departments indicating that there is limited access for un-well patients unable to reach them. Respondents reported that most centers used patients’ auditory complaints as the main reason for referral. This represents a drawback, as substantial cochlear damage would have occurred already. There was strong evidence of limited awareness/ agreement of responsibilities between team members so that audiologists’ roles were underutilized for counseling and rehabilitation and clinicians had limited understanding of the criteria for ototoxicity and test battery used to diagnose it. In conclusion, it was recommended to develop UK-wide clinical guidelines for ototoxicity monitoring and to develop professional education programmes, as advocated by the World Health Organisation (WHO, 1994), to increase the profile and standardisation of ototoxicity 38 monitoring in clinical practice. This can only be possible through active collaboration between the associated professional bodies. Acknowledgements: Special thanks are due to all the clinicians that responded to the survey. References AAA. 2009. American Academy of Audiology Position Statement and Clinical Practice Guidelines: Ototoxicity monitoring [Online]. http://www.audiology.org/resources/documentlibrary/Documents/OtoMonPositionGuideline .pdf: ASHA. [Accessed 13/10/2013. ASHA. 1994. American Speech-Language-Hearing Association Guidelines for the audiologic management of individuals receiving cochleotoxic drug therapy. [Online]. http://www.asha.org/policy/GL1994-00003.htm. [Accessed 03/10/2013. WHO 1994. Report of an informal consultation on strategies for prevention of hearing impairment from ototoxic drugs In: ORGANISATION, W. H. (ed.). http://www.who.int/pbd/deafness/ototoxic_drugs.pdf. Everything an Audiologist needs to know about Speech in Noise D. Hewitt, Audiology Department, Queen Alexandra Hospital, Portsmouth, UK I am an NHS Audiologist and a member of the Portsmouth team based at Queen Alexandra Hospital. Working in Adult Rehabilitation I am reminded every week that the problem for a significant number of people is hearing Speech in Noise and not hearing Speech in Quiet. Not quite so regularly but often enough I meet people who have normal hearing as measured by PTA but feel strongly they are worse than others when it comes to hearing Speech in Noise. And experience has taught me to ignore the marketing brochures and not promise patients that hearing aids will be of any great assistance in noisy places. All of this means that I do not feel I am helping my patients with their hearing problems as much as I want to and that my theoretical knowledge has too many gaps. My theoretical training was the Audiology MSc at the Southampton University ISVR and my dissertation required the digestion of many Speech in Noise research papers. Ever since I have frequently searched through the latest published papers on the topic. My reading to date has included several neuroscience journals such as Brain and Journal of Neuroscience as well as the more audiology related journals such JASA and JRRD. Whilst, as always, the more questions you get answers to, the more additional questions seem to appear, I have recently began to feel some of my knowledge gaps had been filled. This presentation is an attempt to capture what I think I now know about Speech in Noise and I frequently use its content in the conversations I have with patients about their hearing difficulties. I hope you find it similarly useful. Declaration of interest The author reports no conflicts of interest. The authors alone are responsible for the content and writing of the presentation. 39 Poster Abstracts: General sensation and perception 1. Test-retest reliability of the Test of Everyday Attention (TAIL) B.S. Stirzaker, H.J. Stewart and S. Amitay, MRC Institute of Hearing Research, Nottingham, UK The Test of Attention in Listening (TAIL) is a computerized behavioural test which quantifies the attentional processes that mediate auditory processing performance (Zhang et al, 2012). Through making decisions regarding the frequency or location of two non-verbal tones TAIL extracts five independent outcome measurements: information processing efficiency, involuntary orientation of attention to both frequency and location, and conflict resolution for frequency and location. The TAIL has been successfully validated as a test of auditory attention; Stewart and Amitay (2013) found that the TAIL’s measures successfully capture a range of attentional types, and that TAIL’s outcome measures significantly correlate with existing tests of attention which use visual and/or auditory stimuli. While Zhang et al (2012) showed that the attentional effects measured through TAIL are resistant to rapid learning, its test-retest reliability is yet to be formally assessed. The current study examined the test-retest reliability of the TAIL. On the basis of previous testretest procedures of attention measures (e.g., Chan, Hoosain & Lee, 2002), the TAIL was administered at time 1 (i.e. baseline test), and was delivered again (time 2) in one of three interassessment trial intervals: immediately, after 90 minutes, or between seven and fourteen days following baseline assessment. Normal hearing participants were recruited from a healthy population, and randomly allocated to one of the three inter-assessment trial intervals. TAIL scores were compared between time 1 and 2 for each of the three groups, the scores were found to be consistent across all time frames. These preliminary findings confirm our initial prediction, that the five outcome measures derived from TAIL have test-retest reliability across all proposed time intervals. These results further support the use of the TAIL as an evidence-based clinical and research tool for assessing the effect of auditory attention on auditory processing. Acknowledgements Supported by the Medical Research Council, UK. References Chan R.C., Hoosain R. & Lee T.M. 2002. Reliability and validity of the Cantonese version of the test of everyday attention among normal Hong Kong Chinese: a preliminary report. Clin Rehabil, 16, 900-909. Stewart H.J. & Amitay S. 2013. How the Test of Attention in Listening (TAIL) relates to other tests of attention. Presented at British Society of Audiology Annual Meeting, Keele, United Kingdom, Abstract #84. Zhang Y.-X., Barry J.G., Moore D.R. & Amitay S. 2012. A new test of attention in listening (TAIL) predicts auditory performance. PLoS One, 7, e53502. 2. Effects of attention and intention on auditory streaming A.J. Billig, M.H. Davis and R.P. Carlyon, MRC Cognition & Brain Sciences Unit, UK We investigate the extent to which attention and intention affect the perception of repeated ABApatterns, in which A and B represent pure tones of two different frequencies, and “-” indicates a 40 silent gap of the same duration. These sequences can either be perceived in an “integrated” form (with a characteristic galloping rhythm) or as two “segregated” streams, each containing tones of a single frequency. The integrated pattern is typically experienced initially, with phases of segregated and integrated perception then alternating in a bistable fashion. In the first experiment, the ABA- sequence was presented in one ear. In the “Attend” condition, listeners indicated their percept throughout and/or detected occasionally delayed B tones. In the “Switch” condition, listeners performed a demanding task in the contralateral ear before switching their attention to the tones for the final few seconds of each sequence. Both measures indicated that segregated perception was significantly less likely toward the end of the “Switch” condition, compared to the same point in the “Attend” condition. This may indicate that streaming is reduced in the absence of attention, or that the act of switching attention increases the probability of resetting perception to integration. We partially replicate previous findings and show that subjective and objective measures of streaming can be collected simultaneously. Performance on the objective task was better when the pattern was reported as integrated than as segregated. In a second experiment, we will objectively investigate whether the perception of ABApatterns can be intentionally influenced by attending either to the triplet as a whole (to promote integration), or selectively to tones of one frequency (to promote segregation). Deviants of two types will be embedded in the sequences, the first of which will be easier to detect during integration, and the other during segregation. Listeners will be asked to detect both types simultaneously, while either allowing their perception to wander freely, or attending to the tones in such a way as to promote integration or segregation. In a third experiment, listeners will continuously report their percept while we manipulate intention as in the previous experiment and record electro- and magnetoencephelograpic activity. We will examine whether electrophysiological markers of streaming are elicited in a manner that supports intention affecting not only subjective reports, but perception per se. Initial data from these second and third experiments will be presented and discussed. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 3. Listening effort and fatigue: Insights from pupillometry R. McGarrigle, K. J. Munro, P. Dawes, and A. Stewart, School of Psychological Sciences, University of Manchester, UK. Individuals with hearing loss commonly report feelings of fatigue and stress as a result of the mental demands of listening in everyday situations. Previous attempts have been made to characterise this phenomena by measuring listening effort objectively. One such objective measure of listening effort is pupillometry. Pupil size is sensitive to changes in attention and memory load, and has been shown to reflect the mental effort required to understand speech in challenging listening environments (Zekveld et al, 2010). Pupil size has also been shown to reflect subjective fatigue (e.g. from sleep deprivation) (Morad et al, 2000). However, it remains unclear whether or not pupil size can characterise listening effort and/or listening-related fatigue in more naturalistic listening tasks. We developed a novel speech-picture verification (SPV) task which involves speech passages being presented with multi-talker babble noise in two different listening conditions (Easy/Hard). Pupil size was recorded using an eyetracker, and the pupil response was recorded over the course of speech processing. Four 3-second time-bins were created to analyse the ‘effort’ and ‘fatigue’ responses separately. The main findings were that pupil sizes were significantly smaller in the ‘Hard’ versus the ‘Easy’ condition in time-bin 4 (the final epoch of each trial). Post-hoc analysis 41 reveals that baseline pupil sizes decreased as the experiment progressed, and the ‘fatigue’ effect was largest in the final epoch of the experiment. No significant difference in pupil size was found between listening conditions in the first time-bin. These findings suggest that pupillometry can capture listening-related fatigue. An objective measure of listening-related fatigue may serve as a supplementary tool used to assess hearing aid benefit in the clinic. Acknowledgements Supported by the Castang Foundation. References Morad, Y., Lemberg, H., Yofe, N. & Dagan, Y. 2000. Pupillography as an objective indicator of fatigue. Curr Eye Res, 21, 535-542. Zekveld, A., Kramer S. & Festen, J. 2010. Pupil response as an indication of effortful listening: The influence of sentence intelligibility. Ear Hear, 31, 480-490. 4. Using adaptation to investigate the neural mechanisms of attention in the human auditory cortex J. de Boer, S. Gibbs and K. Krumbholz, MRC Institute of Hearing Research, Nottingham, UK Single neuron recordings in auditory cortex have suggested that attention causes a sharpening of neural frequency selectivity (Fritz et al., 2003). In humans, neuroimaging studies have reported similar effects when using notched-noise masking to estimate cortical frequency selectivity (Ahveninen et al, 2011). There is a possibility, however, that these results were confounded by differences in attentional load between different masking conditions. Here, we tested the effect of selective attention on cortical frequency tuning directly, using an adaptation paradigm previously used in the visual system (Murray & Wojciulik, 2004). In this paradigm, the feature selectivity of cortical neurons is assessed by measuring the degree of stimulus-specific adaptation of the gross evoked response as a function of the difference between the adapting stimulus and the subsequently presented probe stimulus. If attention causes an increase in neural selectivity, it would be expected that adaptation becomes more stimulus specific, that is, more strongly dependent on the adapter-probe difference. Auditory-evoked potentials (AEPs) were recorded from 18 participants performing a dichotic listening task. Tone sequences comprising four equally probable frequencies were presented to one ear, while, concurrently, a sequence of waxing or waning noises was presented to the other ear. The participants were instructed to attend either the tones or the noises (alternating every 2.5 min) and to detect rare oddballs within the attended stimulus sequence. Only AEPs to the tones were recorded. As expected, AEP amplitude was significantly larger when the tones were attended than when they were ignored. Furthermore, the effect of attention on AEP amplitude was significantly greater when the evoking tone was immediately preceded by a tone of a different frequency than when it was preceded by a tone of the same frequency. This suggests that the adaptation caused by preceding tones was more frequency-specific when the tones were attended than when they were unattended, implying that selective attention caused an increase in cortical frequency selectivity. Acknowledgements Supported by the MRC. References Fritz J., Shamma S., Elhilali M. & Klein D. 2003. Nat. Neurosci, 6 (11), 1216-1223. 42 Ahveninen J., Hamalainen M., Jaaskelainen I.P., Ahlfors S.P., Huang S. & Lin F.H. 2011. Proc Natl Acad Sci, USA, 108, 4182-4187. Murray S.O. and Wojciulik E. 2004. Nat Neurosci,7, 70-74. 5. Differential effects of training regimen on the learning of speech and non-speech discrimination S. Amitay* and K. Banai§, *MRC Institute of Hearing Research, Nottingham, UK, §Department of Communication Disorders, University of Haifa, Israel While perceptual learning of speech was shown to generalise more widely when trained using a variable regimen (e.g., Lively et al, 1993), this was not the case for the perceptual learning of acoustic feature discrimination (e.g., Banai et al, 2010). Nevertheless, the two types of stimuli were rarely assessed using comparable training schedules. Listeners concurrently trained on discrimination of either four synthetic syllable contrasts or four tone durations. During training, stimuli were presented either in a blocked (160 trials each stimulus, ran consecutively) or a roving (all four stimuli randomly interleaved within each 160-trial block) training schedule. All listeners trained for a total of 640 trials per session for 9 sessions, interspersed over 1-2 months. All completed a pre- and post-test that included testing on all stimuli, presented either in blocks or roved for separate groups. Regardless of training schedule, speechtrained listeners showed significant improvement on speech stimuli compared to duration-trained listeners, who showed no learning effects whatsoever. Learning on the speech stimuli did not generalise to the acoustic stimuli. Although roving impeded naïve discrimination performance, training with either blocked or roved speech stimuli resulted in similar post-training thresholds for both blocked and roved speech stimuli. It is unclear why no learning was observed for the blocked duration training. We conclude that roving and blocked training regimens were equally efficient (or inefficient) for perceptual learning. Whereas speech learning appears robust, duration-discrimination learning seems highly sensitive to the particulars of the training schedule. Acknowledgements Supported by The National Institute of Psychobiology in Israel and the Medical Research Council, UK. References Banai, K., Ortiz, J. A., Oppenheimer, J. D., and Wright, B. A. 2010. Learning two things at once: differential constraints on the acquisition and consolidation of perceptual learning. Neuroscience, 165(2), 436-444. Lively, S. E., Logan, J. S., Pisoni, D. B., 1993. Training Japanese listeners to identify English /r/ and /l/. II: The role of phonetic environment and talker variability in learning new perceptual categories. J Acoust Soc Am, 94(3 Pt 1), 1242-1255. 6. A novel method for predicting speech intelligibility based counting time-frequency bins S. Man Kim, N. Moslemi Yaldaei and S. Bleeck, ISVR, University of Southampton, Southampton, UK Finding an objective intelligibility measure that predicts subjective speech intelligibility is an important on-going research area because existing measures are not good in all situations, and 43 subjective listening tests are costly and time consuming (Loizou, 2013). Several objective measures have been proposed which rely, amongst others, on objective speech quality measures such as signal-to-noise ratio (SNR), signal-to-distortion ratio (SDR) or correlation coefficients (Loizou, 2013). However, none of these methods are perfect, because they do not provide cues for example on how intelligibility of noisy speech depends on the type of noise and why they do not correctly predict intelligibility changes for de-noising algorithms. We propose here an objective measure for predicting the intelligibility of noisy speech that is based on counting the number of perceived time-frequency (T-F) speech component bins. The intelligibility is determined by frame wise counting the number of T-F bins above and below threshold. This strategy is motivated by the assumption that human beings have suprathreshold speech intelligibility enhancing mechanisms, so the number of actually perceived speech components is a more important factor for speech intelligibility than the SNR or SDR. For evaluation, the speech intelligibility predictions of the proposed method are compared against speech intelligibility scores from subjective listening trials against the results of state-of-theart conventional speech intelligibility predicting algorithms such as STOI (Taal et al, 2011) and CSII (James & Kathryn, 2005). Preliminary results show that the proposed method is superior in various noise and noise reduction situations. Further results from ongoing objective measures and subjective listening tests will be shown on the poster. Acknowledgements The work leading to this deliverable and the results described therein has received funding from the People Programme (Marie Curie Actions) of the European Union’s Seventh Framework Programme FP7/2007-2013/ under REA grant agreement n° PITN-GA-2012-317521. References Loizou P. C. 2013. Speech Enhancement: Theory and Practice, Second Edition. CRC Press. Taal C. H., Hendriks R. C., Heusdens R. & Jensen J. 2011. An algorithm for intelligibility prediction of time–frequency weighted noisy speech. IEEE Transactions on Audio, Speech, and Language Processing, 19(7), 2125-2136. James M. K. & Kathryn H. A. 2005. Coherence and the speech intelligibility index. J Acoust Soc Am, 117(4), 2224-2237. 7. Informational masking of monaural speech by a single contralateral formant B. Roberts* and R.J. Summers*, *Psychology, School of Life and Health Sciences, Aston University, Birmingham, UK The role of the time-varying properties of speech in separating a target voice from an interfering voice remains unclear. Recent evidence suggests that when masking is primarily informational, the impact of a single extraneous formant on intelligibility becomes greater as the depth of variation in its frequency contour increases (Roberts et al, 2014). However, three design features constrained the range of formant-frequency variation tested and restricted the generality of the conclusions. First, the three target formants constituting the speech analogues were split between the two ears (left=F1; right=F2+F3). Second, listeners were able to hear each stimulus up to six times before transcribing it. Third, the extraneous formant was added in the same ear as the target F1, which restricted the range of formant-frequency variation that could be tested without crossovers or close approaches between adjacent formants. This study evaluated further the impact of formant-frequency variation in an interferer on speech intelligibility using a method adapted to overcome these constraints. Three-formant analogues of natural sentences were synthesized using second-order resonators and a monotonous glottal source (F0=140 Hz). Target formants were always presented monaurally; 44 the ear receiving them was assigned randomly on each trial. A competitor for F2 (F2C) was added in the contralateral ear (target ear = F1+F2+F3; other ear = F2C), which listeners must reject to optimize recognition. Listeners heard each stimulus only once before entering their transcription. Different competitors were created by inverting the frequency contour of F2 about its geometric mean and varying its depth over a range extending from constant (0%) to twice that of the natural utterances (200%). F2C always had the same F0, level, and amplitude contour as the target F2. Despite the differences from earlier studies, in which the target formants were split between ears and repeat listening was permitted, the adapted method was effective at assessing the impact of F2C on intelligibility. The results confirm and extend those of Roberts et al (2014). Adding F2C typically reduced intelligibility, presumably as a result of informational masking; the impact was least for 0%-depth (constant) F2Cs and increased with depth up to ~100% scaling, but not thereafter. This outcome is consistent with the idea that the ability to reject an extraneous formant does not depend on target-masker similarity in depth of formant-frequency variation. Acknowledgements Thanks to Peter Bailey. Work supported by ESRC. References Roberts B., Summers R.J. & Bailey P.J. 2014. Formant-frequency variation and informational masking of speech by extraneous formants: Evidence against dynamic and speech-specific acoustical constraints. J Exp Psychol Hum Percept Perform. Online First Publication, May 19, 2014. 8. Assessing the interplay of linguistic factors and background noise characteristics to successful speech in noise listening A. Heinrich*, S. Knight*, M. Young*, R. Moorhouse§ and J. G. Barry* *MRC Institute of Hearing Research, University of Nottingham, Nottingham, UK; § University of Nottingham, Nottingham, UK Contextual information supports speech-in-noise listening (Bilger et al. 1984) with the relative importance of this information increasing for older adults (Pichora-Fuller et al, 1995). Studies investigating contextual information are usually based on the SPIN-R sentence pair test (SPIN-R; Bilger et al. 1984), where the same sentence-final word can either be predicted or not from the preceding context. However, these sentence pairs do not differ purely in the relative usefulness of the sentential content: they also vary in duration, stress pattern and semantic richness (highpredictability sentences are semantically varied, while their low-predictability counterparts rely on repeated stock phrases). Additionally, the 12-talker babble traditionally used as masking noise has a steep frequency roll-off above 1.5kHz, so is rumbly and muffled. Stimuli must therefore be presented at very high, ecologically improbable SNRs for suitable levels of masking to occur. In this study, we investigated the role of sentence structure and noise effects in determining overall performance on SPIN-type tasks. The study compared the performance of 12 young normally-hearing listeners on original SPINR sentences and recently developed SPIN-UK sentences. The SPIN-UK sentence pairs are matched for duration, stress pattern, and semantic complexity. All sentences were recorded using a male Standard British English speaker. They were presented either in the original 12-talker babble or new 12-talker babbler with a natural 6dB/oct roll-off at SNRs of -2dB (SPIN-UK) and -15dB (SPIN-R). 45 These levels were chosen to produce comparable intelligibility levels for high predictability sentences. There were main effects of sentence set and predictability, and two-way interactions of babble type x predictability and sentence set x predictability. The main effect of sentence set showed that accuracy was higher for SPIN-UK than SPIN-R sentences. The main effect of predictability showed that accuracy was higher for high- than low-predictability sentences. The babble type x predictability interaction revealed that traditional SPIN babble produced higher accuracy than the new babble, but only within low predictability sentences. The sentence set x predictability interaction revealed that the SPIN-UK sentences produced higher accuracy than the SPIN-R sentences, but only within low predictability sentences. These results suggest that in addition to semantic predictability, broader linguistic factors like prosody and semantic richness contribute to speech-in-noise listening. Design of the masking noise also makes an important contribution to performance and can influence conclusions about general speech-in-noise listening. In future studies we plan to reassess the listening abilities of older adults relative to younger adults using the more ecologically valid SPIN-UK sentences. Acknowledgments Supported by the MRC., Action on Hearing Loss, and Biotechnology and Biological Sciences Research Council (BBSRC). References Bilger, R. C., Nuetzel, J. M., Rabinowitz, W. M. & Rzeczkowski, C. 1984. Standardization of a test of speech perception in noise, J Speech Hear Res, 27(1), 32-48. Pichora‐Fuller, M. K., Schneider, B. A. & Daneman, M. 1995. How young and old adults listen to and remember speech in noise, J Acoust Soc Am, 97(1), 593-608. 9. Developing a speech-in-noise test to measure auditory fitness for duty for military personnel: introducing using the Coordinate Response Measure H.D. Semeraro* D. Rowan*, R.M. van Besouw* and A. Allsopp§, *Institute of Sound and Vibration Research, Southampton, UK, § Institute of Naval Medicine, Gosport, UK. The ability to listen to commands in noisy environments and understand acoustic signals, whilst maintaining situational awareness, is an important skill for military personnel, and can be critical for mission success. Pure-tone audiometry (PTA) is currently used in the UK to assess whether personnel have sufficient hearing for duty, though it is known that PTA is poor at predicting speech intelligibility in noise (SIN). The aim of this project is to develop and validate a SIN test suitable for assessing the auditory fitness for duty of military personnel. Following a review of existing SIN tests, the Coordinate Response Measure (CRM) was selected, partly due to the high face validity when compared to command structure (CRM sentence format: “Ready [call-sign], go to [colour] [number] now”). The CRM was re-recorded using NATO call-signs and a larger selection of call-signs (18) and colours (9) than previous versions. Psychometric functions were estimated with 20 normal-hearing listeners for each call-sign, colour and number in speech-spectrum noise; correction factors were applied to equalise the intelligibility of the stimuli. Analysis of confusion matrices for the call-signs, colours and numbers showed that no two words were being consistently confused. Next, the CRM was implemented in a 2-down, 1-up adaptive procedure, again, using speechspectrum noise. Speech recognition thresholds were estimated four times using two scoring methods, first responding to the call-sign, colour and number and second only responding to the colour and number, and using the Triple Digit Test (TDT) for comparison, with 30 normal-hearing 46 participants. Any learning effect across the repeats was less than 0.7 dB. The test-retest reliability, expressed as the 95% confidence interval on any one SRT estimate on the CRM, was 2.1 dB when responding to all three parts of the sentence and 1.9 dB when only responding to two parts, compared to 2.6 dB on the TDT. A similar study of 30 hearing impaired military personal (bilateral hearing losses ranging from mild to severe) is on-going; assuming a satisfactory outcome from that study future work will focus on assessing the external validity of the CRM using simulated combat scenarios. Acknowledgement This work is funded by the Surgeon General of the Ministry of Defence. 10. Speech understanding in realistic conditions: effects of number and type of interferers, and of head orientation J. F. Culling and S. Priddy, School of Psychology, Cardiff University, Park Place, Cardiff, U.K. In real life, interfering noise is frequently uninterrupted, multi-source spatially distributed and reverberant. Three experiments measured speech reception thresholds (SRTs) against uninterupted interferers based on continuous discourse in different virtual environments. Experiments 1 and 2 compared the effects of speech, reversed speech, speech-shaped noise and speech-modulated speech-shaped noise as a function of the number of interfering sources at a constant overall noise level (controlled at source). Experiment 1 manipulated reverberation using a simulated room, while Experiment 2 used recordings from an acoustic manikin in a real, but unoccupied, restaurant. Experiment 3 examined the effect of the manikin’s head orientation in the presence of eight interferers (speech or speech-shaped noise) in the same restaurant. Experiment 1 found that SRTs were elevated in reverberation and, except in the case of reverberated speech-shaped noise, for larger numbers of interferers. The effect of reverberation decreased with increasing numbers of interferers. Reversed speech produced the lowest and continuous noise the highest SRTs in most cases. The results suggest beneficial effects of masker periodicity and modulation for 1 or 2 interfering sources (SRTs lower for modulated than continuous sources, and lower for single-voice speech and reversed speech sources than for modulated noise). There were detrimental effects of masker intelligibility (lower SRTs with reversed than with forward speech interferers) with 2 interferers, and there was evidence of modulation masking with 8 interferers (lowest SRTs for continuous noise). Experiment 2 found a similar pattern of results to the reverberant case in Experiment 1. Experiment 3 found consistently beneficial effects of head orientation away from the target source. This benefit was driven by improved target-speech level at the ear turned towards the source. As with 8 interferers in Experiments 1 and 2, speech interferers produced 2 dB higher SRTs than noises. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 11. The effect of changes in the acoustic environment on speech intelligibility S. Aspeslagh*§, F. Clark§, M.A. Akeroyd* and W.O. Brimijoin*, *MRC/CSO Institute of Hearing Research - Scottish Section, Glasgow Royal Infirmary, Glasgow, UK, §Department of Computing, University of the West of Scotland, Paisley, UK 47 As listeners walk around a room or move from one room into another they experience changes in the characteristics of the reverberation around them. It is reasonable to hypothesize that listeners are constantly engaged in analysing the acoustics of the space around them in order to optimize speech intelligibility. Previous studies have shown that prior exposure to reverberation is associated with rapid improvements in speech intelligibility (Brandewie & Zahorik, 2010). However, reverberation is not the only acoustic element that can change over time: static background noise changes in level at the ear as listeners move closer or further from the source, and can completely change as background noises themselves start or stop. Previous work from our laboratory has shown that listeners’ ability to identify simple target words briefly drops as they are introduced to a novel simulated acoustic environment. It takes on average around three seconds to return to optimum performance and this time-course is the same regardless of level of hearing impairment, age, or type of background noise or amount of reverberation. Considering that in the previous experiment each environment consisted of a unique combination of noise and reverberation that was entirely novel to the listener, we wished to measure any drop in intelligibility when more subtle changes were made to the acoustics. Therefore, in the current experiment only a single element or a subset of elements were changed rather than everything about a given environment. The three elements included were background noise, reverberation, and target/noise angles. Listeners were asked to identify an ongoing stream of simple words, and the noise, reverberation, angle, or a combination of these three were changed every nine seconds. The results suggest that when all three of the above elements of an acoustic environment change, speech intelligibility is more affected than when only one or two elements change. These results agree with previous work in that listeners’ experience with novel acoustic environments increases speech intelligibility, but also demonstrate that the extent to which speech intelligibility drops after a change is dependent on how substantially the acoustics are altered. As before, we observed no significant difference in adaptation time between normal and hearing impaired listeners, suggesting that the adaptation to changes in the acoustic environment is a process independent of hearing impairment. Acknowledgments: The first author is supported by a MRC PhD Studentship. Supported by MRC (U135097131), the Chief Scientist Office (Edinburgh), and the University of the West of Scotland. References Brandewie, E., & Zahorik, P. 2010. Prior listening in rooms improves speech intelligibility. J Acoust Soc Am, 128(1), 291-299. 12. Native and non-native sentence comprehension in the presence of a competing talker H. Valdes-Laribi*, D. Wendt§, E. MacDonald§, M. Cooke‡ and S. Mattys*, *Department of Psychology, The University of York, York, UK, §Department of Electrical Engineering, Technical University of Denmark, Lyngby, Denmark, ‡Language and Speech Laboratory, University of the Basque Country, Vitoria, Spain In everyday environments, we often have to attend to one person’s speech (target speech) while ignoring another (competing speech). Competing speech can mask the target through energetic masking (acoustic degradation at the periphery) and informational masking (higher-order, cognitive aspects of masking). In this study, we tested the hypothesis that informational masking depletes processing resources that would otherwise be allocated to recognising and understanding the target signal. 48 Using a speeded picture selection task, we investigated native English listeners’ understanding of sentences varying in syntactic complexity, namely subject relatives (easy) and object relatives (hard), played against a competing talker (energetic and informational masking) vs. speechmodulated noise (energetic masking only). Object relative sentences have been shown to require more processing resources than subject relative sentences. We hypothesized that informational masking depletes processing resources that would otherwise be allocated to recognizing and understanding the target speech. The detrimental effect of the competing talker relative to speechmodulated noise should therefore be greater in the object relative than the subject relative condition. Although participants were slower at responding to object relative sentences, there was no effect of mask type or interaction with sentence type. Thus, within the constraints of this experiment, there was no evidence that a competing talker requires greater processing resources than energetic masking alone. We then hypothesised that processing our sentences as a non-native listener might exacerbate the interaction between mask and syntactic complexity because of the greater cognitive effort involved in processing an L2. However, when native Danish speakers were tested on our stimuli, their results were broadly equivalent to those of the native speakers, only slower. Thus, contrary to research by other groups, we found no evidence that a competing talker requires greater processing resources than energetic masking alone. An ongoing eye-tracking version of this experiment will establish the nature of the discrepancy and the psycholinguistic mechanisms behind informational masking. Acknowledgements This project has received funding from the European Union’s Seventh Framework Programme for research, technological development and demonstration under grant agreement no FP7-PEOPLE2011-290000. 13. Specificity effects within and beyond spoken word processing D. Strori*, O. Scharenborg§ and S. Mattys*, *Department of Psychology, University of York, UK, § Centre for Language Studies, Radboud University Nijmegen, The Netherlands Previous research indicates that lexical representations might include both linguistic and indexical (talker-related) specifications (Goldinger, 1998). Recent evidence suggests that non-linguistic sounds co-occurring with spoken words are also incorporated in our lexical memory. Pufahl and Samuel (2014) paired spoken words with environmental sounds and observed that perceptual identification accuracy for previously heard words dropped when the paired sound changed from exposure to test. The authors attributed this effect to the word-sound association and claimed that the specificity (indexical) effects extend beyond the speech domain. We argue that this "soundspecificity effect" might not be due so much to a word-sound association as to the different acoustic glimpses of the words that the associated sounds create. Our first experiment replicated the “voice specificity effect”, such that recognition accuracy for previously heard words (spoken by one of two talkers) dropped when the voice changed from exposure to test. The next experiments investigated the “sound specificity effect”, from an energetic masking angle. In an analogy to the two voices, we paired the spoken words (spoken by the same talker) with one of two car honk sounds and varied the level of energetic masking from exposure to test. The honk sounds had the same on/off pattern and differed only in the frequency domain. During the exposure phase participants (university students) heard word-honk pairs played binaurally and completed a semantic judgement task on the word, to ensure its lexical encoding in memory. After a short delay, participants completed the test phase, in which half of the previously played words were paired with the other honk sound. Their task was to decide whether they had heard the word in the first phase or not. 49 We did not observe a drop in recognition accuracy for previously heard words when the paired sound changed as long as energetic masking was controlled. This was achieved by preserving the temporal overlap between words and honks in both phases. However, when we manipulated the temporal overlap to create an energetic masking contrast, accuracy dropped. The finding suggests that calling for an expansion of the mental lexicon to include non-speech auditory information might be premature. It also indicates that specificity effects do not easily extend beyond the speech domain. Acknowledgements This project has received funding from the European Union’s Seventh Framework Programme for research, technological development and demonstration under grant agreement no FP7-PEOPLE2011-290000. References Goldinger, S. D. 1998. Echoes of echoes: An episodic theory of lexical access. Psych Rev,105, 251– 279. Pufahl, A. & Samuel, A.G. 2014. How lexical is the lexicon? Evidence for integrated auditory memory representations. Cog Psych, 70, 1-30. 14. Artificial grammar learning in young and older listeners C. Füllgrabe*, E. Clifford* and B. Gygi§‡,*MRC Institute of Hearing Research, Nottingham, UK, § Nottingham Hearing Biomedical Research Unit, Nottingham, UK, ‡US Department of Veterans Affairs, Martinez, California, USA Speech identification benefits from the fact that words are arranged in formal grammars which provide context and enhance expectations for certain words. Similarly, non-speech environmental sounds tend to occur in certain settings and sequences, and listeners are aware of these regularities (Gygi & Shafiro, 2011). It is well established that speech intelligibility, especially in the presence of background noise, declines with age. However, age-related peripheral hearing loss alone often fails to explain the entire deficit. Hence, the aim of this study was to investigate whether or not (i) the ability to learn grammars for short (5-item) sequences declines with age, and (ii) the knowledge acquired through grammar learning enhances identification performance in noise. Participants were 17 young (age range = 19-30 yrs; mean = 22 yrs) and 17 older (age range = 62-76 yrs; mean = 68 yrs) listeners with bilaterally normal hearing (defined as audiometric threshold ≤ 20 dB HL between 0.125 and 4 kHz). All older listeners were screened for neurological dysfunction using the Mini Mental State Examination; all scored 27/30 or higher. The Digits Forward part of the Digit Span test (Wechsler Adult Intelligence Scale, 1999) was administered to all participants to ascertain that all could maintain in short-term memory sequences of at least five items. Target (i.e., grammatical) sequences, containing five consecutive sounds (A-E), were generated according to an artificial grammar based on a finite-state machine. Non-target sequences were composed of the same A-E sounds but violated the grammatical grammar rules. The sounds were five different environmental chimerae, obtained by combining the spectrum of one environmental sound (e.g. whistle) with the temporal envelope of another (e.g. footsteps). Participants were initially trained to recognize the five chimerae in isolation. Then, after listening passively to target sequences, participants were asked to recognize these sequences amongst non-target sequences. To assess if participants had indeed learned the underlying artificial grammar, the ability to classify new grammatical and a-grammatical sequences as target or non50 target sequences, respectively, was tested. Finally, participants had to identify the 4th sound in target and non-target sequences in the presences of masking noise at different signal-to-noise ratios. Three additional cognitive tests were conducted to explain inter-individual and -group differences: the Digits Backward from the Digit Span test and the Letter-Number-Sequencing test which are assumed to probe working memory capacity, and the Verbal-Paired-Associates test to assess explicit episodic memory performance. Results show that both young and older learned the artificial grammar and transferred their knowledge to new sequences. There was also a comparable benefit of grammaticality on sound-innoise identification in the two age groups. We conclude that age does not affect the use of top-down information (such as grammaticality) to aid listening in noisy environments. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper.' References Gygi, B. & Shafiro, V. 2011. The incongruency advantage for environmental sounds presented in natural auditory scenes. J Exp Psychol Human, 37, 551-565. 15. Sensitivity of SPiN-UK sentences to differences in children’s listening abilities J.G. Barry*, S. Knight*, A. Heinrich*, R. Moorhouse§ and H. Roebuck*, *MRC Institute of Hearing Research, Nottingham, UK, § University of Nottingham, Nottingham, UK The SPIN sentence-pair test (Bilger et al, 1984) was designed to separately assess contributions to successful listening in noise from auditory-phonetic speech discrimination abilities and cognitive and linguistic abilities. Sentence pairs share the same final single syllable word, but differ in the ease with which this word can be predicted from the preceding context. We recently developed the SPiN-UK sentence pair test (Heinrich et al, 2014) and here, we consider the suitability of these sentence pairs for assessing individual differences in speech-in-noise listening in children aged 8 to 10 years. The children were subdivided using the ECLiPS (Barry & Moore, 2014) according to whether they were typically-developing (TD-Child) or had listening difficulties (AP-Child). SPiN sentences (predictable vs unpredictable) were presented in a British English four-talker babble (Heinrich) at a high (+2dB) and low (-2dB) SNR. Listeners (young adults, n = 18; TD-Child, n = 14; AP-Child, n = 13) repeated each sentence as accurately as possible and the number of correct sentence final words for each group was compared. No floor effects were apparent for any group on the task suggesting the sentences and SNRs used were suitable for the age range and listening abilities of the participants. There was a significant main effect of group (p<.001). The Child-AP group made more errors on all sentence types than the Child-TD group, who in turn made more errors than the adults particularly in the unfavourable SNR. There was an interaction of Group x Sentence type x SNR (p<.001). Both groups of children made more errors in the low SNR, they also got less benefit from the sentence context than the adults in this condition. The Child-AP group performed equally poorly regardless of relative predictability of the final word. This suggests listening support from language abilities (high predictability sentences) cannot offset their greater speech perception difficulties (low predictability sentences). Overall, the UK-IHR SPiN sentences can be used successfully with different child populations and can provide interesting insights into age and disorder-related difficulties. 51 Acknowledgements Supported by the MRC, Action on Hearing Loss, Biotechnology and Biological Sciences Research Council (BBSRC). References Barry, J.G. & Moore, D.R. (2014). Evaluation of Children’s Listening and Processing (ECLiPS) Manual, Edition 1. Heinrich, A., Knight, S., Young, M., & Barry, J. (2014). Assessing the interplay of linguistic factors and background noise characteristics in successful speech in noise listening. BSA Abstract. Bilger, Nuetzel, J., Rabinowitz, W., & Rzeczkowski, C. (1984). Standardization of a Test of Speech-Perception in Noise. J Speech Hear Res, 27(1), 32-48. 16. Generalization of sensory and cognitive learning in typically developing children C.F.B. Murphy*, D.R. Moore§ and E. Schochat*, *Department of Physical Therapy, SpeechLanguage Pathology and Occupational Therapy, University of São Paulo, São Paulo, Brazil, §Communication Sciences Research Center, Cincinnati Children’s Hospital Medical Center, Cincinnati, Ohio, USA. Both bottom-up (sensory) and top-down (cognitive) mechanisms are involved in learning following auditory training, as demonstrated by the far-transfer of auditory sensory learning to higher level language skills (Moore et al, 2005) and top-down neural modulation of auditory sensory learning (Bajo et al, 2010). However, direct comparison between sensory and cognitive training on untrained auditory sensory and cognitive skills in children has not previously been attempted. This study aimed to assess the extent to which generalization of learning occurred following each type of training in typically developing children. Sixty 5-to 8-year-old normal hearing children were quasi-randomly assigned to 5 training groups: attention (AG), memory (MG), auditory sensory (SG), placebo (PG; drawing, painting) and control, untrained (CG). Compliance measures (last level achieved during the game and the number of game blocks played), as well as mid- transfer (same domain; auditory digit span, sustained auditory attention, time compressed speech) and far-transfer (different domain; phonological processing, word reading) measures were applied before and after training. All trained groups received eleven 50-min training sessions; the CG did not receive any intervention. All trained groups showed near-transfer; significant learning (higher levels) on the trained task. No significant correlation was found between the numbers of game blocks played and the last level achieved. On pre- to post-training measures, most groups showed (test-retest) improvements on most tasks. There was a group x training interaction for the digit span task (F (4,117) = 2,73, p = 0.038), with higher spans after MG training. The results show that both bottom-up and top-down training can lead to learning on the trained task, as well as mid-transfer learning on a task (auditory digit span) in the same domain as those trained. Acknowledgements Supported by the CAPES/Coordenação de Aperfeiçoamento de Pessoal de Nível Superior. References Bajo V.M., Nodal F.R., Moore D.R. & King A.J. 2010. The descending corticocollicular pathway mediates learning-induced auditory plasticity. Nat Neurosci, 13(2):253-60. Moore D.R., Rosenberg J.F. & Coleman J.S. 2005. Discrimination training of phonemic contrasts enhances phonological processing in mainstream school children. Brain Lang; 94(1):72-85. 52 17. Imaging brain structures in children with auditory neuropathy spectrum disorder H. Cooper*, L. Halliday§, D. Bamiou‡ and C.Clark*, * Institute of Child Health, University College London, UK, § Division of Psychology and Language Sciences, University College London, UK, ‡ Ear Institute, University College London, UK It is believed that the brain requires both normal intrinsic and extrinsic factors in order for typical development to take place. Auditory neuropathy spectrum disorder (ANSD) often results in fluctuating hearing loss and auditory temporal processing abnormalities leading to degraded and inconsistent auditory input (Rance et al, 1999). ANSD is a collection of test results rather than a specific disorder in its own right. The pattern of test results required for a diagnosis of ANSD is present otoacoustic emissions and/or cochlear microphonic with absent or severely abnormal auditory brainstem response (Starr et al, 1996). The abnormal subcortical transmission of sound experienced by children with ANSD therefore may result in the disruption of normal cortical development. Very little is known about cortical neuromaturation and neural plasticity in children with ANSD. Children from around the UK who are age six and over and have a diagnosis of ANSD are currently being invited to take part in this study. A detailed test battery has been developed in order to comprehensively describe and analyze ANSD. The study uses a variety of magnetic resonance imaging (MRI) techniques to examine both structural and functional brain connectivity in children with ANSD. Diffusion tensor imaging is a quantitative MRI technique that is capable of measuring the microstructure of the brain tissue. Tract based spatial statistics (Smith et al, 2006) is being used to build a map of structural differences in brain white matter networks in children with ANSD compared to normally hearing controls. Resting state fMRI is used to examine differences in functional connectivity between children with ANSD and controls. Cortical thickness is being measured using conventional T1-weighted MRI. Up to 40% of cases of ANSD are thought to have a genetic origin (Manchaiah et al, 2001). Genetic markers for non-syndromic ANSD are being evaluated, including mutations in the otoferlin gene and pejvakin. Tests of auditory performance are used to fully characterize ANSD, including conventional behavioural assessment along with psychoacoustics testing. The psychoacoustics test battery includes assessment of binaural masking level difference, frequency discrimination, temporal integration and temporal modulation transfer function in order to quantify the processing difficulties experienced by children with ANSD. Correlations between tests of auditory function, and differences in structural and functional brain connectivity between children with ANSD and normally hearing controls are being assessed. Preliminary results show wide variations, particularly in behavioural testing, including speech testing and psychophysics assessment. Acknowledgements Hannah Cooper is funded by a National Institute for Health Research CSO Healthcare Science Doctoral Research Fellowship. References Manchaiah V. K. C., Zhao F., Danesh A. A. & Duprey R. 2011. The genetic basis of auditory neuropathy spectrum disorder (ANSD). Int J Pediatr Otorhinolaryngol, 75(2), 151-158. Rance G., Beer D. E., Cone-Wesson B., Shepherd R. K., Dowell R. C., King A. M., Rickards F. W. & Clark G. M. 1999. Clinical findings for a group of infants and young children with auditory neuropathy. Ear Hear, 20(3), 238-252. 53 Smith S. M., Jenkinson M., Johansen-Berg H., Rueckert D., Nichols T. E., Mackay C. E., Watkins K. E., Ciccarelli O., Cader M. Z., Matthews P. M. & Behrens T. E. 2006. Tract-based spatial statistics: voxelwise analysis of multi-subject diffusion data. Neuroimage, 31(4), 1487-1505. Starr A., Picton T. W., Sininger Y., Hood L. J. & Berlin C. I. 1996. Auditory neuropathy. Brain, 119 (Pt 3), 741-753. 18. The development of audiovisual integration in children listening to noise-vocoded speech D.W. Maidment, H-J. Kang, H.J. Stewart and S. Amitay, MRC Institute of Hearing Research, Nottingham, UK While visual information improves speech perception in adults when the fidelity of the auditory speech signal is reduced by background noise, typically developing children benefit comparatively less than adults from observing visual articulations (Ross et al. 2011). Degradation of the speech signal itself, as experienced by hearing impaired listeners, represents another type of adverse listening that can be potentially enhanced by visual cues. Noise-vocoding attempts to simulate speech as heard via a cochlear implant in normal-hearing listeners, the study of which can arguably improve our understanding of speech perception difficulties experienced by this population. Nevertheless, whether additional visual cues improve normal-hearing children’s perception when the speech signal is degraded (e.g. noise-vocoded) is yet to be explored. In the current study, normal-hearing children (n=82) and adults (n=15) were presented noisevocoded sentences from the Children’s Co-ordinate Response Measure (Rosen, 2011) in auditoryonly (AO) or audiovisual (AV) conditions. The number of bands was adaptively varied to modulate the degradation of the auditory signal, with the number of bands required for ~79% correct identification calculated as the threshold. The data revealed two distinct findings: First, we replicated previous work (Eisenberg, et al. 2000) showing that, regardless of visual information, younger children between the ages of 4-5 years required a greater number of frequency bands in order to identify speech in comparison to older children (6-11 years) and adults. Second, we found that visual information improved vocoded speech perception in all but the youngest (4-5 year old) children. Taken together, the data suggests that children do not begin to benefit from accompanying visual speech cues when the auditory signal is degraded until 6 years of age. This evidence not only has implications for understanding the development of speech perception skills in normal-hearing children, but may also inform the development of new treatment and intervention strategies that aim to remediate speech perception difficulties in paediatric cochlear implant users. Acknowledgements Supported by the Medical Research Council. References Eisenberg, L. S., Shannon, R. V., Martinez, A. S., Wygonski, J., & Boothroyd, A. 2000. Speech recognition with reduced spectral cues as a function of age. J Acoust Soc Am, 107, 2704-2710. Rosen, S. 2011. The complexities of understanding speech in background noise. Paper presented at the First International Conference on Cognitive Hearing Science for Communication, Linköping, Sweden. Ross, L. A., Molholm, S., Blanco, D., Gomez Ramirez, M., Saint Amour, D., & Foxe, J. J. 2011. The development of multisensory speech perception continues into the late childhood years. Eur J Neurosci, 33, 2329-2337. 54 19. Exploring the role of synchrony in spatial and temporal auditory-visual integration G.P. Jones, S.M. Town, K.C. Wood, H. Atilgan and J.K. Bizley, UCL Ear Institute, London, UK. Animals and humans integrate sensory information over time, and combine this information across modalities in order to make accurate and reliable decisions in complex and noisy sensory environments. Little is known about the neural basis of this accumulation of information, nor the cortical circuitry that links the combination of information between the senses to perceptual decision making and behaviour. Most previous examples of multisensory enhancement have relied on synchrony dependent mechanisms, but these mechanisms alone are unlikely to explain the entire scope of multisensory integration, particularly between senses such as vision and hearing, which process multidimensional stimuli and operate with vastly different latency constraints. Presented here are two audio-visual behavioural tasks, one spatial, one temporal (adapted from Raposo, et al, 2012), requiring subjects (ferrets and humans) to accumulate evidence from one or both senses over time. In the temporal task subjects estimated the average rate of short auditory and/or visual events embedded in a noisy background (20 ms white noise bursts or flashes) over a defined time period (1000 ms). Instantaneous event rates throughout this time period varied, meaning the accuracy with which the event rate could be estimated increased over time. Similarly, in the spatial task, subjects were required to report whether a greater event rate was presented to the left or right of space. Discrimination was assessed in both unisensory auditory and visual conditions as well as in synchronous and asynchronous multisensory conditions. In the temporal task, accuracy and reaction times for humans were improved in both synchronous and asynchronous multisensory conditions (accuracy: 71% and 72%, reaction times: 275±9 and 300±10 ms, mean ±SE, respectively), relative to the auditory and visual unisensory performance (accuracy: 64% and 62%, reaction times: 345±14 and 338±10 ms, respectively). To investigate how the additional information available in the multisensory conditions leads to optimised listening performance, these data are being analysed in the context of previously published drift diffusion models. Lastly, to better understand how such signals are represented in the brain, recordings are presented from the auditory and visual cortex of anesthetised ferrets. Acknowledgements Supported by the Wellcome Trust / Royal Society References Raposo D., Sheppard J.P., Schrater P.R. & Churchland A.K. 2012. Multisensory Decision-Making in Rats and Humans. J Neurosci, 32(11), 3726-3735. 20. Visual –auditory interactions in ferret auditory cortex: The effect of temporal coherence H. Atilgan, S.M. Town, G.P. Jones, K.C. Wood and J.K. Bizley, Ear Institute, University College London, London, UK We have previously observed that when the size of a visual stimulus was coherently modulated with amplitude of a target auditory stream, human listeners were better able to report brief deviants in the target stream than when the visual stream was coherent with an non-target auditory stream (Maddox et al., submitted). Here, we used modified versions of the same stimuli to explore whether the coherence between temporally modulated auditory and visual stimuli can influence neuronal activity in auditory cortex. The auditory stimuli were continuous vowel sounds which were amplitude modulated with a <7 Hz noisy carrier envelope and had a duration of 20 seconds. Embedded within the continuous vowel were brief (200 ms) timbre deviants generated by altering 55 the first and second formant values of the reference vowel. Two such vowels (/u/ and /a/) were generated, with different fundamental frequencies, and modulated with independent envelopes. These were then presented either separately or concurrently. In both cases the auditory stimuli were accompanied by a luminance-modulated visual stimulus, whose envelope matched one of the auditory streams such that in the single stream case the auditory and visual stimuli were either temporally coherent or independent, and in the dual stream case the visual stimulus was coherent with one of the two streams. Recordings were made in anesthetised ferret auditory and visual cortex and in the auditory cortex of awake, passively listening ferrets. We explored how the visual stimulus influenced both the representation of the timbre deviants as well as the ability of neurons to represent the modulation envelope of the sound in auditory cortex when single auditory and visual streams were presented. Both timbre sensitive units and envelope sensitive units were encountered. Some units responded only to changes in timbre (16% of recorded units) while others responded only to changes in the amplitude envelope (26%). Over one third of the driven units (36%) showed sensitivity to both timbre and AM changes. Next, we quantified whether the coherence between the auditory and visual streams influenced auditory cortical responses. 32% of units showed a significant effect of visual envelope coherence; the main effect of the visual stimulus coherence appeared to be on how well a unit represented the changes in the amplitude modulation, with the responses to timbre changes being relatively uninfluenced. When both streams were presented some neurons responded preferentially to one stream over the other (presumably reflecting tuning for a particular pitch or timbre). In a small number of cases preliminary analysis suggests that some neurons responded preferentially to the auditory stream with which the visual stimulus was coherent. Acknowledgements Action on Hearing Loss, Wellcome Trust, Royal Society, BBSRC 21. Multisensory integration in the ferret A. Hammond-Kenny, V.M. Bajo, A.J. King and F.R. Nodal, Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, UK Our unified perception of the world relies on the ability to integrate information from different sensory modalities with dedicated neural circuits. The reported locations of multisensory integration in the mammalian brain are numerous and include primary sensory cortices. In order to explore the role of auditory cortex in multisensory integration, we aim to record neural activity from the ferret auditory cortex while animals perform a multisensory localization task, thus enabling correlation of behavioural performance, a measure of perceptual ability, with simultaneously recorded neural activity. However, prior to implantation of bilateral multi-electrode arrays, it is necessary first to obtain, for this species, baseline behavioural measurements of multisensory integration. Three ferrets were trained by positive operant conditioning to trigger a stimulus, presented from 1 of 7 sites separated by 30° intervals around the frontal hemifield of a circular arena, by licking a central spout. Stimuli were either unisensory (auditory or visual) or multisensory (audio-visual) of varying durations (20-2000ms) and, for auditory stimuli, of different intensities (56-84dB SPL). Animals received a water reward for approaching the correct target. A comparison of reaction times estimated by the latency of head-orienting movement and perceptual accuracy indicated by the percentage of correct approach-to-target responses between unisensory and multisensory stimulus conditions was used as a metric of multisensory integration. As expected, perceptual accuracy was greater for multisensory versus unisensory stimulus conditions with the greatest gains in percentage correct scores observed at the shortest stimulus durations. In addition, facilitation of reaction times exceeded those predicted by a probability 56 summation mechanism alone, as demonstrated by violation of the Race Model. Together, these results show behavioural enhancement during multisensory testing conditions and validate the use of this behavioural paradigm to identify the neural correlates of multisensory integration. Our initial recordings of neural activity from the auditory cortex, during the behavioural task, have revealed various types of neural responses to stimulus presentations although their multisensory properties remain to be established. Acknowledgements Supported by the Wellcome Trust 22. Auditory-visual interactions between amplitude-modulated sounds and shapes M.P. Laing, A. Dryden, H.H.I. Tung, A. Rees and Q.C. Vuong, Institute of Neuroscience, Newcastle University, Newcastle upon Tyne, UK. Experimental evidence demonstrates that visual speech information has a marked effect on what we hear, with respect to content (e.g. McGurk effect) and perceived location (ventriloquist effect). Previous studies of AV interactions using non-speech related stimuli have often used transient sounds that lack key features of speech. Here we address the hypothesis that auditory visual (AV) interactions observed with speech and faces generalise to simpler stimuli occurring on similar time scales to speech in both brain and behaviour. We have developed a paradigm using sinusoidally amplitude-modulated tones and a visual shape (a cuboid which expands and contracts at the modulation rate) that enables us to manipulate the extent to which listeners fuse the auditory and visual components of the stimuli. Listeners heard pairs of modulated tones (1.5 s duration, 250 Hz carrier frequency, 2 Hz modulation rate) and judged whether their modulation depth (range 20-52% in 8% steps) was the same or different (Audio Only). On cross-modal trials, we paired the shape with the tone with the greatest (AV >) or least (AV <) modulation depth. We found that the modulated shape influenced how listeners perceived the modulation depth of the tones, but modulation of the shape was not influenced by the modulated tone. The auditory-visual interaction was nearly eliminated when the tone and shape differed in their modulation frequency (e.g., 1 Hz shape modulation paired with 2 Hz amplitude modulation of the tone). Furthermore listeners reported that when the modulation frequency in both modalities was the same, they perceived the shape to be the source of the tone, thus indicating crossmodal integration. This merged percept was weak or absent when the modulation frequencies did not match. In a preliminary functional magnetic resonance imaging study using data acquired by sparse imaging, we found that superior-temporal, intraparietal and frontal areas responded more when the tone and shape had the same rather than different modulation frequencies. Our findings provide evidence (1) that auditory-visual interactions observed with speech and faces generalise to simpler stimuli occurring over a similar time scale to speech; and (2) that temporal information, such as modulation frequency, is important for fusing hearing and vision in both brain and behaviour. Acknowledgements MPL is supported by a studentship from the Wellcome Trust. 23. It’s all in the mix: Randomly mixing different auditory amplitude-modulation tasks during training disrupts learning 57 H. Swanborough, D.W. Maidment, and S. Amitay, MRC Institute of Hearing Research, Nottingham, UK There is a growing corpus of work in the auditory learning domain investigating the efficacy of mixed-task training. One concern surrounds the conditions under which mixed training of different tasks does or does not disrupt learning. A previous study from our laboratory examined training of two amplitude-modulation (AM) tasks involving different cues – AM-detection (AMD) and AMrate discrimination (AMR). We have shown that when the two tasks are interleaved in either 60-trial or 360-trial blocks, significant learning is only observed for the AM task completed last in a training session (Gill et al., 2013; Maidment et al., 2013). This suggests that the acquisition and consolidation of the ultimate AM task within a training session disrupts learning of the preceding task. The objective of the current study was to investigate whether learning on two similar AM tasks is disrupted when neither task has an ultimate block. Training consisted of interleaving AMD and AMR tasks on either a trial-by trial basis in each training block, or in separate, 60-trial blocks. We expected no learning to occur in the former mixed regimen, because neither task will have an ultimate block with a sufficient number of trials to allow acquisition and initiate consolidation. Preliminary results replicate our original finding that when training is mixed in separate, 60trial blocks, learning is only observed for the ultimate AM task. Furthermore, as predicted, learning on both AM tasks is not observed when neither task has an ultimate block that disrupts previous training. We conclude two similar tasks should not be mixed randomly on a trial-by-trial basis during training, as this prevents the acquisition and consolidation of both tasks. These findings have implications in the construction of training programmes intended for treatment and intervention strategies, particularly those aimed at the habilitation of cochlear implant users that rely on AM cues to enhance speech intelligibility. Acknowledgements Supported by the Medical Research Council. References Gill, E.C., Maidment, D.W., Kang, H-J., & Amitay, S. (2013). Disruption to learning during the acquisition phase. Poster presented at the British Society of Audiology Annual Conference, Keele University, UK. Abstract #88. Maidment, D.W., Kang, H-J., & Amitay, S. (2013). Auditory perceptual learning: Initiation of consolidation prevents dual-task disruption. Poster presented at the British Society of Audiology Annual Conference, Keele University, UK. Abstract #87. 24. Comodulated notched-noise masking R. Grzeschik*, B. Lübken* and J.L. Verhey*, *Department of Experimental Audiology, Otto von Guericke University Magdeburg, Magdeburg, Germany Detection thresholds of a masked sinusoidal signal are lower when the masking noise is coherently modulated over a wide frequency range compared to a masking condition with the same spectrum but using an unmodulated masker. This phenomenon is referred to as comodulation masking release (CMR; Hall et al., 1984). Here we tested how masker comodulation affects notched-noise data. Thresholds for sinusoids were measured in the presence of a diotic notched-noise masker for comodulated and unmodulated conditions. The signal frequency of the target was 500 Hz. The masker was a broadband filtered noise with lower- and upper cut-off frequencies of 30 and 1000 Hz, respectively. The masker had a spectral notch at the signal frequency. Notch widths were 0, 50, 58 100, 200, and 400 Hz. The arithmetic centre frequency of the notch was equal to the signal frequency. The data of eleven participants (4 male, 7 female, aged 25-37 yrs, mean 30 yrs) entered the analysis. As expected, thresholds decreased with increasing notch width for both masking conditions. The CMR decreased continuously with notch widths as well. CMR was almost 9 dB for no notch. At 400 Hz, CMR was less than 2 dB. The experimental data were compared with predictions of a modulation filterbank model that was previously shown to simulate CMR in the bandwidening type of CMR experiments (Piechowiak et al., 2007). Acknowledgements Supported by the German Research Foundation (SFB/TRR31). References Hall JW, Haggard MP, Fernandes MA (1984) Detection in noise by spectrotemporal patternanalysis. J Acoust Soc Am 76: 50-56. Piechowiak T, Ewert SD, Dau T (2007) Modeling comodulation masking release using an equalization-cancellation mechanism. J Acous. Soc Am 121 (4): 2111-2126. 25. Individual differences in auditory performance of naive listeners D.E. Shub, School of Psychology, University of Nottingham, Nottingham, UK A psychophysical experiment was conducted to examine differences between so called ‘good’ and ‘bad’ listeners. A total of 97 naive young healthy listeners were divided into three groups with listeners from each group being measured on two tasks: (1) frequency and level discrimination, (2) interaural time difference (ITD) and interaural level difference (ILD) discrimination, and (3) N0S0 and N0Spi detection. Two estimates of threshold were made with both a low (0.5 kHz) and a high (4 kHz) frequency tonal signal in every task, except for ITD discrimination, where only the low frequency signal was used. There was substantial correlation between the first and second estimates of threshold (r equal to 0.73, 0.69, and 0.91), but no statistically reliable differences between the thresholds (p equal to 0.72, 0.63, and 0.69). The correlation between the estimates of threshold at the two frequencies (r equal to 0.69, 0.69, and 0.64) was similar to the test-retest correlation. Relative to the test-retest correlation, the correlation between the N0S0 and N0Spi tasks (r=0.92) was similiar, but the correlation between the level and frequency tasks (r=0.37) and ITD and ILD tasks (r=0.4) were substantially reduced. The reduced correlation in the level and frequency tasks suggests that the relative performance of listeners is task dependent. This, however, seems contradictory to both the lack of a change in the N0S0-N0Spi correlation, where the two tasks are nominally measuring different aspects of detection, and the presence of a change in the ITD-ILD correlation, where the two tasks are nominally measuring spatial sensitivity. These discrepancies are potentially related to the generally poor performance of the naive subjects. The level of performance in the ILD task means that there were not only changes in the spatial position, but also changes in the loudness that are absent in the ITD task. Conversely, the N0S0 and N0Spi tasks may have been measuring the same thing since sensitivity to interaural differences is so poor in navie listeners and detection thresholds were so elevated that listeners would not need to rely on interaural decorrelation in the N0Spi task. This is supported by the lack of a binaural masking level difference (BMLD) with a low frequency signal (p=0.46), however, inconsistent with the 3.8 dB BMLD (p<0.001) with a high 59 frequency signal. Overall, the results suggest that subjects are not universally good/bad performers, but that the relative quality of the performance is task, but not frequency, specific. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 26. Frequency tuning of the efferent effect on cochlear gain V. Drga*, I. Yasin* and C. J. Plack§, *Ear Institute, University College London, UK, §School of Psychological Sciences, University of Manchester, UK Behavioural and physiological evidence suggests that the amount of gain applied to the basilar membrane may change during the course of acoustic stimulation due to efferent activation of the cochlea (Liberman et al, 1996). The present study used the Fixed Duration Masking Curve (FDMC) method (Yasin et al, 2013) to estimate human cochlear gain, in the presence of a precursor sound of variable frequency. Masker level at threshold was obtained for a 10-dB SL, 4-kHz sinusoidal signal (3-ms raised-cosine onset and offset ramps) in the presence of either an on-frequency (4 kHz) or off-frequency (1.8 kHz) sinusoidal forward masker (2-ms raised cosine ramps). The masker-signal silent interval was 0 ms. Thresholds were measured as a function of signal and masker steady-state duration, while holding the combined masker-and-signal duration constant at 25 ms. A short total duration ensured that accurate estimates of basilar membrane gain could be obtained in the absence of efferent suppression of the signal by the masker. Presentation of a precursor sound prior to the masker-signal stimulus was used to estimate the frequency dependency of the efferent effect. The use of the FDMC technique ensured that the effect of the precursor on gain could be measured independently of any masking effects by the precursor. This has been a confound in previous behavioural studies. FDMCs were obtained from three listeners with and without a 160-ms precursor noise with a bandwidth of 200 Hz (centred at frequencies of 2.5, 3.0, 3.5, 3.25, 4.0, 4.25, 4.5, 5.0 and 5.5 kHz) and a level of 60 dB SPL. The silent interval between precursor and masker was fixed at 0 ms. Results suggest that the efferent effect is tuned in frequency, with the greatest gain reduction for a precursor presented at a frequency at, or just below, the signal frequency of 4 kHz. The amount of gain reduction with a 3.0, 4.25, 4.5 and 5.0-kHz precursor was significantly less than that with a 4-kHz precursor. The 10-dB bandwidth of the efferent filter was approximately 590 Hz. The results are consistent with physiological studies which show efferent fibre frequency-response tuning (Guinan, 2011). Acknowledgements Supported by the EPSRC (grant EP/H022732/1) References Liberman M.C., Puria S. & Guinan J.J.Jr. 1996. The ipsilaterally evoked olivocochlear reflex causes rapid adaptation of the 2f1-f2 distortion product otoacoustic emission. J Acoust Soc Am, 99, 3572-3584. Yasin I., Drga V. & Plack C.J. 2003. Estimating peripheral gain and compression using fixedduration masking curves. J Acoust Soc Am, 133, 4145-4155. Guinan J.J.Jr. 2003. Cochlear efferent innervation and function. Curr Opin Otolaryngol Head Neck Surg, 18, 447-453. 60 27. Temporal integration in the perception of low-frequency and infrasound H. Patel and T. Marquardt, UCL Ear Institute, University College London, UK Hearing threshold and equal-loudness contours (ELCs) were obtained as a function of tone duration from 5 normal hearing subjects for 4 Hz, 16 Hz, 32 Hz and 125 Hz. The duration ELCs were also obtained for 1 kHz. Like for higher frequencies, thresholds decreased with increasing tone duration until the “critical duration”, beyond which threshold remained constant. The trend of an increase in critical duration with decreasing tone frequency, previously reported for the normal frequency range down to 125 Hz, continues into the infrasound range, where mean data indicate a critical duration of approximately 1 s at 16 Hz, and an further increase to at least 4 s at 4 Hz. Rates of threshold decrease below the critical duration were previously reported for higher frequencies to be in the order of 10 dB per ten-fold duration increase, indicating an integration of stimulus power. Our data demonstrate that this rate holds true down to infrasound. Although the critical durations in the perception of loudness are comparable to those for thresholds, the rates of sound pressure adjustment to keep an equal loudness percept with increasing tone duration (below the critical durations) are considerably lower, and fit well with a rule of 10 phon per ten-fold duration increase. Acknowledgements: This study was supported by the EMRP project HLT01 (EARS). T.M. is supported by the related EMRP Researcher Excellence Grant HLT01-Reg1. The EMRP is jointly funded by the EMRP participating countries within EURAMET and the European Union. 28. Mind the gap: two dissociable mechanisms of temporal processing in the auditory system L.A. Anderson*, J. Mattley* and J.F. Linden*, § *Ear Institute, University College London, London, UK, §Dept. Neurosci., Physiol. & Pharmacol., University College London, London, UK The high temporal acuity of auditory processing underlies perception of speech and other rapidly varying sounds. A common measure of auditory temporal acuity is the threshold for detection of brief gaps in noise. Gap-detection deficits, frequently observed in neurodevelopmental disorders, are considered evidence for 'sluggish' auditory processing. Here we show that neural mechanisms of auditory temporal processing are stimulus-specific, and gap-detection deficits do not imply a general loss of auditory temporal acuity. Extracellular recordings from auditory thalamus of mice with neocortical ectopia (a mouse model of thalamocortical developmental disorder) reveal reduced neural sensitivity to brief gaps in noise, but normal sensitivity to other rapidly changing sounds. We demonstrate, through experiments and mathematical modelling, that thalamic gap-detection deficits arise from weakened neural activity following noise offsets, not onsets. We also present preliminary results from further experiments aimed at identifying the anatomical origins of the observed physiological deficit, examining neural responses at the level of the inferior colliculus and the auditory brainstem. The results reveal dissociable mechanisms of central auditory temporal processing involving sound-onset-sensitive and sound-offset-sensitive channels, and suggest that auditory deficits in developmental disorders arise from abnormal neural sensitivity to sound offsets. Acknowledgements Wellcome Trust, Action on Hearing Loss 29. Build-up and resetting of auditory stream segregation: Effects of timbre, loudness, and abrupt changes in these properties 61 S.L. Rajasingam, R.J. Summers* and B. Roberts*, *Psychology, School of Life and Health Sciences, Aston University, Birmingham, UK Better characterization of the effects of acoustic change on the dynamics of stream segregation is important for understanding the role of streaming in everyday listening. The auditory system’s ability to segregate sounds into separate streams is commonly investigated using rapid, unchanging ABA– triplet sequences (alternating low- and high-frequency pure tones, followed by silence). Presenting long, steady sequences results in ‘build-up’ – an increased tendency towards a segregated percept – which tends to reset if an abrupt change occurs. Two experiments explored the effects of timbre and loudness on build-up, and the consequences of abrupt changes in these perceptual properties. 12 normal-hearing listeners heard 20-s-long ABA– sequences (tone duration = 100 ms; base frequency = 1 kHz; Δf = +4-8 semitones). Participants continuously monitored each sequence and used key presses to report when they heard the sequence as integrated or segregated (cf. Haywood & Roberts, 2013). Experiment 1 used pure tones and narrowly spaced (±25 Hz) tone pairs, dyads, to investigate the effects of timbre. Both stimuli evoke similar peripheral excitation patterns but dyads have a “rougher” timbre. Sequences consisting only of dyads induced a highly segregated percept, with limited scope for further build-up. Most notably, alternating abruptly every 13 triplets between pure tones and dyads produced large changes in the extent of reported segregation. Dyad-to-tone transitions produced substantial resetting, but tone-to-dyad transitions elicited greater segregation than for the corresponding time interval in dyad-only sequences (i.e., overshoot). Experiment 2 used pure-tone sequences and examined the effect of varying triplet loudness. Louder and softer sequences with a maintained level difference (12 dB) showed similar patterns of build-up, but a clear asymmetry occurred for alternating sequences. Specifically, soft-to-loud transitions caused a significant reduction in segregation (resetting) whereas loud-to-soft transitions had little or no effect (cf. Rogers & Bregman, 1998). Together, the results suggest that timbre can strongly affect the tendency to hear streaming, even in the absence of peripheral-channelling cues, but moderate differences in overall level do not. Furthermore, abrupt changes in sequence properties have variable effects on the tendency to hear segregation – including resetting, overshoot, and asymmetries in the effect of transition direction. Acknowledgements Supported by Aston University. References Haywood, N.R. & Roberts, B. 2013. Build-up of auditory stream segregation induced by tone sequences of constant or alternating frequency and the resetting effects of single deviants. J Exp Psychol Hum Percept Perform, 39, 1652-1666. Rogers, W.L., & Bregman, A.S. 1998. Cumulation of the tendency to segregate auditory streams: Resetting by changes in location and loudness. Percept Psychophys, 60, 1216-1227. 30. Influence of context on the relative pitch dominance of individual harmonics H.E. Gockel, S. Alsindi, C. Hardy and R.P. Carlyon, MRC Cognition and Brain Sciences Unit, Cambridge, UK There is evidence that the contribution of a given harmonic in a complex tone to residue pitch is 62 influenced by the accuracy with which the frequency of that harmonic is encoded: Harmonics with low frequency difference limens (FDLs) contribute more to residue pitch than harmonics with high FDLs (Moore et al, 1985; Gockel et al, 2005). We investigated whether listeners adjust the weights assigned to individual harmonics based on acquired knowledge of the reliability of the frequency estimates of those harmonics. Importantly, the acquired knowledge was based on context trials in which the reliability of the frequency estimates of individual harmonics was manipulated, rather than on the reliability of estimates in the test trials themselves. In a two-interval forced-choice task, seven listeners indicated which of two 12-harmonic complex tones had the higher overall pitch. In context trials (60% of all trials), the fundamental frequency (F0) was 200 Hz in one interval and 200+ΔF0 Hz in the other. In different blocks, either the 3rd or the 4th harmonic, plus (always) the 7th, 9th, and 12th harmonics were replaced by narrowband noises that were identical in the two intervals. Harmonics and narrowband noises had rms levels of 50 dB SPL. Feedback was provided. In randomly interspersed test trials (40% of all trials), the F0 was 200+ΔF0/2 Hz in both intervals and either the 3rd or the 4th harmonic was shifted slightly up or down in frequency. There were no narrowband noises. Feedback was not provided. Performance (d′) for judging the residue pitch was calculated separately for each mistuned harmonic (3rd or the 4th) and each context condition. In the test trials, d′ for a given harmonic was slightly but significantly smaller when it was replaced by a noiseband in the context trials, compared to when a different harmonic was replaced. The results are consistent with the notion that listeners give smaller weights to a harmonicfrequency region when they have learned that this frequency region does not provide reliable information for a given task. That is, the relative contribution of individual harmonics to pitch may depend not only on the accuracy of encoding, but also on the listener’s experience of how reliable the information is likely to be. Acknowledgements Supported by intramural MRC-funds. References Gockel H., Carlyon R.P. & Plack C.J. 2005. Dominance region for pitch: Effects of duration and dichotic presentation. J Acoust Soc Am, 117, 1326-1336. Moore B.C.J, Glasberg B.R. & Peters R.W. 1985. Relative dominance of individual partials in determining the pitch of complex tones. J Acoust Soc Am, 77, 1853-1860. 31. Effects of temporal context on frequency selectivity of cortical adaptation in primary auditory cortex O. Woolnough, J. de Boer, K. Krumbholz and C. Sumner, MRC Institute of Hearing Research, Nottingham, UK Adaptation, the reduction in neural responses to repeated stimuli, is a ubiquitous phenomenon throughout human and animal sensory systems. It has been shown that adaptation in auditory cortex is frequency specific (Brosch and Schreiner, 1997), with greater frequency separation between adapter and probe tones reducing the degree of adaptation. The aim of the current study was to compare frequency specificity of adaptation in human and guinea pig (GP) auditory cortex, and to investigate how adaptation specificity depends on the properties of the adapter and probe stimuli. A recent study has shown that using repeated presentation of adapters can enhance frequency specificity of this adaptation (Briley and Krumbholz, 2013). Here we have characterised the effects of both adapter repetition and duration on the tuning of cortical adaptation. 63 Adaptation was measured in auditory cortex, recorded through EEG in humans and via EEG and LFPs in GPs. Pure tone adapter-probe sequences were presented diotically with adapter frequencies of 0, 0.5 and 1.5 octaves higher than the 1 kHz probe. We analysed the resultant auditory evoked potential (AEP) amplitudes, measured as the N1-P2 amplitude, and quantified adaptation as the reduction in response size in the adapted probe compared to the unadapted response. The human EEG results confirm the previous results (Briley and Krumbholz, 2013) that repeated adapter presentation increases the frequency specificity of adaptation. Our results further show how these effects depend on number of repetitions and that no sharpening is observed when adapters are prolonged rather than repeated. The GP EEG results show comparable trends. However, frequency selectivity of adaptation was overall much greater in GPs than in humans (GPs showed a 50-105% reduction in adaptation over 1.5 octave frequency separation vs 15-45% in humans). In fact at the largest adapter-probe frequency separation, the GPs showed significant facilitation, increasing the AEP amplitude beyond the unadapted amplitude, whereas no facilitation was evident in humans. Our results suggest, in both humans and GPs, adaptation contributes to refine the frequency tuning of adaptation with increasing numbers of adapting tones. However, there are apparent qualitative differences in EEG responses across species. Further work will investigate how the underlying representation in neural activity and local field potentials contribute to far field potentials. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Briley P.M. & Krumbholz K. 2013. The specificity of stimulus-specific adaptation in human auditory cortex increases with repeated exposure to the adapting stimulus. J Neurophysiol, 110, 2679–2688. Brosch M. & Schreiner C.E. 1997. Time course of forward masking tuning curves in cat primary auditory cortex. J Neurophysiol, 77, 923–943. 32. Functional organisation of familiar melody recognition as revealed using octavegeneralisation of melodic structure R.E. Millman*, H. Biddles§, R. Garcia§, E. Giannakopoulou§, M. Kulkarni§ and M. Hymers* *York Neuroimaging Centre, University of York, York, UK, §Department of Psychology, University of York, York, UK Deutsch (1972) demonstrated that the accuracy of familiar melody recognition decreased significantly when successive notes of the melody were played in one of three octaves chosen at random (i.e. octave-scrambled). Further, it was reported that on listening to the original melody before the octave-scrambled version of the same melody, participants were better able to recognise the melody in the octave-scrambled condition. Here we used the phenomenon of Deutsch’s “mysterious melody” to study the neural processes underlying familiar melody recognition. Notably Deutsch’s “mysterious melody” illusion permits control for low-level acoustical confounds between “unfamiliar” and “familiar” conditions in a study of familiar melody recognition. First we tested the effects of four categories of melody “scrambling” (octave-scrambled, randomised-note, retrogradation and shuffled-note) on behavioural ratings of melody familiarity. Participants listened to a given “scrambled” version (pre-reveal) of a melody followed by the original melody (reveal) 64 and finally the “scrambled” version again (post-reveal). Following each melody presentation, participants were asked to give a 4-point scale familiarity rating. Post-reveal ratings of melodic familiarity were significantly greater for the octave-scrambled condition than for the other melody “scrambling” conditions. Secondly, we investigated the functional organisation of familiar melody recognition in an Interleaved Silent Steady State (ISSS) functional magnetic resonance imaging (fMRI) experiment (Schwarzbauer et al, 2006). The fMRI experiment closely followed the design of the behavioural experiment and octave-scrambled stimuli were used to manipulate the percept of familiar melodies. The responses to randomised-note stimuli were incorporated into an interaction term that controlled for temporal order effects in the fMRI experiment. The fMRI analyses of the original melody vs. “baseline” (silence) revealed activity in bilateral auditory cortices, bilateral superior temporal lobes, bilateral frontal orbital cortices, right precentral gyrus and left inferior frontal gyrus. The fMRI contrast of interest, based on the interaction term, revealed differential activity in the right inferior parietal lobe, right angular gyrus and left intraparietal sulcus. These findings support the recruitment of a bilateral network of supramodal brain regions in familiar melody recognition. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Deutsch, D. 1972. Octave Generalization and Tune Recognition. Percept Psychophys, 11, 411412. Schwarzbauer, C, Davis, M.H., Rodd, J.M. & Johnsrude, I. 2006. Interleaved silent steady state (ISSS) imaging: a new sparse imaging method applied to auditory fMRI. NeuroImage, 29, 774782. 33. A steady-state evoked response to periodic changes in carrier IPD – the effect of IPD size on the magnitude of the response N.R. Haywood, J.A. Undurraga, T. Marquardt and D. McAlpine, UCL Ear Institute, University College London, London, UK A steady-state evoked response that is dependent on sensitivity to interaural timing differences (ITDs) can be observed using electroencephalography (EEG). The stimuli comprised a five-minute sinusoidally amplitude modulated tone, presented at 80 dB SPL (carrier frequency = 520 Hz, modulation frequency = 41 Hz). The modulation envelope was diotic, and the carrier tone was presented with one of five possible interaural phase differences (IPDs): 22.5º, 45º, 90º, 112.5º, or 135º. At the end of every six successive modulation cycles, the carrier IPD was ‘switched’ abruptly, so that carrier phase led in the opposite ear than previously. This resulted in a stimulus which alternated periodically between right- and left-leading, but for which the overall size of the IPD was held constant. IPD switches were presented at a rate of 6.8 per second, and the timing of each switch coincided with a minimum in the modulation cycle. This ensured that the salience of monaural phase cues were minimised. Responses from ten normally-hearing subjects were recorded with a 16 channel EEG system. Subjects did not attend the stimuli during the recording session. The periodic IPD switches evoked a reliably detectable 6.8 Hz steady-state following response. All IPDs tested elicited a robust following response, but the magnitude of this response was typically largest when the IPD was set to 90º. Notably, following response magnitude decreased at the larger IPD values tested (112.5º, 65 135º). A possible explanation for this finding is that is that subjects applied greater weighting to the envelope IPD cues once the carrier IPD extended beyond 90º. As the modulation envelope was diotic, any increased weighting applied to the envelope IPD may have reduced the perceived lateralisation of the stimulus, and/or reduced the salience of the abrupt lateralisation changes during the stimulus. Previous psychophysical studies that have presented stimuli with envelope and carrier ITDs in opposing directions have observed greater envelope ITD weighting once the carrier IPD extended beyond ~ 45-90º (Buell et al., 1991; Dietz et al., 2009). Preliminary psychophysical estimates of the lateralisation of modified versions of the current stimuli are consistent with these previous findings. Acknowledgements The research leading to these results has received funding from the European Union’s Seventh Framework Programme (FP7/2007-2013) under ABCIT grant agreement number 304912. References Buell, T. N., Trahiotis, C. & Bernstein, L. R. (1991). Lateralization of low‐frequency tones: Relative potency of gating and ongoing interaural delays. J Acoust Soc Am, 90(6), 3077-3085. Dietz, M., Ewert, S. D, & Hohmann, V. (2009). Lateralization of stimuli with independent finestructure and envelope-based temporal disparities. J Acoust Soc Am, 125(3), 1622-1635. 34. Detection of interaural time processing by means of frequency following responses J.A. Undurraga, N. Haywood, T. Marquardt and D. McAlpine, University College London Ear Institute, London, UK Hearing loss in one or both ears can decrease the ability to use binaural cues, potentially making it more difficult to understand speech and to localize sounds. Moreover, in the case of bilaterally implanted cochlear implant (CI) users, a place mismatch between the channels in each ear may reduce the binaural performance. Thus, the use of a reliable and direct objective measure of the binaural processing may help to understand the mechanisms of binaural processing and also be used as a clinical tool to aid matching across-ear electrodes. In this study we developed an objective measure to assess binaural processing in normal hearing listeners; specifically, we measured frequency following response to abrupt interaural phase changes (FFR- IPCs) imposed on amplitude modulated signals in normal hearing listeners. The phase of the carrier signal presented to each ear was manipulated to produce discrete Interaural Phase Changes (IPCs) at minimums in the modulation cycle. The phase of the carrier was symmetrically opposed in each ear. For example, when the carrier phase was set to +5.5◦ in one ear, it was set to −5.5◦ in the other ear, resulting in an effective Interaural Phase Difference (IPD) of 11◦. The IPDs tested in this study ranged between 11 and 135◦. IPCs were presented at three different rates (3.4, 6.8, and 13.6 switches/s) and the carrier was amplitude modulated using several rates (27 to 109 Hz). Measurements were obtained using an EEG system with two channels. Recordings demonstrated that FFR-IPC could obtained from all participants in all conditions. However, the Signal-to-Noise Ratio was largest when IPCs were presented at 6.8 switches/s. FFRIPC were largest generally for IPDs between 45 and 90◦. Overall, FFR-IPC increased with modulation rate. We concluded that FFR-IPC can be used as a reliable objective measurement of binaural processing and may be a more suitable objective measure than the binaural interaction component (BIC) to match across-ear electrodes. Acknowledgements: 66 The research leading to these results has received funding from the European Union’s Seventh Framework Programme (FP7/2007-2013) under ABCIT grant agreement number 304912 35. What does gunfire sound like when you are the target? Investigating the acoustic characteristics of small arms fire Z.L. Bevis*, R.M. van Besouw*, M.C. Lower*, D. Rowan*, G.S. Paddan§, *Institute of Sound and Vibration Research, Southampton, UK, §Institute of Naval Medicine, Gosport, UK Locating the source of small arms fire is an important hearing task for infantry personnel. Bevis et al. (2014) report that personnel claim to have varying confidence in their ability to localise small arms fire and often rely on visual cues to confirm its origin. In order to measure localisation ability amongst personnel, a greater understanding of the acoustic characteristics of small arms fire is needed. Small arms fire (from SA80 rifles) was recorded using binaural (Knowles Electronic Manikin for Acoustic Research) and omnidirectional (B&K2250 sound level meter) microphones placed at 50 m intervals from the firer (up to 300 m) and within 30 cm of the bullet trajectory. The sounds created by small arms are unlike other short duration stimuli such as a click. It is known that a single shot creates two sounds: a high amplitude, high frequency ‘crack’ created by the supersonic bullet parting the air as it travels, followed by a quieter, lower frequency ‘thump’ sound, created by the muzzle blast as the bullet leaves the weapon (Beck et al, 2011). Acoustic analysis of the recorded stimuli shows that the delay between crack and thump increases as distance from the firer increases; a phenomenon described by personnel in the study of Bevis et al (2014). The thump sound produces interaural level differences (ILDs) and interaural time differences (ITDs) as expected for a subsonic sound; however, the supersonic crack results in decreased ILDs and ITDs as the sound itself propagates from the bullet at a point close to the head. The recordings give an insight into the cues available for localisation and the results of this study contribute to the (currently sparse) literature surrounding small arms fire acoustics. These data will inform the development of tests of localisation ability for personnel, to ensure that they can perform hearing tasks critical to the success of military missions. Acknowledgement Funding provided by the Surgeon General of the Ministry of Defence. References Bevis Z.L., Semeraro, H.S., van Besouw R.M., Rowan D., Lineton B. et al. 2014. Fit for the frontline? A focus group exploration of auditory tasks carried out by infantry and combat support personnel. Noise Health, 69, 127–135. Beck S.D., Nakasone H. & Marr K.W. 2011. Variations in recorded acoustic gunshot waveforms generated by small firearms. J Acoust Soc Am, 129, 1748–59. 36. Decoding sound location from auditory cortex during a relative localisation task K.C. Wood, S. Town, H. Atilgan, G.P. Jones and J. K. Bizley, The Ear Institute, University College London, UK Many studies have investigated the ability of human listeners to localise sounds in space to an absolute location (e.g. Stevens & Newman, 1936). Other studies have compared the minimum discriminable difference in spatial location that a listener can reliably discern; the minimum audible angle (Mills, 1958). However, very few studies have investigated relative sound localisation, i.e. 67 reporting the relative position of two sequentially presented sources. Determining the relative location of two sound sources or the direction of movement of a source are ethologically relevant tasks. Here we report multi-unit activity from auditory cortex of ferrets performing a relative localisation task. Ferrets were trained in a positively conditioned 2AFC task to report whether a target sound originated from the left or right of a preceding reference. In standard testing, the target and reference stimuli were 150 ms noise bursts separated by a 10 ms gap. The reference was presented from one of 6 locations in the frontal 180° and the target was presented from a speaker 30° to the left or right. We recorded using 32 individually moveable tungsten electrodes, in a 4x4 array in each side of the head. We have a total of 1284 recordings which showed sound-evoked activity (p<0.05, t-test of mean firing rate in 50 ms before and after stimulus onset) from 207 unique recording sites in two ferrets. Of the 1284 unit recordings, 34% and 26% of recordings from ferret 1 and ferret 2 showed significant tuning to the location of the reference sound (p<0.05 ANOVA on firing rate during reference presentation and reference location, post-hoc analysis to look at tuned locations). Overall, 39% of unit recordings were tuned to contralateral space, 10% tuned to midline locations, and 51% were ipsilaterally tuned. Preliminary ROC analysis on the target-evoked spike count showed that a small fraction of the tuned unit recordings showed a significant choice probability (16%) and a similar fraction (16%) reliably reported the direction of the stimulus. On-going analysis is exploring the contribution of ILD and ITD cues to spatial tuning through the use of band-pass stimuli, and relating other response measures, such as cortical location and frequency tuning, to spatial tuning and the likelihood of observing a significant choice probability. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Mills A.W. 1958. On the minimum audible angle. Journal of the Acoustical Society of America. Stevens S. & Newman E.B. 1936. The localization of actual sources of sound. The American Journal of Psychology. 37. The binaural masking level difference (BMLD) in single cells of the auditory cortex H.J. Gilbert, T.M. Shackleton, K. Krumbholz and A.R. Palmer, MRC Institute of Hearing Research, University Park, Nottingham, UK The binaural masking level difference (BMLD) is an often used model for the improvement in masked signal detectability when the signal and masker are separated in space. In the BMLD a tone signal that is identical at the two ears (S0), masked by a noise that is also identical at the two ears (N0) can be made 12-15 dB more detectable by sign-inverting either signal or noise at one ear (Sπ or Nπ). Single-cell responses to BMLD stimuli were measured in the primary auditory cortex of Urethane-anaesthetised guinea pigs. Firing rate was measured as a function of the presentation level of 500 Hz S0 and Sπ pure tone signals in the presence of N0 and Nπ maskers. The maskers were white noise, low-pass filtered at 5 kHz, with a spectrum level of 23 dB SPL. Tone and masker were both 100-ms duration and switched on and off simultaneously. Responses were similar to those previously reported in the inferior colliculus (IC: Jiang et al, 1997 a,b; McAlpine et al, 1996; Palmer et al, 2000). At the lowest tone signal levels the response 68 was dominated by the noise masker, at higher signal levels the firing rate either increased or decreased. Signal detection theory was used to determine detection threshold. Very few neurones yielded measurable detection thresholds for all four stimulus conditions, and there was a wide range in thresholds. However, estimated from the lowest thresholds across the entire population, BMLDs were 14-15 dB, consistent with human psychophysical BMLDs. As in the IC, the shape of the firing-rate vs. signal-level functions depended on the neurons’ selectivity for interaural time difference (ITD). In summary, like in the IC, the responses were consistent with a cross-correlation model of BMLD with detection facilitated by either a decrease or increase in firing rate. Declaration of interest: The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Jiang D., McAlpine D. & Palmer A.R. 1997a. Detectability index measures of binaural masking level difference across populations of inferior colliculus neurons. J Neurosci 17, 9331-9339. Jiang D., McAlpine D. & Palmer A.R. 1997b. Responses of neurons in the inferior colliculus to binaural masking level difference stimuli measured by rate-versus-level functions. J Neurophysiol 77, 3085-3106. McAlpine D., Jiang D. & Palmer A.R. 1996. Binaural masking level differences in the inferior colliculus. J Acoust Soc Am 100, 490-503. Palmer A.R., Jiang D. & McAlpine D. 2000. Neural responses in the inferior colliculus to binaural masking level differences created by inverting the noise in one ear. J Neurophysiol 84, 844852. 38. Distribution and morphology of cells expressing calcium binding proteins in the rat inferior colliculus and their excitatory and inhibitory inputs. B.M.J. Olthof, L.D. Orton, C. Jackson, S.E. Gartside and A. Rees, Institute of Neuroscience, Newcastle University, Newcastle upon Tyne, UK. The calcium binding proteins (CBP), calbindin (CB), parvalbumin (PV) and calretinin (CR) are the most abundant fast acting cytoplasmic calcium buffers in the central nervous system and are often implicated in neurological disorders and age related decline. Expression of these proteins has been described in all major subdivisions of the inferior colliculus, although the specific roles of the cells expressing them are largely unknown. To gain a better insight into the different functions these neurons play within the IC, we examined the glutamatergic and inhibitory inputs these cells receive. Furthermore, we aimed to establish what proportion of the CBP expressing neurons are GABAergic. It has often been suggested that most if not all PV expressing cells are GABAergic, while this is uncertain for CB and CR (Ouda & Syka, 2012). Adult Wistar rats were perfused with PBS followed by 4% paraformaldehyde. The tissue was cryoprotected in sucrose and 40 μm frozen sections were cut on a cryostat. PV, CB, CR, vesicular transporters for glutamate (VGLUTs 1 and 2), the inhibitory neurotransmitters GABA and glycine (VGAT), and GABA were detected using monoclonal and polyclonal antibodies. Binding of all antibodies was visualised by using appropriate secondary antibodies conjugated with fluorophores. In line with previous reports, we found that most PV expressing cells were located in the central nucleus and most CR and CB positive cells were in the dorsal and lateral cortices. Excitatory synapses containing VGLUT1 and 2 were found to be abundant in the dorsal and lateral cortices. In the central nucleus the labelling of these synapses varied along the rostral caudal axis with denser labelling in the caudal areas of the IC. Most somata, dendrites, and axons of PV expressing cells 69 were found to be surrounded by VGLUT2 labelled terminals that are known to have their origins in the IC and other brainstem nuclei. VGLUT1 labelling, characteristic of descending inputs from the auditory cortex (Ito & Oliver, 2010), did not show this pattern. Terminals expressing the inhibitory marker VGAT were found to make contact with the dendrites and axons, but rarely the somata, of PV positive cells. As observed in other studies, most, but not all, PV cells were found to co-express GABA. Acknowledgements Supported by Action on Hearing Loss. B.M.J.O is supported by a studentship from Newcastle University. References Ito, T. and Oliver, D.L. 2010. Origins of glutamatergic terminals in the inferior colliculus identified by retrograde transport and expression of VGLUT1 and VGLUT2 genes. Front Neuroanat. 4, 135 Ouda, L. and Syka. J. 2012. Immunocytochemical profiles of inferior colliculus neurons in the rat and their changes with aging. Front Neural Circuits, 2012, 6, 68 Poster Abstracts: Hearing loss and hearing aids 39. A comprehensive survey of questionnaires on hearing K. Foreman, J. Holman, M.A Akeroyd, MRC/CSO Institute of Hearing Research - Scottish Section, Glasgow, United Kingdom. As part of wider work on outcome measures for hearing interventions, we collated copies of every published hearing difficulty questionnaire that we could find. The search was part incremental, part systematic, but still essentially comprehensive: we believe that we have included every major questionnaire, most of the others, and a fair fraction of those obscure. We excluded those wholly devoted to tinnitus, children, or cochlear implants. In total we found 178 questionnaires. We classified 109 as “primary” (e.g., the SSQ), while the remaining are 69 are variants or derivatives (e.g. SSQ-B, SSQ-C, and SSQ-12). In total, there were 3614 items across all the primary questionnaires. The median number of items per questionnaire was 20; the maximum was 158. We report here various summaries of the questionnaires and of the items included (e.g. grouped by whether they related to hearing per se, the effects of hearing aids, or the repercussions or a hearing loss). Across all items, about one third were concerned with the person’s own hearing, another third with the repercussions of it, and about a quarter with hearing aids. Within the repercussions grouping, nearly half asked about social situations, a quarter with feelings, and a fifth with family or friends. Many questionnaires had common themes; for instance, including some form of “How much difficulty do you have with your hearing” or “how annoyed does your partner get at your hearing loss”. These themes essentially represent what hearing scientists and clinicians have thought important enough about hearing to be worth asking. But it is also clear that the choice of items is an indirect result of the genealogy of the questionnaires, in that many questionnaires take some of their items from an earlier questionnaire, which took them from an earlier one, and so on (e.g., parts of the SSQ were based on the AIAD, the PHAP, and Noble et al.’s (1995) unnamed questionnaire on spatial hearing). We conclude with some recommendations for future questionnaire designs. 70 Acknowledgements Supported by intramural funding from the Medical Research Council (grant number U135097131) and the Chief Scientist Office of the Scottish Government. References Noble, W., Ter-Horst, K. & Byrne, D. 1995. Disabilities and handicaps associated with impaired auditory localization. J Am Acad Audiol, 6(2), 129–140. 40. Developing a perceptual benchmark for speech-intelligibility benefit D. McShefferty, W. Whitmer and M. Akeroyd, MRC/CSO Institute of Hearing Research – Scottish Section, Glasgow Royal Infirmary, UK How large does a change in speech in noise need to be before for it to be meaningful to someone? We here attempt to answer this question using objective and subjective methods. First, we measured the just-noticeable difference (JND) in signal-to-noise ratio (SNR) to find the lower limits of perceptually relevant features (e.g., noise reduction). Second, we measured the minimum SNR change necessary to spur someone to seek out an intervention using different subjective-comparison paradigms. To measure an SNR JND, we used a variation on the classic level discrimination paradigm using equalised sentences in same-spectrum noise with various controls and roves to ensure that the task could only done by listening to SNR, not level per se. Averaged across participants, the SNR JND was 3 dB. This value was corroborated using different participants in a fixed-level task. JNDs were not correlated with hearing ability. To measure the subjective import of an increase in SNR, we presented paired examples of speech and noise: one at a reference SNR and the other at a variably higher SNR. In different experiments, participants were asked (1) to rate how many successive conversations they would tolerate given each example, (2) to rate the ease of listening of each example, (3) if they would be willing to go to the clinic for the given increase in SNR, and (4) if they would swap the reference SNR for the better SNR example (e.g., their current device for another). The results showed that the mean listening-ease ratings increased linearly with a change in SNR (experiments 1-2), but an SNR change of at least 6 dB was necessary to motivate participants to seek intervention (experiments 3-4). To probe individual variance, a questionnaire of general and hearing health was also administered to participants in the latter experiments. Overall, the results indicate not only the difference limen for SNR, but also how large a change in SNR is needed for it to be meaningful to someone. While an SNR increase less than 3 dB may have relevance to speech-recognition performance, it may not be enough of an SNR improvement to be reliably recognized and, furthermore, may be too little increase to motivate potential users. Acknowledgements Supported by intramural funding from the Medical Research Council (grant number U135097131) and the Chief Scientist Office of the Scottish Government. 41. Assessing the contributions of cognition and self-report of hearing difficulties to speech perception tests A. Heinrich*, H. Henshaw§,M. Ferguson§, *MRC Institute of Hearing Research, Nottingham, §NIHR Nottingham Hearing Biomedical Research Unit, Nottingham, UK. 71 Perceiving speech, particularly in noise, becomes more difficult as people age. Listeners vary widely as to their level of speech perception abilities even when hearing sensitivity is similar. Speech perception abilities can be measured behaviourally using speech tasks, or subjectively by assessing everyday hearing difficulties using self-report questionnaires. Here, we examined how behavioural measures of speech perception performance relate to cognition and self-report of hearing difficulties, while controlling for any confounding effects of age and hearing loss. A group of 44 listeners aged 55-74 years with mild sensorineural hearing loss were tested with speech tests that differed in linguistic complexity (phoneme discrimination in quiet, and digit triplets (DT) and sentences in noise)(Ferguson et al., 2014). Self-report questionnaires assessed hearing abilities (e.g. Speech, Spatial and Qualities of Hearing) and Quality of Life (QoL; EuroQol). Cognitive tests included attention (Test of Everyday Attention) and memory (Visual Letter Monitoring, Digit Span). Exploratory regression models were used to assess the relationship between speech perception performance with both cognition and self-reported hearing ability/QoL for each speech perception task, while controlling for age and hearing loss. Test-retest effects were measured in 21 participants. Cognition explained more variance in speech perception performance (28%) than hearing loss and age (20%), although most speech perception variance was explained by self-report measure scores (61%). The amount of explained variance differed for individual speech tests, which was greater for self-report than cognitive measures. On retesting, the explained variance attributed to hearing loss and age decreased, whilst that attributed to cognition increased. For the DT test, the relative contributions of cognition and self-report differed according to which stimulus (speech or noise) was adjusted adaptively. DT performance with variable noise was better predicted by self-report, whereas DT performance with variable speech was better predicted by cognition. Researchers and clinicians need to be aware that the relationships between cognition, self-report of hearing and hearing loss, as well as methodological issues, are important when using speech tests as outcome measures for people with mild hearing loss. Acknowledgements This research was funded by the National Institute for Health Research and the Medical Research Council. References Ferguson M.A., Henshaw H., Clark D.P. & Moore D.R. 2014. Benefits of phoneme discrimination training in a randomized controlled trial of 50-to 74-year-olds with mild hearing loss. Ear Hear, 35, e110-121. 42. Auditory and working memory training in people with hearing loss M. Ferguson, A.Tittle and H. Henshaw. NIHR Nottingham Hearing Biomedical Research Unit, Nottingham, UK. One in ten older adults (55-74 years) has a significant hearing impairment. Auditory training (AT) is an intervention designed to help listeners compensate for degradation in the auditory signal. Auditory training is hypothesised to aid listening in two ways; bottom up refinement of sensory perception and development of top-down cognitive control. For home-based training interventions to benefit people with hearing loss (PHL), any task-specific learning needs to transfer to functional benefits in real-world listening. A randomised controlled trial (RCT) of 44 adults with mild sensorineural hearing loss (SNHL) showed significant on-task learning for a phonetic-discrimination task (p<.001), and generalisation 72 to improvements in self-reported hearing (p<.05), with the greatest improvement shown for a challenging listening condition (Ferguson et al., 2014). Significant improvements were shown for complex measures of divided attention (p<.01) and working memory (WM) (p<.05), with the greatest improvements in complex conditions. No improvements were shown for speech-in-noise. These findings suggest that benefits of auditory training may be associated with the development of cognitive abilities over and above the refinement of sensory processing. A double-blind RCT aimed to assess whether directly training cognition can offer increased benefit to HA users with mild-moderate SNHL. Participants (n=57) were randomly allocated to an active-control group (ACG) or an adaptive-training group (ATG) and completed 30-40 minutes of Cogmed® WM training, 5 days/week over 5 weeks. Participants in the ATG received adaptivedifficulty training based on individual performance. Participants in the ACG received a nonadaptive practice-level version of Cogmed®. All training was delivered online. Optimal measures of speech perception, cognition and self-reported hearing were assessed pre- and post-intervention (Henshaw & Ferguson, 2014). Significant on-task learning generalised to improvements in trained WM (backwards digit span, p<.001) and untrained WM (visual letter monitoring, p<.001) tasks. No improvements were shown for a competing-speech perception task. It is argued that training is most effective when it targets core deficits. Thus, training interventions that target cognitive skills embedded within auditory tasks are most likely to offer generalised benefits to the real-world listening abilities of people with hearing loss. Acknowledgements This research was funded by the National Institute for Health Research. References Ferguson M.A., Henshaw H., Clark D.P. & Moore D.R. 2014. Benefits of phoneme discrimination training in a randomized controlled trial of 50-to 74-year-olds with mild hearing loss. Ear Hear, 35, e110-121. Henshaw H. & Ferguson M.A. 2014. Assessing the benefits of auditory training to real-world listening: identifying appropriate and sensitive outcomes. In: T. Dau (ed.) Proc ISAAR 2013: Auditory Plasticity - Listening with the Brain. Nyborg, Denmark.pp. 45-52. 43. A James Lind priority setting partnership for mild to moderate hearing loss – what should research listen to? H. Henshaw*, L. Sharkey§, V.Vas* and M. Ferguson* on behalf of the JLA Mild to Moderate Hearing Loss PSP Steering Group. *NIHR Nottingham Hearing Biomedical Research Unit, UK. § Hearing Link. Adult-onset hearing loss is the 13th most common disease burden worldwide (WHO, 2004). Yet funding for research into hearing remains limited with just £47 per disability-adjusted life year (DALY) spent on hearing research, compared to the significantly higher amounts spent on other priority areas, such as vision (£99 per DALY). It is estimated that up to 85% of funds invested in medical research have no direct value to aid patient need and satisfaction (Chalmers and Glasziou, 2009). There is general consensus that in order for research to have an impact on those who may benefit from it, the research process itself should engage clinicians, patients and members of the public as much as possible. The James Lind Alliance (JLA) Priority Setting Partnership (PSP) is an initiative that unites patients, clinicians and professional bodies to be active members in research prioritisation. The JLA PSP will offer patients of a given health condition and their clinicians the opportunity to identify 73 and prioritise “uncertainties” about interventions they think should be at the forefront of current research. A JLA PSP has been recently formed to identify and prioritise uncertainties for the treatment of mild to moderate hearing loss. The project will be driven forward by a Steering Group representing patients (patient representatives, Hearing Link, AoHL, GP), clinicians, and their representative bodies (BSA, BAA, BSHAA and ENT UK). A survey will be developed and distributed through the stakeholders to raise awareness to as many clinicians and patients as possible. This survey is designed to determine uncertainties across a range of aspects along the patient-pathway for those with mild to moderate hearing loss (e.g. prevention, screening, diagnosis, management). Responses will be collated, grouped into themes, and indicative research questions generated for prioritisation. Indicative questions will be identified as ‘true’ uncertainties by comparing to existing research literature. A final prioritisation workshop will be held in which patients and clinicians rank the top 10 uncertainties, to be disseminated to research commissioners and funding bodies. The uncertainties will be formatted for inclusion in the UK Database of Uncertainties about the Effects of Treatments (UK DUETs). Acknowledgements This research was funded by the Nottingham University Hospitals NHS Trust Charity and the National Institute for Health Research. References World Health Organization (2004). The global burden of disease: 2004 update. Switzerland: WHO press. Chalmers I, Glasziou P (2009). Avoidable waste in the production and reporting of research evidence. Lancet 374:86-89. 44. Challenges in testing binaural temporal fine structure cues in hearing impaired listeners R.K.Hietkamp, L.Bramsløw, M.Vatti and N.H.Pontoppidan, Eriksholm Research Centre, DK Recent studies (e.g. Lunner et al, 2012) have shown that temporal fine structure (TFS) information plays a role in complex listening situations. Therefore it is important to investigate the degraded nature of TFS perception in hearing-impaired (HI) listeners. TFS abilities can be tested both monaurally, e.g. frequency discrimination, or binaurally, e.g. spatial cues in the form of interaural time differences (ITD) or interaural time envelope differences (ITED). The present study focuses on the challenges of testing TFS binaurally versus monaurally. We tested 8 normal hearing and 26 hearing impaired listeners with a 2 alternative-forced-choice (2AFC) test. Prior to testing, listeners were subjected to a stepwise training (Hietkamp et al, 2010) where the magnitude of the stimuli was considerably larger than in the test itself. NH listeners showed stable monaural and binaural test results. Binaural TFS (ITD) and binaural envelope (ITED) thresholds for this group were below the boundary of real-life sounds imposed by the human head size. For HI listeners, the results of the binaural TFS (ITD) and binaural envelope (ITED) tests showed increased variability compared to the monaural 2AFC test. Furthermore, ITD and ITED thresholds often exceeded the boundary of real-life sounds; the consequence was that the presented stimuli alternately were below and above this boundary, evoking an ambiguity in percept, going from mid towards one side, mid towards both ears or even change of loudness. Because of the nonlinear and multidimensional relation of the percepts to the magnitude of the manipulations in the binaural ITD and ITED tests, an effective stepwise training was not possible. Lack of significant results might be due to an inability of providing sufficient training. 74 Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Hietkamp R.K., Andersen M.R. & Lunner T. 2010. Perceptual audio evaluation by hearingimpaired listeners – some considerations on task training. In: Buchholz J., Dalsgaard J.C., Dau T. and Poulsen T. (Eds.). Binaural Processing and Spatial Hearing. Proc 2nd Int Symp Audiological Auditory Res, ISAAR, p. 487-496. Lunner, T., Hietkamp, R.K., Andersen, M.R. & Hopkins, K. 2012. Effect of Speech Material on the Benefit of Temporal Fine Structure Information in Speech for Young Normal-Hearing and Older Hearing-Impaired Participants. Ear Hear, 33, 377-388. 45. Evaluation of individualized binaural earpiece for hearing-aid applications F. Denk*#, M. Hiipakka*#, B. Kollmeier*# and S.M.A. Ernst*#, #Cluster of Excellence "Hearing4all", *Medizinische Physik, University of Oldenburg, Oldenburg, Germany Modern hearing aid signal processing is increasingly able to offer great benefit to listeners. Already a target group with moderate hearing loss or even near-to-normal hearing could profit from some hearing aid features, such as binaural directional filters, noise suppression or speech enhancement. However one important challenge still is to achieve high perceptual plausibility and a natural sound impression, i.e. acoustical transparency of the used hearing devices in order to gain acceptance especially among potential first-time users. This work evaluated a new open-fitting hearing aid earpiece prototype that can be individually calibrated in situ to provide acoustic transparency, i.e. it is individually equalized to produce a similar frequency response as the open ear canal. Each prototype includes two balanced-armature receivers and three microphones. The microphones are located inside the ear canal, at the ear canal entrance, and in the concha. Both of the receivers and two of the microphones are located inside a small tube that can easily be inserted into individually shaped earmolds. Evaluation methods include technical measurements using a dummy head with exchangeable ear canal simulator (DADEC) and listening tests with human test subjects. The degree of perceived transparency was determined using a three-interval two-alternative (ABX) forced choice procedure, comparing the new earpiece with a simulated open ear channel reference. 25 repetitions of 20 short audio samples divided in three categories (speech, music and environmental) were presented in random order. 10 normal hearing subjects (5 experts, 5 non-experienced) participated in the experiment. A statistical analysis of the obtained results shows that an accurate distinction between the two presentation conditions was not significantly possible for the categories music and speech. However for technical sounds (e.g. ringtones) with considerable energy up to high frequency the subjects were able to differentiate with higher accuracy. In a second test, the earpiece was evaluated for subjective listening quality when driven with a selection of hearing aid algorithms using a combined discrimination and classification (CoDiCl) test routine. Overall it could be shown that the new earpiece offers a high degree of signal control, which allows for individual high quality and transparent sound presentation. It is therefore promising to serve as a high-fidelity assisting listening device, which can be scaled up to a complete hearing aid when necessary, but also is accepted by near-to-normal hearing patients. Thus, it can guide the way to a sooner and therefore more effective treatment of hearing loss. Acknowledgements Supported by BMBF 13EZ1127D and DFG FOR1732. 75 46. Obstacle avoidance using echolocation and sensory substitution information A.J. Kolarik*§, A.C. Scarfe*, S. Cirstea*, B.C.J. Moore§* and S. Pardhan*, *Vision and Eye Research Unit, Anglia Ruskin University, Cambridge, UK, §Department of Psychology, University of Cambridge, Cambridge, UK In the absence of vision, sound can provide important spatial information to aid navigation. Bats and dolphins echolocate by emitting sound bursts and use the returning echoes to detect objects in the vicinity. Some blind people also use self-generated clicking sounds to echolocate, and both blind and sighted people can be trained to echolocate (Rowan et al, 2013; Kolarik et al, 2014). Very few studies have investigated the effectiveness of echolocation for human navigation, which requires precise motor responses to avoid collisions when walking around obstacles. Empirical evidence showing that echolocation can guide human locomotion is lacking. Sensory substitution devices (SSDs) that work on an echolocation principle have been developed as electronic travel aids for the blind. Tactile SSDs use an ultrasound source and a receiver to detect signal reflections from nearby obstacles, and the rate of vibration applied to the skin informs the participant about distance to the obstacle. However, few studies have assessed the accuracy of navigation using SSD information. The current study compared performance using echolocation, vision, and a tactile SSD for blindfolded sighted participants navigating around a single flat 0.6 x 2 m obstacle. It was expected that vision would provide much better locomotor control than echolocation or the SSD. The obstacle was positioned at the midline, or 25 cm to the right or left of the participant, and the position was varied randomly from trial to trial. Participants approached the obstacle from distances of 1.5 or 2 m. Motion analyses were conducted using a Vicon 3D system, which allows the accurate quantification of human movement. Buffer space (clearance between obstacle and shoulder), movement times, and number of impacts were smallest with visual guidance, larger for the SSD, and greatest for echolocation. The changes in movement times suggest increased caution with the SSD and echolocation. The results show that echolocation can be used to guide human movement, but movement is not as accurate as when an SSD or visual information is used. Acknowledgements The research was supported by the Vision and Eye Research Unit (VERU), Postgraduate Medical Institute at Anglia Ruskin University. References Kolarik A.J., Cirstea S., Pardhan S. & Moore B.C.J. 2014. A summary of research investigating echolocation abilities of blind and sighted humans. Hear Res, 310, 60-68. Rowan D., Papadopoulos T., Edwards D., Holmes H., Hollingdale A., Evans L. & Allen, R. 2013. Identification of the lateral position of a virtual object based on echoes by humans. Hear Res, 300, 56-65. 47. Growing cochlear fibrocyte cultures on transplantable matrices D.N. Furness, M.Z. Israr, N.C. Matthews, A.H. Osborn, K. Riasat and S. Mahendasingam, Institute for Science and Technology in Medicine, Keele University, Keele, UK Lateral wall fibrocytes play an important role in maintaining cochlear fluids by recycling potassium from endolymph to perilymph. It is thought that fibrocyte degeneration may lead to hearing loss so replacing fibrocytes may prevent this (Mahendrasginam et al, 2011). Fibrocyte cultures are a source 76 of potentially transplantable cells, but when grown in standard culture conditions theys do not resemble native fibrocytes and have proved difficult to transplant successfully in vitro. We have developed methods of fibrocyte culture in injectable matrices of collagen I. Here we evaluate the morphology and ultrastructure of these cells. We prepared gels in two ways. (i) 3 mg/mL collagen I gels were made in standard 24 well plates and solidified. Cells grown in 2D cultures in T25 flasks were harvested and seeded at a density of 10 000 – 50 000 cells per cm2 onto the surface of the gels and grown for 1 – 4 d. (ii) we developed a simple way of culturing cells in the collagen gel solidified within a pipette tip. Approximately 100 L of 3 mg/mL final concentration collagen I gel solution was drawn up into a 1 mL syringe through a cut down yellow pipette tip. This was followed by 50 L of cell suspension. The gels became solidified overnight and were kept for 1 – 2 d. Gels were then removed from the wells or extruded from the tips, fixed and prepared for either scanning or transmission electron microscopy, or for immunocytochemistry using our standard methods. Cells survived for at least 2 d on all the gels. Immunocytochemistry showed that almost all expressed aquaporin, a marker characteristic primarily of native type III fibrocytes. However, some cells also expressed S-100 or Na,K,ATPase, characteristic of other fibrocyte types, suggesting a mixed phenotype. Scanning and transmission electron microscopy revealed cell morphologies that were partially consistent with the suggestion that the cells were phenotypically type III, but they also had other features that were less reminiscent of native cochlear fibrocytes. These data suggest that it is possible to grow viable fibrocyte cultures in gels that could be used for transplantation. The extrusible gels can be manipulated relatively easily from their pipette tip, providing a potentially ideal way of introducing them into the lateral wall under control, without damage to the culture cells themselves. Acknowledgements This research was funded by Action on Hearing Loss.and the Freemasons’ Grand Charity References Mahendrasingam S., Macdonald J.A. & Furness D.N. 2011. Relative time course of degeneration of different cochlear structures in the CD/1 mouse model of accelerated aging. J Assoc Res Otolaryngol 12, 437-453. 48. Ultrastructural localization of TMHS in mouse cochlear hair cells C.M. Hackney*, S. Mahendrasingam§, D.N. Furness§, and R. Fettiplace‡, *Department of Biomedical Sciences, University of Sheffield, UK, §Institute for Science and Technology in Medicine, Keele University, UK, ‡Department of Physiology, University of Wisconsin-Madison, USA. Hair cells are the mechanosensors in the cochlea where they convert the mechanical vibrations caused by sound into electrical signals to be carried along the auditory nerve to the brain. Deflection of the stereocilia in hair cells is believed to place a strain on these channels which then open, letting through K+ ions which results in depolarization of the cell. The molecular components of the mechanotransduction channels of the hair cells are still under investigation. Recently, mechanotransduction has been shown to be impaired in mice lacking a tetraspan membrane protein, TMHS. Immunofluorescence suggests this protein is found near the stereociliary tips (Xiong et al, 2012). Here we have utilized post-embedding immunogold labelling and stereological analysis to refine and more precisely localize this protein in p5 mice. Our data confirm the suggested stereociliary tip localization but they also show TMHS in the cytoplasm as well as in the membranes suggesting that during development, this protein is being shipped to the membranes from the manufacturing areas of the cell. It would be interesting to know if this shipment continues into adult life as the turnover time of this protein may be relevant to 77 understanding some kinds of hearing loss. As mutations of the gene for TMHS lead to deafness and because it may show the precise anatomical position of the mechanotransduction channels, we believe that ultrastructural localization could provide further clues to its function. Immunogold postembedding labelling gives a way of doing this, particularly if combined with quantitative analysis of precise localization. Acknowledgements This work has been sponsored by the NIH USA. References Xiong W., Grillet N., Elledge H.M., Wagner T.F., Zhao B., Johnson K.R., Kazmierczak P., Müller U. 2012. THMS is an integral component of the mechanotransduction machinery of cochlear hair cells. Cell, 151(6), 1283-129. 49. A real-time hearing aid simulator on the open development platform X. Yang, U. K. Ravuri, A. Gautam, S. Kim and S. Bleeck, Institute of Sound and Vibration Research, University of Southampton, UK With the development of digital technology, hearing aids have become more sophisticate. Advanced signal processing strategies have become possible, such as noise reduction, feedback cancellation, frequency compression/lowering and bilateral processing, because of the explosion of possible techniques. Evaluation of various acoustic processing schemes in realistic environment is essential. To facilitate this, we develop a real-time evaluation platform ― the "Open Development Platform" (ODP), to implement real-time acoustic system. The main goals are: To enable usage of wide range of users with a minimum of programming knowledge Using MATLAB® codes in a simple Simulink® environment (using the Simulink Real-Time® Toolbox) To support a wide range of different hardware platforms (currently, ODP supports Speedgoat® target machines, Texas Instruments® DSP-based systems, and Raspberry PI® boards) To provide a library of 'standard' acoustic signal processing algorithms that can be used out-ofthe-box (filters, feedback cancellation, compression, amplification, and also more complex auditory models, etc.). To enable online parameter tuning at runtime To open this platform on a GNU license to facilitate growths form the community To provide sufficient support, documentation and help Our goal is to enable an average MSc student to build their own take-away real-time hearing aid within a day. We demonstrate in the poster an example of a basic multi-band unilateral hearing aid on a target machine. Acknowledgements 78 The work leading to this deliverable and the results described therein has received funding from the People Programme (Marie Curie Actions) of the European Union’s Seventh Framework Programme FP7/2007-2013/ under REA grant agreement n° PITN-GA-2012-317521 50. Do hearing impaired individuals tolerate different output delay lengths than normal hearing individuals? T. Goehring, J.L. Chapman, J.J.M. Monaghan and S. Bleeck, Institute of Sound and Vibration Research, University of Southampton, UK Current hearing aids issued to hearing impaired listeners are designed to provide less than 10 milliseconds of output delay; during this time, complex digital signal processing algorithms are applied to enhance signal quality. Although increased output delay could allow for heightened signal quality and so result in increased hearing aid use and quality of life, excessive delay can be bothersome to the listener. Previous literature (Stone & Moore, 1999-2008) (Agnew & Thornton, 2000) investigating delay tolerance has focused predominantly on normal hearing participants. The present study aimed to highlight the importance of testing hearing impaired and normal hearing listeners under the same conditions using real-time processing and measuring delay tolerance for application to hearing aid use. It also aimed to investigate long-term acclimatisation to output delay. Twenty hearing impaired and twenty normal hearing participants took part and their tolerance to various delays ranging from 10 to 50 milliseconds for their own voice and an external speaker were assessed using subjective ratings. Findings revealed a significant increase in delay tolerance for hearing impaired compared to normal hearing listeners, demonstrating the importance of recruiting hearing impaired participants. An increase in tolerance was also found for increasing degrees of hearing loss, but this failed to reach significance. Although significance was not reached, an increase in tolerance was found for experienced hearing aid users compared to new users. This could indicate long-term acclimatisation to delays. These findings support a possible increase of the output delay of current hearing aid technology resulting in more sophisticated algorithms with better performance and/or lower energy consumption. Acknowledgements The work leading to this deliverable and the results described therein has received funding from the People Programme (Marie Curie Actions) of the European Union’s Seventh Framework Programme FP7/2007-2013/ under REA grant agreement n° PITN-GA-2012-317521. References Stone, M. A., & Moore, B. C. (1999-2008). Tolerable hearing aid delays. I-IV. Ear and Hearing, 20(3), 182. Agnew, J., & Thornton, J. M. (2000). Just noticeable and objectionable group delays in digital hearing aids. Journal of the American Academy of Audiology, 11(6), 330-336. 51. Android application of hearing aids on Google Glass A. Gautam, U.K. Ravuri, X. Yang, T. Goehring, S. Kim and S.Bleeck, Institute of Sound and Vibration Research, University of Southampton, UK. 79 Google Glass is a high tech hands-free augmented reality computer developed by Google with an aim to replace existing computers and change the future of assistive technology. It clubs the real and the virtual world within a single frame. A user can interact with the device through Google cloud server, which is accessible through Wi-Fi. The gadget uses Bluetooth to communicate with other external devices. For programming, Google provides the ‘Mirror Application Programming Interface (API)’ to facilitate application development which is implemented using Java. Sound is played back to the user via a bone conductor via the temporal bone. Our objective is to investigate the usability of the glasses as an android-based hearing aid. The envisaged application will interact with the sound processor to perform the functions of amplification, compression and noise reduction as implemented in state-of-art hearing devices. Bone conduction audio technology improves the feedback cancellation performance to a high extent. The voice control feature is particularly useful for controlling hearing aid for setting parameters like noise reduction, volume variation etc. We aim to produce a system that will be useful not only for mild and moderate hearing impaired users and also for normal-hearing people. On our poster we will explore the potential of Google Glass as a hearing aid, thereby providing overview on technical issues like latency, processing capabilities, frequency range, audio quality etc. We will also demonstrate a demo version of the glasses hearing aid. Acknowledgements The work leading to this deliverable and the results described therein has received funding from the People Programme (Marie Curie Actions) of the European Union’s Seventh Framework Programme FP7/2007-2013/ under REA grant agreement n° PITN-GA-2012-317521. 52. MicUp! A proposed head-controlled multichannel wireless system for hearing aids W.O. Brimijoin* and A.W. Archer-Boyd§*, *MRC/CSO Institute of Hearing Research, Scottish Section, Glasgow, UK, §Institute of Communication Acoustics, Ruhr-Universität Bochum, Bochum, Germany Hearing aids with directional microphones and noise reduction programs cannot always provide sufficient speech intelligibility in reverberant and noisy environments. As an alternative means of increasing SNR, several hearing aid manufacturers offer additional hand-held microphones that can be linked wirelessly to the hearing aid. When the microphone is held close to the talker’s mouth, it can transmit a low-noise signal directly to the hearing impaired listener. Providing a “clean” signal directly to the ear has been shown to provide a significant intelligibility benefit (e.g., Whitmer et al, 2011). The primary issue with this solution, however, is that while it can be used to improve the intelligibility for a single talker close to the microphone, many typical listening situations involve group conversations, where the focus shifts rapidly from one person to another. These situations demand that the microphone be passed from talker to talker during a conversation, a requirement that is often impractical. Here we propose an interactive multichannel wireless system to address the above issue. In this system each talker is given a special badge (see Hart et al, 2009) comprised of a microphone, an infrared LED, a short-range FM transmitter, and a battery. Each badge’s LED flashes at a unique rate and transmits the microphone signal continuously at a predefined FM frequency. The listener uses hearing aids mounted with FM receivers for each of the badges, an infrared micro-camera, and a low-power processor. The infrared camera identifies each badge by its flashing rate and uses its position in space to dynamically adjust the overall level (highest for signals in front) and – using anechoic binaural impulse responses – the apparent lateral location of each of the wireless signals. 80 This system allows the listener’s natural head orientation to function as a source selector and level controller. The use of impulse responses to place the signals at the correct angles in space provides auditory localization cues to the listener, potentially further improving both speech intelligibility and making it simpler to locate and orient towards the next talker in a multi-talker situation. We have built two headphone prototypes of the system, one integrated using an Arduino microcontroller and FM receiver, and one using real time spatialization on a laptop. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Whitmer, W.M., Brennan-Jones C.G., & Akeroyd M.A. 2011. The speech intelligibility benefit of a unilateral wireless system for hearing-impaired adults. Int J Audiol, 50.12: 905-911. Hart, J., Onceanu, D., Sohn, C., Wightman, D., & Vertegaal, R. 2009. The attentive hearing aid: Eye selection of auditory sources for hearing impaired users. In: Human-Computer Interaction, Springer Berlin Heidelberg, pp. 19-35. 53. The use of the Analytical Hierarchy Process to evaluate an algorithm designed to reduce the negative effects of reverberation N.S. Hockley, C. Lesimple and M. Kuriger, Bernafon AG, Bern, Switzerland When someone is speaking in a room, the sound field consists of direct sound originating from the speaker and indirect sound in the form of reverberation. The intensity of the speaker’s sound decreases with increasing distance, however, the reverberation is assumed to be uniform in intensity throughout the room. When the speaker is close to the listener, the direct sound will be more intense than the reverberation. As the speaker moves away from the listener, the intensity of the direct sound will decrease, however, the intensity of the reverberation will remain constant. Reverberation can have detrimental effects for the hearing-impaired individual (e.g., Nabelak & Robinette, 1977) and also challenges the directional performance of a hearing aid. These negative effects could be addressed by a dedicated algorithm (Walden et al, 2004). The challenge when designing such an algorithm is to remove the potentially negative effects of reverberation while maintaining sound quality and speech intelligibility. Two questions were addressed in this investigation. The first is, does the algorithm provide a perceptible attenuation of reverberation? The second is, does the algorithm have a noticeable effect on the quality of speech within a reverberant environment? Eleven normal hearing subjects participated in this investigation. The methodology used was based on the Analytical Hierarchy Process (AHP) (Saaty, 2008). The AHP is a structured technique for organizing and analyzing complex decision-making. It provides an analytic measure of multidimensional scaling and is based on paired comparisons that are embedded within a reciprocal matrix. Paired-comparison ratings were used to measure the amount of reverberation, speech naturalness, and overall preference. These criteria were derived from recommendations from the International Telecommunication Union (ITU, 1996; 2003). ITU (2003) specifies the methodology to evaluate speech communication systems that include noise suppression algorithms. As a high level of reverberation can be considered to be a useless signal, like noise, a reverberation reduction algorithm might have some similar behaviour as a noise suppression algorithm. The reverberation reduction algorithm was tested in a reverberant condition using four different settings (strengths) of the algorithm and unprocessed (off). In all of the paired comparisons, the processed conditions were preferred regardless of the settings. 81 The subjects responded consistently and additionally there was a high degree of consensus between the subjects. It was found that the algorithm provided a clear and constant reduction of the amount of reverberation. With a stronger setting, this difference was even more noticeable. Speech naturalness was found to be better with the algorithm than without, regardless of the tested setting. The implication of these findings for further investigations will be discussed along with the potential feasibility of using this algorithm within a hearing aid. Acknowledgements: The authors would like to thank Matthias Bertram for his help in developing the prototype and Bernhard Künzle for his useful comments and guidance. The authors are all employees of Bernafon AG. The authors alone are responsible for the content and writing of the paper. References International Telecommunications Union (ITU-T) (1996). Recommendation P.800: Methods for subjective determination of transmission quality. Geneva Switzerland. International Telecommunications Union (ITU-T) (2003). Recommendation P.835: Subjective test methodology for evaluating speech communication systems that include noise suppression algorithm. Geneva, Switzerland. Nabelek A. K. & Robinette L. (1978). Influence of the precedence effect on word identification by normally hearing and hearing-impaired subjects. J Acoust Soc Am, 63(1), 187-194. Saaty T. L. (2008). Decision making with the analytic hierarchy process. International J Serv Sci, 1(1), 83 - 98. Walden B. E., Surr R. K., Cord M. T., & Dyrlund O. (2004). Predicting hearing aid microphone preference in everyday listening. J Am Acad Audiol, 15(5), 365-396. 54. Validation of the Spatial Fixed-SNR (SFS) test in anechoic and reverberant conditions N.S. Jensen*, S. Laugesen*, F.M. Rønne*, C.S. Simonsen§ and J.C.B. Kijne§, *Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark, §Oticon A/S, Smørum, Denmark Adaptive speech-intelligibility tests are widely used in the assessment of hearing aids. Typically, the administration of such tests is easy and the results are reliable. However, one problem associated with adaptive tests is the fact the different listeners are tested at different signal-to-noise ratios (SNRs), introducing a possible confounding effect of the SNR on the outcome of the test (Bernstein, 2011). Furthermore, the test SNRs are often substantially lower than the typical daily-life SNRs (Smeds et al, 2012), for which the hearing aids under test were designed. In order to address these problems, the SFS test was developed. It is based on a paradigm where some of the parameters in a spatial speech-on-speech test are set individually in order to ‘manipulate’ all listeners in a given study towards the same speech reception threshold (SRT). This will enable a comparison of hearing aids at the corresponding fixed SNR without encountering floor or ceiling effects. Based on findings from a previous experiment (Rønne et al, 2013), the variable test parameters were scoring method, masker gender, and the target-masker spatial separation. The target speech material was HINT sentences, while the masker speech material was recordings of running speech. In two separate studies, the Danish SFS test was implemented and validated in an anechoic chamber and in a sound-treated listening room, respectively. N=26 and N=19 hearing-impaired test subjects took part in the two studies. Whereas the two test protocols differed on some aspects, e.g. by including two different experimental hearing-aid contrasts, both studies investigated the ability of the SFS test to bring listeners to perform at a given SNR, and both studies assessed the test-retest reliability. 82 The results indicated that the ‘SRT manipulations’ were reliable and similar in the anechoic and reverberant test conditions. In both conditions, the test-retest standard deviation was estimated to be around 8.5%. Better overall test performance was observed in the anechoic chamber, probably due to the reverberation present in the listening room. In conclusion, the SFS test is a reliable method to compare different hearing aids at a given fixed SNR. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Bernstein J.G.W. 2011. Controlling signal-to-noise ratio effects in the measurement of speech intelligibility in fluctuating maskers. Proc 3rd ISAAR, Nyborg, Denmark, 33-44. Rønne F.M., Laugesen S., Jensen N.S., Hietkamp R.K. & Pedersen J.H. 2013. Magnitude of speechreception-threshold manipulators for a spatial speech-in-speech test that takes signal-to-noise ratio confounds and ecological validity into account. Proc 21st ICA, Montreal, Canada, 19, 1-8. Smeds K., Wolters F. & Rung M. 2012. Estimation of Realistic Signal-to-Noise Ratios. Poster presented at the International Hearing Aid Conference (IHCON), Lake Tahoe, USA. 55. Identifying an ideal alternative television audio mix for hearing-impaired listeners R.E. Davis*§, D.F. Clark*, W.M. Whitmer§ and W.O. Brimijoin§, *Department of Computing, University of the West of Scotland, Paisley, UK, §MRC/CSO Institute of Hearing Research, Scottish Section, Glasgow, UK Currently, television (TV) viewers cannot adjust the ratio between the speech and background sound of TV broadcasts, only the overall level, which contributes little to speech intelligibility. However, digital broadcast services such as the British Broadcasting Corporation (BBC) Red Button and iPlayer potentially provide a means to simultaneously broadcast different audio versions of TV programmes. If “high intelligibility” audio were broadcast alongside standard audio, viewers could use a button on their TV remotes to swap between them. It was the intent of this study to determine the ideal speech-to-background ratio for such a transmission. The experiment consisted of two main parts, both of which used BBC TV programmes. For part one, subjects were shown short video clips with pre-set sound mixes, and asked to rate the audio on a subjective scale. For part two, subjects were shown a single long clip and asked to adjust the levels for the music and effects tracks separately until they were satisfied by the mix. Both parts of the experiment incorporated a range of -6 to +30 decibel (dB) speech-to-background ratio, with 0 dB providing the broadcast reference mix. The subject group consisted of 10 normal-hearing listeners (6 female) with a median age of 23, and 21 hearing-impaired listeners (12 female) with a median age of 63. The results from part one indicated that normal-hearing listeners rated a mix of +12 dB most highly. Hearing-impaired listeners tended to increase their ratings as a function of speech-tobackground ratio, but their ratings for mixes above +12 dB did not increase further. The results from part two indicated that both normal-hearing and hearing-impaired listeners set background levels between -12 and -18 dB below the reference mix, which equates to speech-to-background ratios of +12 and +18 dB. These findings together suggest that an optimal alternative mix would consist of a 12 dB increase in speech-to-background ratio. It should be noted, however, that the results from this experiment may be dependent on the programme material used. Our current testing incorporates a larger range of source material, more natural testing conditions, and trial broadcasts of listenerswitchable audio by BBC Scotland. 83 Acknowledgements The authors thank the BSA Gatehouse Fund for funding this study, the MRC (U135097131), the Chief Scientist Office (Edinburgh), the University of the West of Scotland, and Jim Hunter and Andrew Britton at BBC Scotland. 56. Assessing the impact of hearing impairment on immediate speech intelligibility as talkers change. J. Holman, A. MacPherson, M.A. Akeroyd, MRC/CSO Institute of Hearing Research - Scottish Section, Glasgow, United Kingdom. It is well known that hearing-impaired people report difficulties in multi-talker conversations. For instance, data collected in our lab over a decade shows that the relevant SSQ question (“You are with a group and the conversation switches from one person to another. Can you easily follow the conversation without missing the start of what each new speaker is saying?”) gives one of the lower scores of the speech section. Few studies, however, have used dynamic listening situations to explore experimentally talker changes in conversations. We developed a variation of a new continuous speech test, the Glasgow Monitoring of Uninterrupted Speech Test (GMUST; MacPherson and Akeroyd 2014), to measure an important aspect of the problem: does speech identification drop immediately after a change in talker? The GMUST uses a long (here, 6-min) audio passage of continuous prose together with a text transcript in which some single-word mis-transcripts have been deliberately made. The task is to spot the mismatch between the audio and the text. Here, we edited the audio to change from male to female to male and so on, varying every few seconds, while keeping the flow of the prose. Thirty-six mistranscripts were made, twelve being within 1 second of the change of talker (“early”), twelve between 1-3 seconds later (“medium”), and twelve between 3-9 seconds later (“late”). The hypothesis was that fewer mismatches would be noticed in the early condition than in the late condition. In a control condition there was a single change at 3 mins, with the 36 mismatches placed at arbitrary times unrelated to the change. All the audio was presented along with 4 talker babble as a masker (the SNR was set individually on the basis of a short pre-test; here, they ranged from -2 to +10 dB). At present, 25 hearing-impaired listeners have participated (BEA >= 25 dB HL; aged = 45-74 years). Five listeners did particularly poorly or badly and were excluded. The remainder found, on average, about 20 of the mis-transcripts. There were small differences in performance across early vs. late in the hypothesized direction. But we discount these effects, as there were differences of the same magnitude in the control condition, so overall the experiment did not support the hypothesis. Various potential explanations will be explored. Acknowledgements Supported by intramural funding from the Medical Research Council (grant number U135097131) and the Chief Scientist Office of the Scottish Government. References MacPherson, A. & Akeroyd, M.A. 2014. A method for measuring the intelligibility of uninterrupted, continuous speech. J Acoust Soc Am, 135, 1027-1030 84 57. Comparison of different forms of frequency lowering in different forms of frequency lowering in digital hearing aids for people with dead regions in the cochlea M. Salorio-Corbetto, T. Baer, and B.C.J. Moore, University of Cambridge, Department of Psychology, Downing Street, CB2 3EB, Cambridge, UK. When high-frequency hearing loss is associated with extensive dead regions (DRs) in the cochlea, i.e. regions of the basilar membrane with non-functioning or very poorly functioning inner hair cells and/or neurons, high-frequency amplification via a hearing aid often fails to provide benefit for speech intelligibility. Frequency-lowering hearing aids, which transform high-frequency sounds into lower-frequency sounds, may be used to provide information from high frequencies to people with high-frequency DRs. However, there have been few studies of the effectiveness of frequencylowering hearing aids for people with confirmed extensive DRs. Here, two frequency-lowering prototypes based on a commercially available digital hearing aid (Phonak Exélia Art P) were evaluated. One of the prototype hearing aids used frequency transposition, i.e. frequency components in a high-frequency band were recoded by decreasing the frequency of each by a constant amount in Hz. The other prototype hearing aid used frequency compression, i.e. the downward frequency shift of components in a high-frequency band increased with increasing frequency above the lower edge of the band. Ten subjects with steeply-sloping hearing losses and extensive DRs took part in a single-blind, three-period, three-condition cross-over design trial whose aim was to evaluate the effectiveness of frequency-lowering hearing aids for speech perception. Conditions were: (1) Conventional amplification over the widest frequency range possible with no frequency lowering; (2) Frequency transposition; (3) Frequency compression. The edge frequency (fe) of the DRs ranged between 0.5 and 2 kHz, and was below 0.75 kHz for five of the subjects, and below 1 kHz for seven of the subjects. The two frequency-lowering schemes were fitted taking into account the characteristics of the DRs of the subjects, which led to frequency lowering at frequencies below those used in previous research and in current clinical practice. Outcome measures were: 1) Identification of consonants in nonsense vowel-consonant-vowel (VCV) combinations; 2) Detection of word-final [s] and [z]; 3) Speech reception threshold in noise (SRT); 4) Responses on the Speech, Spatial and Qualities Questionnaire (SSQ). Frequency lowering did not provide a significant benefit for any of the outcome measures relative to the control condition. However, the use of frequency lowering may provide some practical advantages, such as reduced problems with acoustic feedback and lower output level requirements. Acknowledgements: MSC was supported by Action on Hearing Loss and Phonak. TB and BCJM were supported by the Medical Research Council (MRC). 58. Simulating speech understanding in noise for hearing impaired listeners J.J.M. Monaghan, M.N. Alfakhri and S. Bleeck, ISVR, University of Southampton, Southampton, UK Understanding speech in a noisy background is one of the greatest challenges for hearing-impaired (HI) listeners and hearing aids can provide only limited benefit. As well as elevated thresholds for hearing, supra-threshold deficits such as loudness recruitment and reduced frequency selectivity and temporal resolution are thought to contribute to the difficulties experienced by HI-listeners in complex listening environments. Recent experiments in our laboratory, however, suggest that the low intelligibility of speech in noise observed in moderate-to-severely HI listeners is not evident in simulations of the reduced frequency selectivity, threshold elevation and recruitment observed in HI listeners. The aim of the current study, therefore, is to produce a better simulation of speech understanding in noise for HI listeners. To this end, in addition to employing spectral smearing to simulate reduced frequency selectivity (e.g., Baer and Moore, 1993), the effects of deafferentation 85 (Lopez-Poveda and Barrios, 2013) and of broadening the temporal integration window were also simulated. Appropriate parameters for the different simulation conditions were determined by measuring the masking release obtained by adding either a temporal or spectral notch when detecting a tone in a band of noise. Data for normal hearing (NH) participants listening through the simulation were compared with previously measured data obtained in HI listeners (van Esch et al., 2013). Speech recognition scores with BKB sentences were measured for HI listeners and for NH listeners listening under simulations of hearing impairment. To determine whether simulation can be used to predict any benefit provided by speech enhancement algorithms, scores were also measured for noisy speech processed by an algorithm known to improve intelligibility of speech in noise for HI listeners but not for NH listeners. Results will be presented at the meeting. Acknowledgements Work supported by EPSRC grant EP/K020501/1. References Baer, T., and Moore, B. C. J. (1993). "Effects of spectral smearing on the intelligibility of sentences in noise," J. Acoust. Soc. Am. 94, 1229-1241. Lopez-Poveda, E. A., and Barrios, P. (2013). "Perception of stochastically undersampled sound waveforms: A model of auditory deafferentation," Frontiers in Neuroscience 7. Van Esch T. E. M., Kollmeier, B., Vormann, M., Lyzenga, J., Houtgast, T., Hallgren, M., Larsby, B., Athalye, S. P., Lutman, M. E., Dreshchler, W. A. (2013) “Evaluation of the preliminary auditory profile test battery in an international multi-centre study” International Journal of Audiology 2013; 52: 305–321 59. Development of a patient decision aid for adults with hearing loss A.J.Hall*, H. Pryce§, *St Michael’s Hospital, Bristol, UK, §Sirona Health and Social Care, Bath, UK, Shared decision making (SDM) is a process in which patients and clinicians share information and plan clinical care together based on evidence and patient preference (Edwards and Elwyn, 2009). Decision aids are tools to support SDM (Stacey et al, 2011). They present information to patients (e.g. in a pamphlet) on the range of options available giving answers to the most frequently asked questions around options and their implications. The use of SDM and decision tools in clinical practice is promoted within the NHS and is seen as a marker of clinical quality (NICE, 2012). This study aims to develop a patient decision aid for adults with hearing loss in the UK. The decision aid is being developed in the Option Grid format (www.optiongrid.org). Each Option Grid provides key facts to frequently asked questions to support patients and clinicians in deliberating about options. All Grids are evidence-based, aligned, where possible, with NICE guidelines, and updated yearly. They do not attempt to be comprehensive and supplement existing information and support. Patient and clinicians’ input is central to the Option Grid development process and there is a standard procedure for developing Options Grids. We are following the Option Grid Development Framework and iterative development process; the Editorial Team includes six stakeholders, including clinical experts, researchers and one patient representative. In the first phase of the project we have run three Focus groups with patients with hearing loss to determine their decisional needs; also ethnographic observations of five audiology appointments and the interactions between audiologists and patients offered hearing aids to examine current practice and informational need. This research was conducted with patients from Sirona Care and Health Audiology/Hearing Therapy department, Bath. 86 Using the qualitative data from the first phase, the Editorial Team has developed the first version of the Option Grid ready for user testing. Acknowledgements Supported by the British Society of Audiology Research Fund. References NICE Quality Standard 15. 2012. Quality standard for patient experience in adult NHS services. Edwards A. & Elwyn G. 2009. “Shared decision-making in health care: Achieving evidence-based patient choice”. In. Edwards and Elwyn (eds) 2009. Shared decision-making in health care: achieving evidence-based patient choice. 2nd ed. Oxford: Oxford University Press. Stacey D., Bennett C.L., Barry M.J., Col N.F., Eden K.B., Holmes-Rovner M., Llewellyn-Thomas H., Lyddiatt A., Légaré F. & Thomson R. 2011. Decision aids for people facing health treatment or screening decisions. Cochrane Database Syst Rev 5;(10):CD001431. 60. Multimedia interactive reusable learning objects improve hearing aid knowledge, practical handling skills and use in first-time hearing aid users M. Ferguson*, M. Brandreth*, P. Leighton§, W. Brassington‡, H. Wharrad§ *NIHR Nottingham Hearing Biomedical Research Unit, §University of Nottingham, ‡Nottingham Audiology Services, UK There is currently a gap in audiology rehabilitation for an effective intervention to enhance knowledge and educate hearing aid users. Interactive multimedia video tutorials (or re-usable learning objects, RLOs) can enhance learning and motivation in educational and healthcare contexts. To address this in first-time hearing aid users, eight short RLOs (5-10 mins; total duration 1 hour) were developed using a participatory design that included (i) a delphi review of 33 UK audiology/hearing experts, and (ii) 35 hearing aid users. The RLOs were evaluated in an RCT of 203 first-time hearing aid users, where 100 received the RLOs at the fitting appointment (RLO+) and 103 formed a waitlist control group (RLO-). Outcome measures were obtained 6-weeks post-fitting (RLO+, n=79; RLO-, n=88). RLO uptake and compliance was high (78% and 94% respectively). Half the users watched the RLOs at least twice, and 20% watched the RLOs 3+ times. This suggests that users were using the RLOs to selfmanage their hearing loss and hearing aids. The most prevalent delivery medium was DVD for TV (50%), with a surprisingly high number using RLOs through the internet (30%). The RLO+ group had significantly better knowledge of both practical hearing aid and psychosocial issues (p<.001), and significantly better practical hearing aid skills (PHAST: p=.002) than the RLO- group There was no significant difference in overall hearing aid use or benefit using the Glasgow Hearing Aid Benefit Profile. However, there was significantly greater use in the RLO+ group for difficult listening situations (p=.013), and for suboptimal users (use<70%; p=.018) with a large effect size (d=.88). The RLOs were rated as highly useful (9.0/10), and the majority (>80%) agreed the RLOs were enjoyable, improved their confidence, and were preferable to written information. 78% would recommend the RLOs to others. These results were supported by discussion groups of hearing aid users and family members. In addition, it was reported that RLOs were shared amongst other people with hearing loss, and were watched by family members and friends. Thus, the RLOs were beneficial to hearing aid users across a range of quantitatve and qualitative measures. The RLOs have since undergone further development work based on feedback from from RCT and the discussion group participants. The C2Hear RLOs will be commercially available for clinical use and wider distribution from autumn 2014. 87 Acknowledgements This research was funded by the NIHR Research for Patient Benefit scheme. 61. An investigation into the ability to detect and discriminate speech phonemes using nonlinear frequency compression hearing aids in a paediatric population S. Harman and S. Morgan, Audiology, Bradford Royal Infirmary, Bradford, UK This study aims to look at the ability of hearing impaired children to complete the Auditory Speech Stimuli Evaluation (ASSE) test using non-linear frequency compression (NLFC) hearing aids. Recent guidance (Feirn et al, 2014) recommends the use of NLFC hearing aids for children with moderate as well as those with severe or profound hearing losses due to reported evidence predicting positive benefit. It is recommended that NLFC fittings are verified using speech material. Previous research has used the plurals test, consonant confusion testing, and sentence testing, amongst others for verification. ASSE was used to assess NLFC fittings using detection of eight phonemes /a/, /u/, /i/, /m/, /v/, /z/, /s/, /sh/ presented at 65dHBL, and discrimination between seven pairs /i/-/a/, /u/-/a/, /u/-/i/, /z//s/, /m/-/z/, /s/-/sh/, /v/-/z/ presented at 65dBHL with a roving +/-3dB to remove intensity cues. Twenty-nine children with stable sensori-neural hearing losses, who were long-term users of NLFC aids bilaterally, were seen on their routine hearing aid review appointments. Prior to ASSE testing, hearing aids were verified that they were fitted optimally and working within technical specification. All subjects completed both detection and discrimination lists. Eighteen children scored 100% on detection and discrimination. The remaining eleven children could not discriminate between at least one phoneme pair. Of those eleven children nine children could not detect /s/ and of these, four children were unable to detect multiple phonemes. When the patients were grouped by hearing level it was found that all eighteen children who completed ASSE correctly had average hearing losses of mild or moderate level in at least one ear. The eleven children who could not complete ASSE testing had average hearing losses of severe or profound level bilaterally. The significance of these results on a child’s outcomes will require further investigation Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Feirn R., Wood S., Sutton G., Booth R., Meredith R., Brennan S. & Lightfoot G. 2014. Guidelines for Fitting Hearing Aids to Young Infants. MCHAS. 62. Using music therapy and dichotic training to develop listening skills in some children with auditory processing disorders A. Osisanya and A. Adewunmi, Department of Special Education, University of Ibadan- Nigeria Prevalent evidence of poor listening ability among the school- aged population due to manifestation of Auditory Processing Disorders (APD) has become a matter of concern as the disorder places a negative impact on the academic achievement and psycho- social development of persons 88 presenting with it. This has evidently warranted the necessity of appropriate management strategies such as music therapy and dichotic training, both aimed at improving an individual’s listening abilities (BSA APD SIG, 2011). 35 school children, ages between 8 and11 years old, screened down to 15 with normal puretone thresholds (< 25dB hearing level at 1- 4kHz) with no evidence of glue ear and with average cognitive ability (90- 109 TSS on SIT- R3) were diagnosed with APD using a 2.S.D. below the mean from the SCAN C and TALC- R, coupled with abnormal scores on the RGDT, and supported by scores on the CCC- 2 and CHAPPS; were randomly assigned to three groups of music therapy, dichotic listening and both; and exposed to 8 weeks training of 1 hour, 4 times a week session, using live and recorded music/ nursery rhymes, phonetically balanced words from CID W- 22 Test List, and Words/Sentence list developed by the researchers. The 15 participants were pre and post tested on the Words/ Sentence list. Our results showed a significant difference in the pre and post test scores of the music group (t= 7.060; p< 0.05), of the dichotic group (t= 9.791; p<0.05), and of the combined group (t=11.326; p<0.05). Our post- hoc analysis showed that the combined therapy group had the highest adjusted post- test score ( x = 75.398); followed by the dichotic group ( x = 61.519), and the music group ( x =58.283). Further, we found that the independent variables jointly accounted for 61.8% of the variance in the therapy among the participants. Therefore, we concluded that the peculiarity of APD does not warrant a single management protocol, but a combination of several management strategies that involves active participation of the individual presenting with the disorder. ‘Declaration of interest: the authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper.’ References BSA APD SIG. (2011). Practical Guidance: An overview of current management of auditory processing disorder. 1-60 63. Effects of different masker types on speech perception in children and adults H. Holmes, D. Rowan and H.E. Cullington. Institute of Sound and Vibration Research, University of Southampton, UK Normal-hearing adults show improved speech perception with speech maskers over stationary noise maskers, showing a release from masking. Some studies with children have shown a smaller masking release compared to adults (Hall et al., 2002), with large inter-child differences, although there is conflicting literature (Litovsky, 2005). Conflicts may be due to differences in speech masking materials used and it is thought that children may be more susceptible than adults to ‘informational masking’. Adaptive methods used in measuring speech perception result in different tested signal-to-noise-ratios (SNRs). Bernstein and Brungart (2011) propose that the amount of masking release depends on the SNRs used, where high SNRs yield low masking release and vice versa. Therefore, as children generally need higher SNRs than adults in stationary maskers they may also show a smaller masking release, which could be due to differences in the baseline conditions, leading to an SNR confound which may be misinterpreted. This study aims to: 1) Determine differences in performance between adults and children with stationary noise and speech maskers. 2) Investigate whether expected smaller masking release in children compared to adults is due to an SNR confound (Bernstein & Brungart, 2011). Participants were required to listen to and identify target sentences (BKB and IHR sentences) in two different masker types: speech shaped noise and a speech masker. The targets and maskers were presented at five fixed signal-to-noise-ratios for each masker type to determine the percentage of sentences scored correctly in order to estimate the shapes of the psychometric functions. Fifteen 89 sentences were presented for each condition and a repeat was carried out on a separate day. Data collection from children (aged 7-8 years) and adults is currently underway and will be presented. Acknowledgements Supported by a scholarship from the University of Southampton. References Bernstein, J. G. W. & Brungart, D. S. 2011. Effects of spectral smearing and temporal fine structure distortion on the fluctuating-masker benefit for speech at a fixed signal-to-noise ratio. J Acoust Soc Am, 130, 473-488. Hall III, J. W., Grose, J. H., Buss, E. & Dev, M. B. 2002. Spondee recognition in a two talker masker and a speech-shaped noise masker in adults and children. Ear Hear, 23, 159-165. Litovsky, R. Y. 2005. Speech intelligibility and spatial release from masking in young children. J Acoust Soc Am, 117, 3091-3099. 64. Listening and reading in background noise: children versus adults H. Holmes, D. Rowan and H.E. Cullington, ISVR, University of Southampton, UK Every day we are faced with the task of parsing incoming simultaneous signal components to decode meaning and supress interfering sounds when listening and reading, invoking both ‘bottomup’ and ‘top-down’ processes. Many listening studies have found a smaller masking release in children compared to adults when a stationary-noise background is switched to a speech background (e.g. Hall et al., 2002) and such results are understood to reflect maturation of ‘topdown’ processes in children. Bernstein and Brungart (2011) however propose, the amount of masking release may be subject to a signal-to-noise-ratio (SNR) confound so previous findings could be misinterpreted. The first aim of this study was thus to: 1) Determine speech perception performance and investigate whether expected smaller masking release in children compared to adults is due to a SNR confound (Bernstein & Brungart, 2011). Twenty normal hearing children (7-8 years) and twenty adults were required to listen to and identify BKB and IHR target sentences in stationary-noise and speech backgrounds. Percent-correct was measured for fifteen sentences presented at five fixed SNRs for each background to estimate psychometric function shape. A repeat was carried out on a separate day. Masking release was smaller in children compared to adults, but child/adult differences lessened when baseline stationary-noise differences were taken into account. Although reduced, a child/adult difference remained suggesting fundamental differences in the effect speech maskers have on listening tasks in children compared to adults. Such results could suggest children may have difficulties listening in noisy classrooms. Since speech backgrounds may affect ‘top-down’ processes in children, it is of interest to determine how such backgrounds affect these processes only, particularly to tap into the cognitive demands arising from noisy conditions. In order to do this, a reading paradigm was utilised to track eye movements during reading in the presence of different background sounds. The second aim of this study was thus to: 2) Determine which backgrounds are most disruptive to a reading task to explore ‘top-down’ processes and investigate child/adult differences. Data collection from children (8-9 years) and adults is currently underway and results from an eye tracking experiment be presented. 90 Acknowledgements Supported by a Scholarship from the University of Southampton. References Bernstein, J.G.W., Brungart, D.S. (2011). Effects of spectral smearing and temporal fine-structure distortion on the fluctuating-masker benefit for speech at a fixed signal-to-noise ratio. J. Acoust. Soc. Am., 130, 473–88. Hall, J.W., Grose, J.H., Buss, E., et al. (2002). Spondee Recognition in a Two-Talker Masker and a Speech-Shaped Noise Masker in Adults and Children. Ear Hear., 159–165. 65. Using the ECLiPS questionnaire to assess the everyday listening difficulties of children with Otitis Media with Effusion J.G. Barry*, M.L. Young*, M. Daniel§ and J. Birchall§, *MRC Institute of Hearing Research, Nottingham, UK, §Nottingham Hearing Biomedical Research Unit, Nottingham, UK Otitis media with effusion (OME) is known to reduce overall hearing level, but there has been little focus on how it affects hearing and listening in everyday situations. At present, access to interventions is determined by the NICE guidelines, which focus on pure tone audiograms but suggest no standardised or quantifiable means of measuring the impact of OME on a child’s quality of life. The Evaluation of Children’s Listening and Processing Skills questionnaire (ECLiPS) has recently been launched as a method of assessing children’s listening abilities and highlighting a range of potential problems. The scale comprises five factors which examine symptoms of listening, language, social, and cognitive difficulties. It was initially designed to assess children referred for suspected auditory processing disorder. This study investigates: (a) The sensitivity of the ECLiPS to the particular listening difficulties associated with OME. (b) The extent to which it is sensitive to differences in severity of OME. The ECLiPS questionnaire was completed by parents of children aged 3-11 in three groups: ‘severe OME’ recommended for surgery (N=9); ‘mild OME’ with no surgery planned (N=7); and a control group of age- and gender-matched healthy children (N=16). Children with OME showed an opposite pattern of responses to controls on most items, with higher ratings (indicating greater difficulty) for children with severe OME than those with mild OME. ANOVAs revealed significant differences between all groups on the listening and language factors and to a lesser extent, on the factor assessing social interaction skills. Differences were found between OME children and controls on cognitive abilities (‘Memory and Attention’), but there was no effect for severity between mild and severe OME. The results indicate that the ECLiPS is sensitive to listening difficulties of children with OME. Moreover, it also appears to be sensitive to the severity of the condition, particularly in the case of the factors assessing listening and language abilities. The ECLiPS may be useful as a tool for assessing the impact of this condition on a child’s global functioning and could potentially supplement pure tone audiograms to aid clinical decision-making. Acknowledgements Supported by the MRC and Nottingham University Hospitals. 91 References Barry, J.G. (2013). Measuring listening difficulties using parental report. Proceedings of British Society of Audiology annual conference, Keele UK. National Institute for Health and Care Excellence 2008 Surgical management of OME. CG60. London: National Institute for Health and Care Excellence. 66. Attending to narrative and detecting mispronunciations with Auditory Processing Disorder (APD) H. Roebuck and J.G. Barry, MRC Institute of Hearing Research, University Park, Nottingham, UK. Recent research has suggested that Auditory Processing Disorder (APD) may reflect an underlying deficit in an ability to maintain attention (Moore et al, 2010). This suggestion was partly based on evidence of fluctuating threshold levels during psychophysical assessment of auditory processing. Tests of auditory processing are inherently boring and bear little relationship to everyday listening. In this study, we assess the hypothesis that APD is effectively a disorder of sustained attention using a story task designed to mimic everyday listening. The current sample consists of 27 children (8-10 years) with and without a referral for APD, or listening difficulties identified using the ECLIPS questionnaire (Barry and Moore, 2014) (n=13/14). The participants were asked to press a button when they heard a mispronounced word. To assess the role of language, mispronounced words were presented in both predictable and unpredictable contexts. Previous research has shown that children and adults are faster and more accurate at detecting mispronounced words presented in predictable contexts (Cole & Perfetti, 1980). If sustaining attention rather than underlying language difficulty is the primary difficulty for participants with APD we would predict a similar benefit for mispronunciations in predictable contexts, but fewer words would be detected. If however, underlying language deficits impact task performance, we would expect the difference in word context to disappear. Preliminary data suggests the listening difficulty group missed more mispronunciations (15.4 ± 1.9) than their typically developing counterparts (5 ± 1.8, p<0.001). Secondly, both groups detected more words in a predictable context (p=0.003), and identified these words more quickly than in the unpredictable context (p=0.001). Taken together, this suggests that the listening difficulty group derived some benefit from sentence context when detecting mispronunciations, but their attention could not be maintained throughout the duration of the task. The results appear to support the idea that problems with sustained attention contribute to listening difficulties during every day listening tasks, and are not limited to more artificial tasks designed to assess auditory processing. Acknowledgements Supported by the Medical Research Council and Nottingham University Hospitals NHS Trust. References Moore, D.R., Ferguson, M.A., Edmondson-Jones, A.M., Ratib, S., & Riley, A. (2010). The nature of auditory processing disorder in children. Pediatrics, 126, e382-390. Cole, R.A., & Perfetti, C.A. (1980). Listening for mispronunciations in a children's story: The use of context by children and adults. J Verb Learn Verb Beh, 19, 297-315. Barry, J.G. and Moore, D.R. (2014). Evaluation of Children’s Listening and Processing (ECLiPS) Manual edition 1. 67. Evaluation of school entry hearing screening 92 H. Fortnum*, A. Allardice §, C. Benton‡, S. Catterick*, L. Cocking#, C. Hyde+, J. Moody^, V. Nikolaou+, M. Ozolins*, R. Taylor+, O. Ukoumunne+ and J. Watson$. *National Institute for Health Research Nottingham Hearing Biomedical Research Unit, School of Medicine, University of Nottingham, UK, §Nottinghamshire Healthcare NHS Trust, Nottingham, UK, ‡Nottingham University Hospitals NHS Trust, Nottingham, UK, #Peninsula Clinical Trials Unit, Plymouth University, UK, +University of Exeter Medical School, UK, ^Cambridge University Hospitals NHS Foundation Trust, and Cambridgeshire Community Services NHS Trust, Cambridge, UK, $Parent representative, Nottingham, UK The School Entry Hearing Screen (SES) has been around since the 1950s. With the introduction of Universal Newborn Hearing Screening in 2006, the value of the SES has been questioned as around half the cases of permanent hearing loss are now identified at birth. In 2007 a report funded by the Health Technology Assessment (HTA) Programme1 described a lack of good evidence of the effectiveness of the SES. This project evaluates its diagnostic accuracy and cost-effectiveness. We evaluated the diagnostic accuracy (sensitivity and specificity) of two screening tests; a pure-tone sweep test, and the Siemens HearCheckTM device, via comparison with Pure Tone Audiometry (PTA) (gold standard). Children with sensorineural or permanent conductive hearing loss bilaterally (average 20-60dBHL) or unilaterally (≥20 dBHL) were screened at home and the results compared with their latest PTA from audiological records. Children without identified hearing loss were screened at the research unit and their PTA measured in a soundbooth. To compare yield, referral age and source, and clinical pathway for children aged 3-6 years we collected data on referrals to an audiology service with a school screen (Nottingham) and one without (Cambridge). We assessed the practical implementation of the two screens by school nurses. All data will ultimately contribute to an evaluation of cost-effectiveness in an economic model. To date 60 hearing-impaired children and 134 hearing children have been tested. Preliminary data indicates little difference between the two screens for sensitivity (>83%). or specificity (>87%). School nurses have highlighted advantages and disadvantages of each of the screens with neither being overwhelmingly preferred. Details of >3000 referrals for Nottingham and Cambridge have been collected. Analyses of the difference in yield and age of referral between the two services are ongoing. Acknowledgements This project was funded by the National Institute for Health Research Health Technology Assessment Programme (project number 10/63/03). The views and opinions expressed therein are those of the authors and do not necessarily reflect those of the HTA programme, NIHR, NHS or the Department of Health. Reference Bamford J et al. 2007. Current practice, accuracy, effectiveness and cost-effectiveness of the school entry hearing screen. Health Technol Assess, 11(32), 1-168 68. Clinical feasibility and acceptability of recording cortical auditory evoked potentials: results from the first 80 infants Munro, Kevin J‡#, Nassar Ruth#, Purdy, Suzanne C*; O’Driscoll, Martin‡#, Booth, Ruth#, Bruce, Iain‡,#2; Uus Kai‡., ‡School of Psychological Sciences, University of Manchester, UK; #Central Manchester University Hospitals NHS Foundation Trust, Manchester Academic Health Science Centre, Manchester, UK; *School of Psychology, University of Auckland, New Zealand 93 There is growing interest in using supra-threshold obligatory cortical auditory evoked potentials (CAEPs) to complement established paediatric clinical test procedures. As part of an on-going study investigating the clinical usefulness of aided CAEPs in infants, we have been obtaining data on feasibility and acceptability of the procedure within the clinical setting. Responses to short duration stimuli (/m/, /g/ and /t/), will ultimately be recorded in 100 normal-hearing and 10 hearing-impaired infants (between 4 and 39 weeks of age) from a loudspeaker at zero degree azimuth and a presentation level of 65 dB SPL. At the time of writing, we have the following data from the first 80 infants (who all passed new-born screening and have no reported hearing difficulties): CAEP test duration, completion and detection rates, and a parental acceptability questionnaire (9 questions with 7-point scale, 1 being best). The mean test duration was 26 minutes (range 17-89 min): 13 min preparation and 13 min data acquisition. A response was obtained to at least one stimuli in 100% of infants. A response to all three, two and one stimuli was detected in 72%, 23% and 5% of infants, respectively. Responses to /g/, /t/ and /m/ were detected in 97%, 88% and 820% of infants, respectively. So far, 29 parents have completed the acceptability questionnaire and the mean score of individual questions range from 1.2-2.5. The poorest score was obtained for the question enquiring about the difficulty keeping the baby awake and quiet for the duration of the test procedure. In conclusion, the short test duration, high completion and detection rates and good scores on the acceptability questionnaire suggest that the sound field CAEP procedure may be feasible and acceptable for use in infants within the clinical setting. Our next study will investigate the relationship between audibility and CAEP detetection in hearing-impaired infants. Acknowledgements The study was funded by a strategic investment grant from Central Manchester University Hospitals NHS Foundation Trust and facilitated by the Manchester Biomedical Research C and the NIHR/Wellcome Clinical researech Facility. 69. The use of the Analytical Hierarchy Process to evaluate an algorithm designed to reduce the negative effects of reverberation N. S. Hockley, C. Lesimple and M. Kuriger, Bernafon AG, Bern, Switzerland When someone is speaking in a room, the sound field consists of direct sound originating from the speaker and indirect sound in the form of reverberation. The intensity of the speaker’s sound decreases with increasing distance, however, the reverberation is assumed to be uniform in intensity throughout the room. When the speaker is close to the listener, the direct sound will be more intense than the reverberation. As the speaker moves away from the listener, the intensity of the direct sound will decrease, however, the intensity of the reverberation will remain constant. Reverberation can have detrimental effects for the hearing-impaired individual (e.g., Nabelak & Robinette, 1977) and also challenges the directional performance of a hearing aid. These negative effects could be addressed by a dedicated algorithm (Walden et al, 2004). The challenge when designing such an algorithm is to remove the potentially negative effects of reverberation while maintaining sound quality and speech intelligibility. Two questions were addressed in this investigation. The first is, does the algorithm provide a perceptible attenuation of reverberation? The second is, does the algorithm have a noticeable effect on the quality of speech within a reverberant environment? Eleven normal hearing subjects participated in this investigation. The methodology used was based on the Analytical Hierarchy Process (AHP) (Saaty, 2008). The AHP is a structured technique for organizing and analyzing complex decision-making. It provides an analytic measure of multidimensional scaling and is based on paired comparisons that are embedded within a reciprocal 94 matrix. Paired-comparison ratings were used to measure the amount of reverberation, speech naturalness, and overall preference. These criteria were derived from recommendations from the International Telecommunication Union (ITU, 1996; 2003). ITU (2003) specifies the methodology to evaluate speech communication systems that include noise suppression algorithms. As a high level of reverberation can be considered to be a useless signal, like noise, a reverberation reduction algorithm might have some similar behaviour as a noise suppression algorithm. The reverberation reduction algorithm was tested in a reverberant condition using four different settings (strengths) of the algorithm and unprocessed (off). In all of the paired comparisons, the processed conditions were preferred regardless of the settings. The subjects responded consistently and additionally there was a high degree of consensus between the subjects. It was found that the algorithm provided a clear and constant reduction of the amount of reverberation. With a stronger setting, this difference was even more noticeable. Speech naturalness was found to be better with the algorithm than without, regardless of the tested setting. The implication of these findings for further investigations will be discussed along with the potential feasibility of using this algorithm within a hearing aid. Declaration of interest: The authors are all employees of Bernafon AG. The authors alone are responsible for the content and writing of the paper. Acknowledgements: The authors would like to thank Matthias Bertram for his help in developing the prototype and Bernhard Künzle for his useful comments and guidance. References International Telecommunications Union (ITU-T) (1996). Recommendation P.800: Methods for subjective determination of transmission quality. Geneva Switzerland. International Telecommunications Union (ITU-T) (2003). Recommendation P.835: Subjective test methodology for evaluating speech communication systems that include noise suppression algorithm. Geneva, Switzerland. Nabelek, A. K. & Robinette, L. (1978). Influence of the precedence effect on word identification by normally hearing and hearing-impaired subjects. Journal of the Acoustical Society of America, 63(1), 187-194. Saaty, T. L. (2008). Decision making with the analytic hierarchy process. International Journal of Services Sciences, 1(1), 83 - 98. Walden, B. E., Surr, R. K., Cord, M. T., & Dyrlund, O. (2004). Predicting hearing aid microphone preference in everyday listening. Journal of the American Academy of Audiology, 15(5), 365396. 70. The effect of primary tone ramping paradigms on the onset time-course of distortion product otoacoustic emissions (DPOAEs). B. Lineton and L. J. Macfarlane, Institute of Sound and Vibration Research, University of Southampton, UK In this study, low-side band (2f1-f2) and high-side (2f2-f1) distortion product otoacoustic emissions (DPOAEs) are studied both experimentally and using cochlear model simulations. Current theory suggests that the generation of both these DPOAEs side-bands involve a number of interacting mechanisms (Knight & Kemp, 2001; Shera & Guinan, 1999; Young et al., 2012). In this theory, the DPOAEs originate from a directional nonlinear source on the basilar membrane in the primarytravelling-wave overlap region that generates travelling waves preferentially in the forward 95 (anterograde) direction. A second component arises from reflection at the characteristic place of the DPOAE component. The travelling wave from these components then combine, giving rise to DPOAE measured in the ear canal that can have complex properties. The time-course of the DPOAE onset due to the ramping on of the primary tones is also determined by the build-up of the primary travelling waves along the basilar membrane. The time-course of the DPOAE is difficult to measure directly, because the ramping of the primary tones introduces spectral components (by “spectral splatter”) that swamp the DPOAE components. One way of overcoming this problem is to use the primary-tone-phase-variation (PTPV) method (Duifhuis, 2006; Whitehead et al., 1996). In this study, the PTPV method has been used in simulations of the DPOAE onset time-course as the two primary tones (labelled f1 and f2 as per convention) are ramped on and off. Three different primary-tone onset paradigms are studied: First f1 and f2 both ramped simultaneously; second, f1 is held on continuously and f2 is ramped on and off, and third, f2 is held on continuously while f1 is ramped on. The simulation result will be compared with PTPV measurements of the DPOAE time-course, and discussed in terms of theoretical implications. Acknowledgements This work was funded by the University of Southampton. References Duifhuis, H. 2006. Distortion Product Otoacoustic Emissions: A Time Domain Analysis. ORL, 68, 340–346. Knight, R.D. & Kemp, D.T. 2001. Wave and place fixed DPOAE maps of the human ear. J Acoust Soc Am, 109, 1513-1525. Shera, C.A. & Guinan, J.J. 1999. Evoked otoacoustic emissions arise by two fundamentally different mechanisms: A taxonomy for mammalian OAEs. J Acoust Soc Am, 105, 782–798. Whitehead, M.L., Stagner, B.B., Martin, G.K. & Lonsbury-Martin, B.L. (1996). Visualization of the onset of distortion-product otoacoustic emissions, and measurement of their latency. J Acoust Soc Am. 100, 1663–1679. Young, J.A.; Elliott, S.J. & Lineton, B. 2012. Investigating the wave-fixed and place-fixed origins of the 2f(1)-f(2) distortion product otoacoustic emission within a micromechanical cochlear model, J Acoust Soc Am, 131, 4699-4709 Poster Abstracts: Cochlear implants 71. Auditory implant team support for speech and language therapists working with children who use cochlear implants A. Odell*§, H. Fortnum* and P. Leighton‡,*Nottingham Hearing Biomedical Research Unit, Otology and Hearing group, Division of Clinical Neuroscience, University of Nottingham, Nottingham, UK, §Nottingham Auditory Implant Programme, Nottingham University Hospitals, Nottingham, UK, ‡School of Medicine, University of Nottingham, Nottingham, UK After 25 years of paediatric cochlear implantation in the UK, we have a large body of evidence showing how children can benefit from this procedure. We can use some of that evidence to inform community speech and language therapists (SLTs) about what speech and language progress to expect following the surgery and to enable them to be more independent in their management of children with this technology. The initial phase of the study looks at community SLT confidence in working with this client group and what support they need from the auditory implant programmes, specifically the implant 96 centre speech and language therapists (ICSLTs). We used qualitative methods, consisting of focus groups and interviews, analysed using thematic analysis. We found that SLTs value the support and advice of the ICSLTs to demystify the technology, to advise in target setting and in differential diagnosis. Some issues were raised regarding communication between professionals, particularly for liaison about target setting. Those SLTs who specialise in hearing impairment reported feeling confident about working with deaf children who have no additional needs, but valuing ICSLT support when there are other issues to tease out. The more generalist SLTs feel they need more training and information about what to aim for and in what timescale. We hope that further work can propose a model of support which responds to the evidence and is appropriate for the level of information which can be found in the literature as well as the knowledge held by SLTs. Acknowledgements: The research was supported by a grant from the National Institute for Health Research 72. Pitch ranking and speech recognition in children with cochlear implants R. Miah* and D. Vickers§ , *Audiology Department, Princess of Wales Hospital, South Wales § Ear Institute, University College London, U.K Kopelovich et al (2011) and Dawson et al (2000) provided evidence suggesting that children with cochlear implants (CIs) can pitch rank; that this ability improves with age and is related to speech perception. The goal of this research was to determine whether pitch ranking ability of implanted children was related to speech recognition when vocabulary age was accounted for. Fourteen children aged between 4-17 years participated; nine used Med-El, four Cochlear and one Advanced Bionics devices, nine were unilaterally and five bilaterally implanted. Pitch ranking and speech perception tests were delivered using computer generation of stimuli and automated response collection. Sounds were presented through a loudspeaker one metre directly in front of the listener in a sound-attenuating booth. For pitch ranking pairs of pure tones (1000ms duration and 1000ms inter-stimulus interval) were delivered at 65dBA. A two-alternative forced choice paradigm was used and participants indicated which tone was higher in pitch, feedback was provided. Stimuli were twelve tones from 262Hz to 6272Hz separated by half an octave. ±3dB level jitter was used. Speech perception in quiet was assessed using the Chear Auditory Perception Test (CAPT) presented at 50dBA and speech perception in noise was assessed using the Children’s Coordinate Response measure (CCRM); with adaptive speech-shaped noise and fixed level speech (65dBA) measuring the 50% speech reception threshold. Vocabulary age was assessed using the Renfrew word finding test (Renfrew, 1988). When vocabulary age was controlled for there was a moderate correlation between pitch ranking and CAPT (r = 0.54, p=0.021) and also for CCRM (r = -0.62, p = 0.006). The findings showed that pitch ranking can be assessed in children with CIs and that there is a relationship with speech recognition even when developmental age is accounted for. Therefore pitch ranking may potentially be useful for mapping and habilitation for children with CIs. Acknowledgments Thanks to participants and families and CI programme at the Princess of Wales Hospital. The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 97 References Dawson, P.W., McKay, C M., Busby, P A., Grayden D B., and Clark, G M. 2000. Electrode discrimination and speech perception in young children using cochlear implants. Ear Hear. 21 (6): 597 -607 Kopelovich, J C., Eisen, M D and Franck K H. 2010. Frequency and electrode discrimination in children with cochlear implants). Hearing Res. 268: 105-113 Renfrew, C.E. (4th edition) 2009. The Renfrew Language Scales: Word finding vocabulary test. United Kingdom, Speechmark. 73. Cochlear implant listeners at a cocktail party: evaluating CI performance in multi-talker listening situations with the CRM (Coordinate Response Measure). Cooper, H.R. University Hospital Birmingham Cochlear implants (CI) can provide excellent speech discrimination in quiet, with large variance in individual results. Standard tests of speech recognition do not replicate real-life listening conditions where multiple, spatially separated maskers (background noise or competing talkers) create difficulties for CI listeners. Cues available to NH listeners to solve the ‘cocktail party problem’ may not be available to CI listeners (Cherry, 1953 ; Darwin, 2008). The CRM (Coordinate Response Measure) measures the ability to selectively attend and respond to a target sentence in the presence of distracter talkers either co-located or at random spatially separated locations. Stimuli were presented from 9 head-height loudspeakers in the sound field. The target sentence contained a pre-determined ‘call sign’ that listeners listened for. All sentences had the form ‘Ready (call sign) go to (colour)(number) now. Listeners reported the target colour and number (closed set choice of 16 responses (four colours/four numbers)). An adaptive procedure varied the relative level of target and masker sentences to find the SNR for 50% correct performance. For NH listeners (n=12), median SRTs (50% correct), in 3 conditions (A : two talkers, one female/one male, single speaker in front ; B : 7 talkers, locations of target and maskers randomised across 9 loudspeakers, with an 800 ms onset delay between target and masker sentences, and C : 7 talkers, locations of target and maskers randomised across 9 loudspeakers, target and masker sentences presented in pairs with no onset delay) were (A) -19.5 dB ; (B) -10.8 dB and (C); -1.7 dB . Unilateral CI listeners (n=8) had median SRTs in the 3 conditions of (A) +1.72 dB ; (B) +6.57 and (C) +9.54 dB. Only 2 listeners achieved an SRT at a negative SNR, in the 2-talker condition. Bilateral CI listeners (n=8), tested a) with bilateral implants and b) with the better implant alone, all had positive SNRs in all 3 conditions. There was no significant bilateral advantage in condition A (2 colocated talkers) or C (7 talkers without onset delay), but there was a significant advantage in condition B (7 talkers with 800 ms onset delays). Unlike NH listeners, CI users require positive SNRs in this multi-talker listening task. A bilateral advantage was only found when onsets of multiple talkers were non-simultaneous. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Cherry E.C. (1953): Some experiments on the recognition of speech, with one and with two ears. Jl of Acoust Soc Am, 25:975-979 98 Darwin C.J (2008) Listening to speech in the presence of other sounds. PhilTrans R Soc B, 363, 1011-1021 74. Comparing the benefit to sentence recognition in noise through a cochlear implant of a contralateral modulated tone and noisy speech T.Green, S. Rosen and A. Faulkner, Department of Speech, Hearing and Phonetic Sciences, UCL, London, UK Attempts to identify sources of bimodal benefit have shown that cochlear implant users with residual low-frequency acoustic hearing can achieve improved speech recognition relative to implant-alone from combining the implant signal with a contralaterally presented modulated tone conveying voicing, fundamental frequency (F0) and amplitude envelope information (Brown & Bacon, 2009). Earlier work examining F0-controlled sinusoids as an aid to lip-reading for profoundly hearing impaired listeners established that a simple neural network could extract useful voicing and F0 information from noisy speech in real-time (Faulkner et al, 1997). This suggests that there may be circumstances in which bimodal benefit to speech perception in noise would be maximised by presenting to the unimplanted ear a simplified signal conveying extracted F0 and voicing information rather than amplified noisy speech. The present work attempts to establish the range of signal-to-noise-ratios (SNRs) at which typical bimodal CI users are helped by a clean simplified acoustic signal, and to compare the benefit provided to that from simple amplification of noisy speech. Recognition of target BKB sentences in noise was assessed with implant alone and in combination with a contralateral acoustic signal consisting of either the speech and noise mixture, low-pass filtered at 500 Hz, or a noise-free modulated tone. The tone follows the F0 contour of voiced segments of the target speech, and is amplitude modulated with the envelope of the target speech. No tone is presented during unvoiced target segments. Performance is assessed both in steady speech-shaped noise and in noise modulated with an envelope extracted from connected speech from a single-talker. The latter allows assessment of the extent to which bimodal hearing promotes glimpsing of target speech in modulated maskers. Percentage of key words correct was measured for each combination of acoustic input and noise type at each of four SNRs evenly-spaced between 0 and +12 dB. Identifying the circumstances in which the modulated tone provides greater bimodal benefit than low-frequency speech in noise would establish the requirements for real-time F0 extraction that would need to be met for approaches using a modulated tone to be viable in realistic environments. Acknowledgements Supported by Action on Hearing Loss. References Brown, C. A. & Bacon, S. P. 2009. Achieving electric-acoustic benefit with a modulated tone. Ear Hear, 30, 489-493. Faulkner, A., van Son, N. & Beijk, C. M. 1997. The TIDE project OSCAR. Scandanavian Audiology, 26, Suppl 47, 38-44. 75. Improved measures of spatial selectivity in cochlear implant users S. Cosentino, J.M. Deeks and R.P. Carlyon, MRC Cognition and Brain Sciences Unit, Cambridge, UK 99 Cochlear implant (CI) users rely on a finite number of channels to perceive acoustic information. A number of factors including the presence of dead regions in the cochlea or the positioning of an electrode in the lateral wall, limit the selectivity with which electrodes stimulate the desired subset of auditory nerve fibres. The present study aimed to develop a measure of CI spatial selectivity that is sensitive to dead regions. This was achieved by measuring psychophysical tuning curves for both single and dual-masker electrodes using interleaved stimulation. This method was chosen because, we argue, the more commonly used forward-masking method is susceptible to non-sensory effects that can lead to an over-estimation of spatial selectivity and that may miss dead regions. Stimulation rate was chosen to minimize channel interaction between probe and masker pulses. The level of the maskers was adaptively changed to mask a fixed-level probe, to ensure the same amount of excitation at threshold near the probe. Equal currents were applied to the masking electrodes in the dual-masker configuration; this differs from equal loudness setting of the maskers, which, we argue, is adversely influenced by neural survival at the masker, rather than the probe, position. The stimulus consisted of a 48-ms 125-pps pulse train (probe) presented in the middle of a 368-ms 125-pps masker. In the dual-masker condition, a second masker pulse was added 90us after the first, and the presentation order with respect to the probe was alternated between the two masker pulses (e.g., M1 - M2 - P M2 - M1 - P - M1…). The pulses were 60-µs cathodic-first biphasic pulses. A 3I-2AFC procedure was used with the probe fixed at 10 or 20% of its dynamic range, and the masker level of one or two electrodes was varied for different probe-masker electrode separations. CI users implanted with MedEl and Advanced Bionic devices took part. With two masking electrodes, masker levels at probe masked thresholds generally increased with increasing separation between the probe and the maskers. This reflects excitation of the neural population responding to the probe that diminishes as the masker is physically moved away from the probe location. Combined with single-masker thresholds, these data revealed several cases where the probe was detected via neurons closer to adjacent electrodes than to the probe electrode. The relative contributions of electrode position and neural survival to these cases was assessed by combining the masking measures with detection thresholds across the electrode array. Acknowledgements SC is funded by Action on Hearing Loss. 76. Sound localisation by cochlear implant users and normally hearing adults: The role of head movements A.M. Goman*, P.T. Kitterick§ and A.Q. Summerfield*, *Department of Psychology, University of York, UK, § NIHR Hearing Biomedical Research Unit, Nottingham, UK Head movements can help listeners reduce front-back confusions and determine the location of a sound source (Wightman & Kistler, 1999). It is less clear whether head movements improve localisation accuracy for cochlear implant users. This study sought to investigate localisation accuracy and listening effort in a group of cochlear implant users and normally-hearing adults in a variety of listening environments. 24 normally hearing adults and 8 cochlear implant users (4 bilateral) participated. Participants sat in the centre of an array of loudspeakers: Five were positioned in front of the listener (separated by 15°) and five were positioned behind the listener (separated by 15°). In a blocked design, a short (~0.8s), medium (~5.4s) or long (continuous repetition) stimulus was presented from one of 5 (front-only condition) or 10 (front-back condition) locations. Participants indicated where they perceived the location of the sound to be and at the end of each block rated their listening effort on a 100 0-100 scale where zero indicated no effort and 100 indicated extreme effort. Head movement was either permitted or not permitted. Motion tracking enabled changes in yaw to be measured. Bilateral users were more accurate (59% correct) than unilateral users (18% correct) except in the most challenging listening condition (short, front-back). Despite making use of additional time when it was available, performance for unilateral users remained low in all conditions. However, for bilateral users, accuracy improved when head movements were permitted for both the front-only and front-back conditions. Conversely, whilst performance by normally-hearing adults was at ceiling in the front-only condition, permitting head movements worsened performance slightly but significantly in the front-back condition. Normally hearing adults did not benefit from additional time and reported more listening effort was required when movement was not permitted. This study found localisation to be better with two implants than one. Bilateral users benefit from the varied interaural differences which arise as a result of head movements and can determine the location of a sound source if sufficient time is available. In this study, accuracy by normally hearing adults was higher when head movements were not permitted. Possibly cognitive control perceived as effort is increased when head movement is not permitted. It is possible that this cognitive control results in more concentration which in turn results in more accurate performance. Acknowledgements AMG is supported by and ESRC studentship References Wightman, F., & Kistler, D.J. (1999). Resolution of front-back ambiguity in spatial hearing by listener and source movement. JASA, 105, 2841-53. 77. Effects of frequency-to-place mismatch on sound localisation in electro-acoustic hearing S. Morris* and P.T. Kitterick§. *MRC Institute of Hearing Research, Nottingham, UK. §NIHR Nottingham Hearing Biomedical Research Unit, Nottingham, UK Individuals with severe-to-profound deafness in one ear and normal hearing in the other have minimal access to inter-aural cues which support accurate localisation. Cochlear implantation (CI) of the impaired ear can aid localisation despite probable differences in the mapping between frequency and cochlear place in the implanted and non-implanted ears. A frequency-to-place mismatch between the ears can reduce sensitivity to inter-aural level cues (Goupell et al., 2013) and may limit binaural benefits in these individuals. Listening with an implanted and a normal-hearing ear was simulated to address two questions: 1) Does a mismatch between the implanted and nonimplanted ears impair localisation; and 2) Can the mapping of the CI be adjusted to optimise localisation? Binaural recordings were made of monosyllabic words presented from one of seven loudspeakers at 20-degree intervals from -60 to +60 degrees azimuth. Recordings were presented over headphones and roved to disrupt monaural loudness cues. Participants reported the location of each word from the set of possible locations. Signals presented to the right (CI) ear were processed in one of three ways. Ideal processing represented the theoretical ability to deliver a wide range of frequencies to sites in the cochlea with matching characteristic frequencies. Realistic processing delivered the same wide range of frequencies but to an electrode array whose length and position were based on published specifications and surgical outcomes, respectively. Best-Fit processing allocated frequencies across the electrode array based on its position in the cochlea. Nine normalhearing participants completed three sessions, one for each simulation. Before each session, loudness balancing was performed and training was provided. 101 Localisation (mean absolute difference between actual and reported locations) and lateralisation (percentage of trials where actual and reported locations were on the same side) exceeded chance levels for all simulations. A mismatch between the ears impaired performance significantly compared to matched presentation (Realistic vs Ideal, mean differences: localisation +6.0 degrees; lateralisation -7.9%). Allocating frequencies based on electrode position produced numerical but non-significant improvements compared to position-independent allocation (Best-Fit vs Realistic, mean differences: localisation -1.9 degrees; lateralisation +4.6 %). The results suggest that a mismatch in frequency-to-place mapping between an implanted ear and a non-implanted ear impairs localisation. Allocating frequencies based on electrode positioning does not appear to improve localisation after short-term exposure. Acknowledgements Supported by the intramural program of the MRC and infrastructure support from the NIHR. References Goupell M.J., Stoelb C., Kan A., Litovsky R.Y. 2013. Effect of mismatched place-of-stimulation on the salience of binaural cues in conditions that simulate bilateral cochlear-implant listening. J Acoust Soc Am, 133(4), 2272-87. 78. A conceptual framework for studying cochlear implant users’ relationship with music: music-related quality of life G. Dritsakis and R.M. van Besouw, ISVR, University of Southampton, UK The relationship that cochlear implant (CI) users have with music has been the focus of many recent studies (Looi et al. 2012). However, from a review of studies on music and CIs published since 2000, it is evident that there is inconsistent use of terminology. For instance, music appreciation and music enjoyment are sometimes used interchangeably. As a result, it is often difficult to interpret findings and draw comparisons across studies. In addition, the concept of quality of life (QoL) has been overlooked, despite evidence showing the impact of music on the QoL of normal-hearing listeners and significant positive correlations between music enjoyment or sound quality ratings and QoL scores for CI users. Consequently, the psychological and social dimensions of CI users’ relationship with music, such as attitudes towards music and participation in music-related social activities, are under-researched. The aim of the present study is to develop a conceptual framework that will describe CI users’ relationship with music taking into account QoL. The concept of ‘musicrelated quality of life’ (MRQoL) is introduced. It refers to aspects of the QoL of CI users that are particularly related to music. Issues relating to CI users’ relationship with music identified by previous studies, e.g. pitch perception, music appraisal or listening habits, are organised into the physical, psychological and social QoL domains (Hinderink et al. 2000). The use of terms and the rationale behind the development of the framework are explained. The proposed framework is holistic and can be adapted in order to cover additional domains or items. It can be used to structure research in the area of music and CIs and is currently being used as the basis for the development of a music-specific QoL measure. Focus group data from adult CI users will be analysed according to the framework and questionnaire items will be generated. The analysis of the focus group data will also provide evidence concerning the usefulness of the framework in describing CI users’ relationship with music. Acknowledgments Supported by the Faculty of Engineering and the Environment (FEE), University of Southampton 102 References Hinderink, J.B., Krabbe, P.F. & Van Den Broek, P., 2000. Development and application of a healthrelated quality-of-life instrument for adults with cochlear implants: the Nijmegen cochlear implant questionnaire. Otolaryngology-Head and Neck Surgery, 123(6), pp.756–65. Looi, V., Gfeller, K. & Driscoll, V., 2012. Music appreciation and training for cochlear implant recipients: A review. Seminars in Hearing, 33(4), pp.307–334. 79. The effect of stimulus polarity and pulse shape on spread of excitation in cochlear implant users R.P. Carlyon §, J.M. Deeks§ , A. van Wieringen ‡, J. Undurraga‡, O. Macherey§, and J. Wouters‡, § MRC Cognition and Brain Sciences Unit, Cambridge, UK, ‡ Lab. Exp. ORL, Univ. Leuven, Belgium Physiological measurements obtained with animals have shown more restricted spread of neural excitation with bipolar (BP) compared to monopolar (MP) stimulation. However, in BP mode the current at one electrode is a polarity inverted version of that at the other, leading to a possible bimodal spread of excitation, at least for the commonly-used symmetric (SYM) biphasic pulses. Experiments using asymmetric pulses show that, at least at moderately high levels, CI users are more sensitive to anodic than to cathodic current. This raises the possibility that asymmetric pulses in bipolar mode might be used to selectively excite neurons near one electrode. We measured forward-masked psychophysical tuning curves in which 200-ms 1031-pps BP maskers consisted either of either SYM or pseudomonophasic (PS) pulses ; for PS pulses both polarities were tested. The signal consisted of three 97 µs/phase SYM pulses at a rate of 344 pps, separated from the masker by a 6.8-ms silent interval, at a +3 dB sensation level. The first phase of each masker pulse had a duration of 97 µs, and, for PS pulses, was followed by an opposite-polarity phase of four times this duration and one quarter of its amplitude. The hypothesis was that PS maskers should be more effective when the short high-amplitude phase was anodic on the electrode closest to the probe, compared to when it was cathodic. Hence we predicted an interaction between the effects of masker position and polarity. Seven users of the Advanced Bionics CI took part. A repeated-measures ANOVA (masker position X polarity) on just the PS masker data revealed no significant main effects or interactions. However, ANOVAs on the results for individual subjects revealed a significant interaction for six out of the seven subjects. For all subjects there was at least one electrode position where the masker level at threshold differed significantly between the two PS polarities. For five of the seven listeners all of these significant differences were in the predicted direction. However, for two listeners there were significant differences in the opposite direction. We discuss possible reasons for these different patterns of results, including the possibility that some listeners may show greater sensitivity to cathodic stimulation at low levels ; although all of our maskers were well above threshold, the excitation that they produced at the probe place was probably quite low. Acknowledgement: Supported by MRC and KU Leuven. 80. Comodulated notched-noise masking Ramona Grzeschik*, Björn Lübken* and Jesko L. Verhey*, *Department of Experimental Audiology, Otto von Guericke University Magdeburg, Magdeburg, Germany Detection thresholds of a masked sinusoidal signal are lower when the masking noise is coherently modulated over a wide frequency range compared to a masking condition with the same spectrum 103 but using an unmodulated masker. This phenomenon is referred to as comodulation masking release (CMR; Hall et al., 1984). Here we tested how masker comodulation affects notched-noise data. Thresholds for sinusoids were measured in the presence of a diotic notched-noise masker for comodulated and unmodulated conditions. The signal frequency of the target was 500 Hz. The masker was a broadband filtered noise with lower- and upper cut-off frequencies of 30 and 1000 Hz, respectively. The masker had a spectral notch at the signal frequency. Notch widths were 0, 50, 100, 200, and 400 Hz. The arithmetic centre frequency of the notch was equal to the signal frequency. The data of eleven participants (4 male, 7 female, aged 25-37 yrs, mean 30 yrs) entered the analysis. As expected, thresholds decreased with increasing notch width for both masking conditions. The CMR decreased continuously with notch widths as well. CMR was almost 9 dB for no notch. At 400 Hz, CMR was less than 2 dB. The experimental data were compared with predictions of a modulation filterbank model that was previously shown to simulate CMR in the bandwidening type of CMR experiments (Piechowiak et al., 2007). Acknowledgements Supported by the German Research Foundation (SFB/TRR31). References Hall JW, Haggard MP, Fernandes MA (1984) Detection in noise by spectrotemporal patternanalysis. J Acoust Soc Am 76: 50-56. Piechowiak T, Ewert SD, Dau T (2007) Modeling comodulation masking release using an equalization-cancellation mechanism. J Acous. Soc Am 121 (4): 2111-2126. 81. Cochlear implant listeners at a cocktail party: evaluating CI performance in multi-talker listening situations with the CRM (Coordinate Response Measure). Cooper, H.R. University Hospital Birmingham, Bormingham, UK Cochlear implants (CI) can provide excellent speech discrimination in quiet, with large variance in individual results. Standard tests of speech recognition do not replicate real-life listening conditions where multiple, spatially separated maskers (background noise or competing talkers) create difficulties for CI listeners. Cues available to NH listeners to solve the ‘cocktail party problem’ may not be available to CI listeners (Cherry, 1953 ; Darwin, 2008). The CRM (Coordinate Response Measure) measures the ability to selectively attend and respond to a target sentence in the presence of distracter talkers either co-located or at random spatially separated locations. Stimuli were presented from 9 head-height loudspeakers in the sound field. The target sentence contained a pre-determined ‘call sign’ that listeners listened for. All sentences had the form ‘Ready (call sign) go to (colour)(number) now. Listeners reported the target colour and number (closed set choice of 16 responses (four colours/four numbers)). An adaptive procedure varied the relative level of target and masker sentences to find the SNR for 50% correct performance. For NH listeners (n=12), median SRTs (50% correct), in 3 conditions (A : two talkers, one female/one male, single speaker in front ; B : 7 talkers, locations of target and maskers randomised across 9 loudspeakers, with an 800 ms onset delay between target and masker sentences, and C : 7 talkers, locations of target and maskers randomised across 9 loudspeakers, target and masker sentences presented in pairs with no onset delay) were (A) -19.5 dB ; (B) -10.8 dB and (C); -1.7 dB . 104 Unilateral CI listeners (n=8) had median SRTs in the 3 conditions of (A) +1.72 dB ; (B) +6.57 and (C) +9.54 dB. Only 2 listeners achieved an SRT at a negative SNR, in the 2-talker condition. Bilateral CI listeners (n=8), tested a) with bilateral implants and b) with the better implant alone, all had positive SNRs in all 3 conditions. There was no significant bilateral advantage in condition A (2 colocated talkers) or C (7 talkers without onset delay), but there was a significant advantage in condition B (7 talkers with 800 ms onset delays). Unlike NH listeners, CI users require positive SNRs in this multi-talker listening task. A bilateral advantage was only found when onsets of multiple talkers were non-simultaneous. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Cherry,E.C. (1953): Some experiments on the recognition of speech, with one and with two ears. J Acoust Soc Am, 25, 975-979 Darwin C.J (2008) Listening to speech in the presence of other sounds. Phil Trans R Soc B, 363, 1011-1021 Poster Abstracts: Balance and vEMPs 82. Cervical and ocular VEMPs in secondary school age children D. Deval*, C. Shaw§, A. Shetye§, K. Harrop-Griffiths* and W. Pagarkar§,*Nuffield Hearing and Speech Centre, Royal National Throat Nose and Ear Hospital, London, UK, §Department of Audiovestibular Medicine, Hackney Ark, London, UK. VEMPs are myogenic potentials generated following a sound stimulus and most commonly detected by electrodes placed on the sternocleidomastoid muscle (cervical VEMPs or cVEMPs) (Colebatch et al, 1994). More recently recording from the extra ocular muscles has been described (ocular VEMPs or oVEMPs) (Rosengren et al, 2005). The cVEMPs are saccular in origin whereas the oVEMPs are thought to be utricular, although their waveforms show some similarity. Studies describing normative data of oVEMPs in children are scarce (Wang et al 2013). Studies comparing oVEMPs and cVEMPs in adults have been published (Nguyen et al 2010), but there are no published studies of comparison between oVEMPs and cVEMPs in children with normal hearing. This study aims to compare the results of oVEMPs and cVEMPs in secondary school age children with normal hearing. The study recruited 50 secondary school children of both sexes, 11-16 years of age. Children were excluded if they had hearing loss, middle ear dysfunction, history of vestibular symptoms, neurological problems, developmental delay, cervical spine or neck pathology, ocular muscle palsies or migraine headaches. All participants had pure tone audiometry, tympamometry, cVEMPs and oVEMPs testing. The VEMP parameters recorded were: P and N wave latency (miliseconds), amplitude of waveform (microvolts), asymmetry ratio between amplitudes of two ears and threshold. The t test was used to compare the cVEMPs and oVEMPs data. The mean cVEMPs threshold was 75dBnHL (±5.83) for the right ear and 76 dBnHL (±5.33) for the left ear. The mean oVEMPs threshold was 83dBnHL (±5.07) for the right ear and 83dBnHL (±4.49) for the left ear. The mean waveform amplitude of cVEMPs was 140microvolts (±99.19) for the right ear and 143.81microvolts (±91.02) for the left ear. Corresponding values for oVEMPs were 6.47microvolts (±4.37) (right ear) and 6.73microvolts (±6.58) (left ear).The asymmetry ratio was 17.93 (±11.87) for cVEMPs and 25.64 (±16.22) for oVEMPs. 105 There was no significant difference between the mean values of the right and the left ears for cVEMPs and oVEMPs threshold, wave latency and amplitude. The oVEMPs had a significantly smaller amplitude and a higher threshold than the cVEMP. Both cVEMPs and oVEMPs show a wide variation in amplitude. This study shows the differences in the waveforms of the cVEMPs and oVEMPs in children with normal hearing. A larger study is required to compare the cVEMPs and oVEMPs response variations with age and sex. Acknowledgements Supported by the BSA. References Colebatch J.G., Halmagyi G.M. & Skuse N.F. 1994. Myogenic potentials generated by a clickevoked vestibulocollic reflex. J Neurol Neurosurg Psychiatry, 57(2), 190-197. Rosengren S.M., McAngus Todd N.P. & Colebatch J.G. 2005. Vestibular-evoked extraocular potentials produced by stimulation with bone-conducted sound. Clin Neurophysiol, 116(8), 19381948. Wang S.J., Hsieh W.S. & Young Y.H. 2013. Development of ocular vestibular-evoked myogenic potentials in small children. Laryngoscope, 123(2), 512-517. Nguyen K.D., Welgampola M.S. & Carey J.P. 2010. Test-Retest Reliability and Age-Related Characteristics of the Ocular and Cervical Vestibular Evoked Myogenic Potential Tests. Otol Neurotol, 31(5), 793-802. Poster Abstracts: Tinnitus 83. An internet-based approach to characterising the misophonia population V.L. Kowalkowski*§ and K. Krumbholz*, MRC Institute of Hearing Research, UK, University of Nottingham, UK The term ‘Misophonia’ was coined by otolaryngologists approximately fifteen years ago to identify a population of individuals who experience negative emotional responses to certain everyday sounds (Jastreboff & Jastreboff, 2001), but who do not fit the criteria of hyperacusis. Misophonia is thought to be caused by dysfunctional interaction between the auditory and limbic systems (Jastreboff & Jastreboff, 2001), and might therefore share similarities with tinnitus distress. Whereas tinnitus has attracted intense research interest, little systematic research exists on misophonia, probably because the condition is purportedly rare. Online support networks for misophonia, however, suggest a much greater prevalence than seen in the clinic. This is supported by a recent group study of misophonia, based on 42 patients, which reported that recruitment through self-referral was easy (Schröder et al., 2013). We developed an online questionnaire, targeted at online misophonia support groups, to determine to what extent findings by previous case studies (Edelstein et al., 2013, Webber et al., 2014) represent the wider misophonia population. The questionnaire combines psychological and audiological approaches by containing an adapted version of the diagnostic-oriented A-Miso-S questionnaire, as well as two new questionnaires: the Misophonia Handicap Questionnaire (MHQ) and Misophonia Handicap Inventory (MHI), which were derived from the Tinnitus Handicap Questionnaire and Tinnitus Handicap Inventory, respectively. We present analyses of the data from 80 respondents, obtained within three weeks of the questionnaire’s release. The data suggest that, of these 80 respondents, 40% are likely to have 106 hyperacusis instead of, or in addition to, misophonia. The data also support the suggestion from previous studies for a hereditary component to misophonia: 50% of the respondents without suspected hyperacusis reported immediate family members with similar symptoms. To evaluate the measurement properties of the new misophonia questionnaires, we used the methodological approach described by Terwee et al (2007). We found that the MHI correlated well with the A-Miso-S questionnaire (r=0.73, P(>F)<0.001) and did not suffer from ceiling or floor effects, suggesting that it represents a good measure of the effect of misophonia on an individual’s quality of life. The MHQ correlated less strongly with the A-Miso-S (r=0.58, P(>F)<0.001), but did correlate strongly with the MHI (r=0.8, P(>F)<0.001). Future work will be aimed at validating the new questionnaire measures, and assessing their usefulness in clinical and research settings. Some of the questionnaire participants will be invited for EEG recordings of spontaneous brain activity, to be compared with subjects suffering from bothersome tinnitus. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. References Jastreboff & Jastreboff 2001. Components of decreased sound tolerance: hyperacusis, misophonia, phonophobia. ITHS News Lett. 2, 5–7 Edelstein M., Brang D. and Ramachandran V. S. (2013) Front Hum Neurosci 7:296 Webber T. A., Johnson P. L., Storch E.A 2014 Gen. Hosp. Psych. 36:2, 231 Schroder A., Vulink N. and Denys D. 2013 PLOS One 8, 1-5 Terwee C. B., Bot S. D. M., de Boer M. R., van der Windt D. A. W. M., Knol D. L., Dekker J., Bouter L. M., de Vet H. C. W., 2007. Quality criteria were proposed for measurement properties of health status questionnaires. J Clin Epid 60,1,34-42. 84. Source space estimation of oscillatory power and brain connectivity in tinnitus O. Zobay*, A. R. Palmer*, D. A. Hall§,‡, M. Sereda§,‡ and P. Adjamian*, *MRC Institute of Hearing Research, University Park, Nottingham, UK, §National Institute for Health Research (NIHR) Nottingham Hearing Biomedical Research Unit, Nottingham, UK, ‡Otology and Hearing group, Division of Clinical Neuroscience, School of Medicine, University of Nottingham, UK Tinnitus is the perception of an internally generated sound that is postulated to emerge as a result of structural and functional changes in the brain. However, the precise pathophysiology of tinnitus remains unknown. The thalamocortical dysrhythmia model (Llinas et al. 1999, 2005) suggests that neural deafferentation due to hearing loss causes a dysregulation of coherent activity between thalamus and auditory cortex. This leads to a pathological coupling of theta and gamma oscillatory activity in the resting state, localised to the auditory cortex where normally alpha oscillations should occur. Numerous studies also suggest that tinnitus perception relies on the interplay between auditory and non-auditory brain areas. According to the Global Brain Model (Schlee et al, 2011), a network of global fronto–parietal–cingulate areas is important in the generation and maintenance of the conscious perception of tinnitus. Thus, the distress experienced by many individuals with tinnitus is related to the top–down influence of this global network on auditory areas. In this magnetoencephalography (MEG) study, we compare resting-state data of tinnitus subjects and normal-hearing controls to examine effects on spectral power as well as functional and effective connectivity. The analysis is based on beamformer source projection and an atlas-based region-ofinterest approach. We find increased functional connectivity within the auditory cortices in the 107 alpha band. A significant increase is also found for the effective connectivity from a global brain network to the auditory cortices in the alpha and beta bands. We do not find evidence of effects on spectral power. Overall, our results provide only limited support for the thalamocortical dysrhythmia and global brain models of tinnitus. Acknowledgements This research was funded by the Medical Research Council. MEG data was collected at the Sir Peter Mansfield Magnetic Resonance Centre, University of Nottingham. References Llinas R.R., Ribary U., Jeanmonod D., Kronberg E. & Mitra P.P. 1999. Thalamocortical dysrhythmia: A neurological and neuropsychiatric syndrome characterized by magnetoencephalography. PNAS, 96, 15222-15227. Llinas R., Urbano F.J., Leznik E., Ramirez R.R. & van Marle H.J.F. 2005. Rhythmic and dysrhythmic thalamocortical dynamics: GABA systems and the edge effect. Trends Neurosci, 28, 325-333. Schlee W., Lorenz I., Hartmann T., Müller N., Schulz H. & Weisz, N. 2011. A Global Brain Model of Tinnitus. In: Møller A.R., Langguth B., De Ridder D. & Kleinjung T., (eds.) Textbook of Tinnitus. New York: Springer. pp. 161-169. 85. Objectively characterising tinnitus using the postauricular muscle reflex J.I. Berger, B. Coomber, A.P. Carver, M.N. Wallace and A.R. Palmer, MRC Institute of Hearing Research, Nottingham, UK Tinnitus – the perception of sound in the absence of external auditory input – affects 8-15% of adults worldwide. Currently, there are no universally effective treatments, due to heterogeneity in tinnitus aetiologies and a poor understanding of underlying neural pathologies. Assessment of specific tinnitus characteristics relies solely on subjective self-report, which can often be unreliable. Accordingly, an objective test would be desirable both as a diagnostic tool and as an outcome measure for assessing treatment efficacy. A commonly-used behavioural test for tinnitus in animals – gap-induced prepulse inhibition of acoustic startle (GPIAS) - relies on measuring a reflex response to determine the detection of gaps in continuous stimuli. Impaired gap detection following tinnitus induction is thought to be caused by tinnitus ‘filling in’ the gap, thus reducing its salience. Preliminary work in humans measuring GPIAS of eye blink responses showed gap detection deficits in tinnitus subjects, but the underlying mechanisms of this effect were unclear (Fournier & Hébert, 2013). Conversely, gap recognition in a perceptual task was not found to be affected when hearing-impaired tinnitus subjects were compared with controls (Campolo et al., 2013). We have developed a variation of the GPIAS method in which we measure gap-induced prepulse inhibition (PPI) of the Preyer reflex in guinea pigs (Berger et al., 2013). The postauricular muscle reflex (PAMR) is an analogue of the Preyer reflex, present in humans, and may represent a possibility for developing an objective tinnitus test for the clinic. The PAMR is subject to prepulse inhibition (Cassella & Davis, 1986), and represents an alternative to the eye blink reflex. However, gap-induced PPI of the PAMR has yet to be demonstrated. In the present study, we measured PPI of the PAMR in a small group of normal-hearing subjects. PAMR responses were recorded electromyographically with surface electrodes placed over the insertion of the postauricular muscle to the pinna. Gap detection was evaluated with tonal background sounds, presented at either 60 or 70 dB SPL, and startle stimuli comprising very brief broadband noise bursts presented at either 85, 90, 95, or 100 dB SPL. Preliminary data suggest that 108 the PAMR is susceptible to gap-induced PPI, and is dependent on the relative sound levels of background and startle stimuli. Future studies will examine further the parameters that affect gap-induced PPI of the PAMR, and subsequently establish whether subjects with tinnitus exhibit deficits in gap detection. Acknowledgements BC was supported by Action on Hearing Loss (International Project Grant G62). AC was supported by an Action on Hearing Loss summer studentship. References Berger JI, Coomber B, Shackleton TM, Palmer AR, Wallace MN (2013) A novel behavioural approach to detecting tinnitus in the guinea pig. J Neurosci Methods 213: 188-95. Campolo J, Lobarinas E, Salvi R (2013) Does tinnitus “fill in” the silent gaps? Noise Health 15: 398-405. Cassella JV & Davis M (1986) Habituation, prepulse inhibition, fear conditioning, and drug modulation of the acoustically elicited pinna reflex in rats. Behav Neurosci 100: 39-44. Fournier P & Hébert S (2013) Gap detection deficits in humans with tinnitus as assessed with the acoustic startle paradigm: does tinnitus fill in the gap? Hear Res 295: 16-23. 86. Investigating the central gain mechanism following a short period of unilateral earplug use H.E. Brotherton*, M. Maslin*, C.J. Plack*, R. Schaette§ and K.J. Munro*, School of Psychological Sciences, University of Manchester, Manchester, UK, §Ear Institute, University College London, London, UK. Previous research has shown a change in the threshold of the middle ear acoustic reflex after continuous earplug use in one ear for seven days (Munro et al, 2014). The main aim of the present study was to extend the findings of Munro and colleagues and determine (i) the time course of the changes, (ii) if the greatest change occurred at frequencies affected by the earplug and (iii) if the underlying mechanism operated in the ascending auditory pathway. The present study investigated the acoustic reflex threshold in 25 normal hearing listeners after six days of continuous unilateral earplug use. Acoustic reflex thresholds for pure tones (0.5-4 kHz) and broadband noise were measured using a blinded design. Measurements took place at baseline, two, four, and six days after earplug use, as well as four and 24 hours after final removal of the earplug on day six. Reflex thresholds were obtained at a lower sound pressure level when the highfrequency and broadband noise stimuli were presented to the previously plugged ear, and there was no change when stimuli were presented to the control ear. The mean change of around 5 dB was greatest after four days of unilateral earplug deprivation. The onset of change in the reflex threshold was found to be slower than the recovery after the earplug was removed. Most of the recovery had taken place within 24 hours of earplug removal. The findings reveal direct evidence for adaptive plasticity after a period of unilateral auditory deprivation. The changes in the reflex threshold are consistent with a gain control mechanism that operates on a time scale of hours to days and operates in the ascending auditory system at frequencies most affected by the earplug. The findings may have implications for patients with tinnitus or hyperacusis. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 109 References Munro K.J., Turtle C. & Schaette R. 2014. Plasticity and modified loudness following short-term unilateral deprivation: Evidence of multiple gain mechanisms within the auditory system. J Acoust Soc Am, 135, 315-322. 87. Unilateral hearing loss in the ferret: functional characterisation of a novel model of tinnitus J.R. Gold*, F.R. Nodal*, D.N. Furness§, A.J. King* and V.M. Bajo** Department of Physiology, Anatomy and Genetics, University of Oxford, UK, §Institute for Science and Technology in Medicine, Keele University, UK Tinnitus in its subjective form is characterised by the patient-reported presence of a sound percept without any corresponding environmental acoustic source. Animal models have been sought that may reproduce such a percept in the subject, verified using behavioural measures and understood by physiological and anatomical means. However, published data have largely obtained either behavioural or physiological measurements within the same subjects, rarely combining multiple measures within individuals over time. Therefore, we have aimed to address this lacuna in the field by developing a novel model of maladaptive plasticity based upon selective sensorineural hearing loss in adult ferrets. Individual animals (n = 7) were tracked longitudinally (>12 months), with auditory function assessed using an operantly-conditioned gap-detection paradigm and measurement of auditory brainstem responses (ABRs). Upon acquisition of stable behaviour, selective mechanical lesions were made unilaterally in the basal coil level of the spiral ganglion (SG), which was followed by further behavioural testing and ABR measurement at multiple time points post-recovery. We subsequently determined that SG cell density in control cochleas were 1.6x higher than in lesioned specimens. Behavioural performance outcomes were heterogeneous, with bootstrap inference-based analysis revealing a subset of ferrets that, at specific times postlesion, displayed gap-detection psychometric function threshold elevations and slope changes. These animals displayed correlated changes in late-wave ABR shape and amplitude, akin to those observed in certain human tinnitus patient cohorts. In particular, these functional changes were distinct from those observed in animals without behavioural signs of tinnitus. Extracellular recordings performed in the auditory cortex bilaterally of lesioned and control ferrets revealed changes in multiunit spectrotemporal response properties of tinnitus-like animals, including tonotopic rearrangement and elevations of neural synchrony and spontaneous activity, in accordance with published tinnitus-like physiology, and our own behavioural and ABR-defined criteria. Histological processing of the same group of animals revealed that selective SG lesions also caused changes in auditory brain neurochemistry and particularly in NADPH diaphorase expression, within the primary auditory cortex and auditory brainstem nuclei. On the basis of these longitudinallyevaluated physiological changes, a second subset of lesioned ferrets received chronic bilateral fibreoptic implants in the auditory cortex, each of which had been injected with an AAV-based optogenetic construct to enable neural expression of the light-driven outward proton pump, ArchT. The aim of these experiments is to potentially reverse the tinnitus-like percept by silencing abnormal cortical signalling in awake behaving animals, with behavioural testing and analysis presently ongoing. Acknowledgements: This study is supported by the Wellcome Trust and Action on Hearing Loss. 110 88. Changes in resting-state oscillatory power in anaesthetised guinea pigs with behaviourally-tested tinnitus following unilateral noise trauma. V.L. Kowalkowski*§, B. Coomber*, M.N. Wallace*, K. Krumbholz*, MRC Institute of Hearing Research, UK, University of Nottingham, UK Chronic tinnitus affects about one in ten people in the UK, and can be highly bothersome for some individuals. It is often associated with noise trauma or hearing loss. There are several theories for how the tinnitus percept is generated, such as the ‘thalamo-cortical dysrhythmia’, and ‘corticalreorganisation’ models. In the former, deafferentation as a result of noise exposure leads to a slowing of spontaneous oscillatory activity between the auditory thalamus and cortex, as well as an increase in spontaneous high-frequency activity within auditory cortex. The increased spontaneous activity is thought to underlie the tinnitus percept. On the other hand, the latter model posits that tinnitus arises as a result of inhomogeneous (e.g., sloping) hearing loss, which deprives some areas of the hearing range of input, but leaves others intact. This is thought to create an imbalance of inhibitory and excitatory inputs to the intact areas, leading to expansion of their cortical representation, as well as increased spontaneous activity and tinnitus. In this study, we measured spontaneous oscillatory activity from the auditory cortical surface of 12 anaesthetised guinea pigs. Six of the animals were unilaterally exposed to loud noise to induce tinnitus and then allowed to recover for 8-10 weeks. The presence of tinnitus was tested with prepulse inhibition of the Preyer startle reflex, and ABR measurements were used to show that the exposed animal’s hearing thresholds had recovered to within 20 dB of their baseline. Six of the animals were unexposed controls. Consistent with the ‘thalamo-cortical dysrhythmia’ model, some differences in the frequency composition of the spontaneous activity were observed between the exposed and control groups, with a relative increase in mid-frequency activity in the exposed animals. More striking, however, was a highly significant group-by-hemisphere interaction, which seems to accord with the ‘cortical reorganisation’ model in that spontaneous activity in the contralateral hemisphere was reduced in the exposed animals compared to the controls, whilst activity in the ipsilateral hemisphere was increased. The contralateral hemispheres receives predominant input from the exposed ear, whereas the ipsilateral hemisphere receives predominant input from the intact ear. Future work aims to investigate changes in resting-state activity in humans with tinnitus and hearing loss using electroencephalography. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 89. An investigation into the relationships between hearing loss, hearing handicap, tinnitus handicap, and attention R. Lakshman*, N. Mohamad§, D. Hoare§, D.A. Hall§, School of Medicine, University of Nottingham*§, National Institute for Health Research (NIHR) Nottingham Hearing Biomedical Research Unit§ One major complaint of tinnitus patients is the effect tinnitus has on their ability to concentrate and perform attention-demanding tasks. However, whether and to what degree hearing loss or handicap has a confounding effect on the relationship between tinnitus handicap and cognitive function is unclear. To begin to interpret these relationships this study sought first to explore the correlation between pure-tone audiogram (PTA) average and self-reported hearing handicap. Essentially we wished to identify whether hearing loss was a sufficient predictor of hearing 111 handicap to be included as a covariate when considering the relationship between tinnitus and attention for example. One-hundred and twenty-two participants completed PTA, the Hearing Handicap Inventor (HHI), the Tinnitus Handicap Inventory (THI), and Subtest 7 of the Test of Everyday Attention (TEA) which provides a measure of sustained attention. PTA thresholds were measured at 0.5, 1, 2, and 4 kHz with averages calculated for across both ears, and individually (to determine better ear). There was a strong correlation between HHI scores and the average PTA across both ears (r = 0.72), and for the better ear only (r = 0.70) suggesting that either the audiogram data or the selfreport questionnaire would provide a reliable account of the effect of hearing in a model. In contrast, correlations between the THI and hearing measures were weak (r = 0.39 for HHI, 0.13 for PTA average) clearly demonstrating that the THI is measuring a construct different to that of the hearing measures. THI score only weakly correlated with TEA scores (r = -0.18) suggesting tinnitus to have little effect on sustained attention in this instance. We have demonstrated the reliability of using either hearing loss or HHI score to represent the construct of hearing. Further work will demonstrate whether tinnitus has effects on other forms of attention or other cognitive function. Acknowledgements This is a summary of work supported by the University of Nottingham and the National Institute for Health Research. 90. “100 years of amplified music”: a public science experiment on loud music exposure, hearing loss, and tinnitus M.A. Akeroyd*, W.M. Whitmer*, O. Zobay§, D.A. Hall‡, R. Mackinnon‡, H. Fortnum‡, and D.R. Moore#. * MRC/CSO Institute of Hearing Research - Scottish Section, Glasgow, United Kingdom; § MRC Institute of Hearing Research, University Park, Nottingham. ‡ NIHR Nottingham Hearing Biomedical Research Unit & Otology and Hearing Group, School of Medicine, University of Nottingham, Nottingham, United Kingdom; # Communication Sciences Research Center, Cincinnati Childrens Hospital, Cincinnati, OH, United States. It is well known that a lifetime of exposure to loud industrial noise can lead to substantial amounts of hearing damage. The effect of a lifetime of loud music exposure is far less certain, however, despite it being around 100 years since amplified music reproduction first became widely available. We ran a public-science experiment to get data addressing this issue, using the 2013 Centenary of the Medical Research Council to act as a further “pull” for the public to participate. The experiment ran for 18 weeks on a dedicated website. It was designed to assess whether one’s music-listening history across one’s lifetime was a predictor of speech reception thresholds (triple digits) in noise, self reported hearing ability, and self-reported tinnitus. The website combined (a) a 12-item questionnaire on the amount of listening to loud music (e.g. “How often would you say you went to gigs, concerts, and festivals in your twenties?”), (b) a set of standard questions on various aspects of hearing (e.g., “Do you currently have any difficulty with your hearing?” or “How often nowadays do you get noises such as ringing or buzzing in your head or ears (tinnitus) that last for more than 5 minutes?”) and (c) a triple-digit speech-in-noise test specifically crafted for high-frequency hearing losses (Vlaming et al., 2014). In all, 5301 people participated. We found no effect of loud music exposure on speech reception thresholds, but there were links between exposure, self-report hearing difficulty, and self-report tinnitus: those people reporting high exposures were more likely to report moderate or greater hearing difficulty (and tinnitus) than those with low exposures, and, vice versa, those people reporting low exposures were more likely to report no difficulty than those with high exposures. 112 We noted that participation rates were strongly correlated to media coverage, clearly demonstrating its importance for public science experiments. Acknowledgements The experiment was supported by the Medical Research Foundation. The team are supported by the Medical Research Council (grant number U135097131), the Chief Scientist Office of the Scottish Government, and the National Institute for Health Research. References Vlaming M.C.M.G, MacKinnon R.C., Jansen M., & Moore D.R. 2014. Automated Screening for High-Frequency Hearing Loss. Ear and Hearing, in press 91. The NIHR Clinical Research Network (CRN): what do we do for the hearing research community? D.A. Hall*, H. Blackshaw§, M. Simmonds§, A. Schilder§ *NIHR Nottingham Hearing Biomedical Research Unit, §evidENT Team, Ear Institute, University College London. The aim of this poster is to raise the profile of the NIHR Clinical Research Network (CRN), particularly the work of the Ear, Nose and Throat Specialty Group, among the audiology profession in the UK. The NIHR CRN has two broad aims. First, it provides the infrastructure to support highquality clinical research within the NHS, so that patients are given the opportunity to participate in research and ultimately benefit from new and better treatments. This is achieved by helping researchers to plan, set up and deliver clinical studies quickly and effectively; providing health professionals with research training, dedicated systems for gaining approvals, funding for experienced support staff and equipment and both advice and practical help with opening of sites and recruitment.. The second aim of the CRN is to support the Life Sciences Industry to deliver their research programmes. The CRN provides a committed industry team who can offer industries assistance with study design, site intelligence, site identification, study set-up and recruitment including dedicated clinical time and industry managers. The CRN structure comprises of a lead national office and 15 ‘Local CRNs’ (LCRNs) that cover the length and breadth of England. Each LCRN delivers research across 30 clinical specialties, including ENT. Each specialty has a national group of members representing each of the LCRNs. Together these local leads are called the ‘ENT Specialty Group’. The group is chaired by Prof Schilder and local lead members consist of research-active clinicians and practitioners across Ear, Nose, Throat, Hearing and Balance fields. Our job is to ensure that the ENT clinical research studies open in the UK and being supported by the CRN receive the correct specialty-specific advice to ensure the studies are delivered successfully in the NHS. For example, our ‘intelligence’ can help industry identify where the best sites to conduct research are, where specific populations are and how to access the right equipment. We support a UK portfolio of clinical research into the diagnosis, treatment, management and therapy of ENT disorders. Our remit includes the normal development, function and diseases of the ear, nose and throat and related aspects of hearing, balance, smell and speech. The NIHR ENT Specialty Group strongly supports audiology research activities and welcomes queries from interested researchers (www.crn.nihr.ac.uk/ent/). Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 113 92. UK validation of the Tinnitus Functional Index (TFI) in a large research population K. Fackrell*§, D. A. Hall*§, J. G. Barry‡ and D. J. Hoare*§, *NIHR Nottingham Hearing Biomedical Research Unit. § University of Nottingham, ‡ MRC Institute of Hearing Research, Nottingham, UK Questionnaires are essential for measurement of tinnitus severity and treatment-related change but there is no standard measure used across clinical and research settings. Current tinnitus questionnaires are limited to measure either severity or change, but not both. The Tinnitus Functional Index (TFI; Meikle et al, 2012) was developed to be used as both a diagnostic measure of the functional impact of tinnitus and to be a sensitive measure of treatment-related change. In this first study, we evaluate validity of the TFI as a diagnostic measure of tinnitus severity in a UK research population. 283 participants completed a series of questionnaires; the TFI, the Tinnitus Handicap Inventory (THI), Tinnitus Handicap Questionnaire (THQ), tinnitus loudness VAS, tinnitus annoyance percentage scale, Beck’s Depression Inventory (BDI), Beck’s Anxiety Inventory (BAI), and the World Health Organisation Quality of Life Bref (WHOQOL-BREF).One hundred participants completed the TFI at a second visit. Analyses included (1) Exploratory Factor Analysis (EFA) to identify interrelationships without predefining a possible structure; (2) Confirmatory factor analysis (CFA) using the current eight subscales proposed for the TFI development and the newly identified EFA structure; (3) Discriminant and convergent validity; and (4) Test-retest reliability and agreement. EFA indicated that two questions loaded onto factors other than the factor they are currently associated with (Meikle et al, 2012). For CFA, although the approximate fit indices were within an acceptable range (RMSEA =0.6, CFI > 0.95), X2 was still significant (p<0.001), indicating the current overall TFI structure is not optimal. Convergent and discriminant validity of the TFI revealed high correlations with the THI (r = 0.82) and THQ (r =0.82), moderate correlations with a tinnitus loudness VAS (r =0.46) and tinnitus annoyance scale (r =0.58) and moderate correlations with the BDI (r =0.57), BAI (r = 0.38) and WHOQOL (r = -0.48). Test-retest reliability was extremely high (ICC =0.86), whilst test agreement was 93%. Overall, the TFI does appear to reliably measure tinnitus and multiple symptoms domains. However, at this point we cannot confirm the proposed TFI structure. We are currently collecting large scale data in a clinical population to further this analysis and evaluate sensitivity of the TFI to treatment-related change. Acknowledgements This work is part of a NIHR funded PhD studentship. References Meikle, M. B. et al. 2012. The Tinnitus Functional Index: Development of a New Clinical Measure for Chronic, Intrusive Tinnitus. Ear Hear, 33(2), 153-176. 93. Investing in the UK tinnitus support group network D. Stockdale and C. Arthur, British Tinnitus Association The aim of the project was to increase and improve the support available to people living with chronic tinnitus in the UK by strengthening a network of tinnitus support groups around the UK. Specifically, to reverse a trend of decline by aiming to increase the number of groups from 34 to 46 within a two year period, and to support these groups to be run as effectively as possible. 114 Following initial scoping work, investigating the strategies of similar organisations in this area, the British Tinnitus Association (BTA) increased its organisational capacity and skills to better support groups by recruiting a Groups Coordinator. Research was undertaken to examine factors which might affect participation in support groups. Different models of groups were also assessed. The BTA increased the promotion of the value of organising a support group, and highlighted the increased support available from the BTA which was developed in line with identified needs. Events were organised by the BTA including workshops for new group organisers, and a national groups skill sharing and network event. Following initial research, the BTA improved its understanding of the range and activities of groups and their organisers: Identified 3 main types of group: purely voluntary (traditional ‘self help’ groups); groups run by healthcare professionals; groups run by local charities. Identified that the two key drivers of groups folding were declining membership and a difficulty replacing outgoing committee members. Identified a shared trend of decline across many types of local or community groups and a general consensus amongst related networks that changing social patterns had a role notably, that people were increasingly seeking out communities online and less often doing so in their locality; that changing work patterns led to reduced resource for voluntary projects. Identified that key barriers for potential organisers included low confidence, a perceived shortage of technical knowledge, expectations about the time and paperwork involved. In contrast to the decline trend, initial work also identified a good number of potential group organisers. Voluntary, or ‘self help’, and charity-led groups have seen the greatest increases, whilst NHS healthcare professionals appear to face more hurdles getting a new group off the ground. Year one of investment in group support led to a reverse in the decline trend. Group numbers were up 50% by the end of September 2013 from 34 to 52, exceeding the 24 month target of 46 groups. Developments in support offered to groups have been well received and now need further refining and embedding in organisational processes. Success in growing and strengthening the groups network demonstrates the value of national organisations investing in these types of activities. Declaration of interest The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. The End 115