The Centre for Speech Technology Research, The university of Edinburgh

Publications by Karl Isaac

s0976649.bib

@inproceedings{Wolters:2012:HTS:2212776.2223703,
  author = {Wolters, Maria and Isaac, Karl and Doherty, Jason},
  numpages = {6},
  publisher = {ACM},
  doi = {10.1145/2212776.2223703},
  isbn = {978-1-4503-1016-1},
  title = {Hold that thought: are spearcons less disruptive than spoken reminders?},
  url = {http://doi.acm.org/10.1145/2212776.2223703},
  series = {CHI EA '12},
  booktitle = {CHI '12 Extended Abstracts on Human Factors in Computing Systems},
  acmid = {2223703},
  location = {Austin, Texas, USA},
  year = {2012},
  keywords = {irrelevant speech effect, reminders, spearcon, speech, working memory},
  pages = {1745--1750},
  address = {New York, NY, USA}
}
@inproceedings{wolters2010,
  author = {Wolters, Maria K. and Isaac, Karl B. and Renals, Steve},
  title = {Evaluating speech synthesis intelligibility using {Amazon Mechanical Turk}},
  booktitle = {Proc. 7th Speech Synthesis Workshop (SSW7)},
  abstract = {Microtask platforms such as Amazon Mechanical Turk (AMT) are increasingly used to create speech and language resources. AMT in particular allows researchers to quickly recruit a large number of fairly demographically diverse participants. In this study, we investigated whether AMT can be used for comparing the intelligibility of speech synthesis systems. We conducted two experiments in the lab and via AMT, one comparing US English diphone to US English speaker-adaptive HTS synthesis and one comparing UK English unit selection to UK English speaker-dependent HTS synthesis. While AMT word error rates were worse than lab error rates, AMT results were more sensitive to relative differences between systems. This is mainly due to the larger number of listeners. Boxplots and multilevel modelling allowed us to identify listeners who performed particularly badly, while thresholding was sufficient to eliminate rogue workers. We conclude that AMT is a viable platform for synthetic speech intelligibility comparisons.},
  year = {2010},
  pdf = {http://www.cstr.inf.ed.ac.uk/downloads/publications/2010/wolters-ssw2010.pdf},
  pages = {136--141},
  categories = {intelligibility, evaluation, semantically unpredictable sentences, diphone, unit selection, crowd- sourcing, Mechanical Turk, HMM-based synthesis}
}
@inproceedings{Wolters2011,
  author = {Wolters, Maria Klara and Johnson, Christine and Isaac, Karl B},
  title = {Can the Hearing Handicap Inventory for Adults Be Used As a Screen for Perception Experiments?},
  booktitle = {Proc. ICPhS XVII},
  year = {2011},
  address = {Hong Kong},
  pdf = {http://www.cstr.inf.ed.ac.uk/downloads/publications/2011/Wolters_icphs.pdf},
  abstract = {When screening participants for speech perception experiments, formal audiometric screens are often not an option, especially when studies are conducted over the Internet. We investigated whether a brief standardized self-report questionnaire, the screening version of the Hearing Handicap Inventory for Adults (HHIA-S), could be used to approximate the results of audiometric screening. Our results suggest that while the HHIA-S is useful, it needs to be used with extremely strict cut-off values that could exclude around 25\% of people with no hearing impairment who are interested in participating. Well constructed, standardized single questions might be a more feasible alternative, in particular for web experiments.},
  categories = {audiometry,hearing handicap inventory,screening}
}