Hidden Bibliographic Details
Other authors / contributors: | Karakos, Damianos.
Nguyen, Patrick.
Linguistic Data Consortium.
|
ISBN: | 1585635782 9781585635788
|
Notes: | Title from disc label. "Authors, Geoffrey Zweig, Damianos Karakos and Patrick Nguyen"--LDC catalog. Data type: text. Data source: broadcast news. Data application: speech recognition. "Developed by researchers at Microsoft and Johns Hopkins Unviersity (JHU) for the Johns Hopkins 2010 Summer Workshop on Speech Recognition with Conditional Random Fields. The lattices were generated using the IBM Attila speech recognition toolkit and were derived from transcripts of approximately 400 hours of English broadcast news recordings. They are intended to be used for training and decoding with Microsoft's segmental CRF toolkit for speech recogntion, SCARF."--Index.html. "LDC2011T06." Also available on the Internet.
|
Summary: | Boadcast News Lattices, Linguistic Data Consortium (LDC) catalog number LDC2011T06 and isbn 1-58563-578-2, was developed by researchers at Microsoft and Johns Hopkins Unviersity (JHU) for the Johns Hopkins 2010 Summer Workshop on Speech Recognition with Conditional Random Fields. The lattices were generated using the IBM Attila speech recognition toolkit and were derived from transcripts of approximately 400 hours of English broadcast news recordings. They are intended to be used for training and decoding with Microsofts segmental CRF toolkit for speech recogntion, SCARF. The goal of the JHU 2010 workshop was to advance the state-of-the-art in core speech recognition by developing new kinds of features for use in a Segmental Conditional Random Field (SCRF). The SCRF approach generalizes Condtional Random Fields to operate at the segment level, rather than at the traditional frame level. Every segment is labeled directly with a word. Features are then extracted which each measure some form of consistency between the underlying audio and the word hypothesis for a segment. These are combined in a log-linear model (lattice) to produce the posterior possibility of a word sequence given the audio.
|