|
{ |
|
"paper_id": "O05-3005", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T07:58:43.150570Z" |
|
}, |
|
"title": "TAICAR -The Collection and Annotation of an In-Car Speech Database Created in Taiwan", |
|
"authors": [ |
|
{ |
|
"first": "Hsien-Chang", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Jung Christian University", |
|
"location": { |
|
"addrLine": "396 Chang Jung Road, Sec.1", |
|
"settlement": "Kway Jen, Tainan", |
|
"country": "Taiwan, R.O.C" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Chung-Hsien", |
|
"middle": [], |
|
"last": "Yang", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Jung Christian University", |
|
"location": { |
|
"addrLine": "396 Chang Jung Road, Sec.1", |
|
"settlement": "Kway Jen, Tainan", |
|
"country": "Taiwan, R.O.C" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Jhing-Fa", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Jung Christian University", |
|
"location": { |
|
"addrLine": "396 Chang Jung Road, Sec.1", |
|
"settlement": "Kway Jen, Tainan", |
|
"country": "Taiwan, R.O.C" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Chung-Hsien", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Jung Christian University", |
|
"location": { |
|
"addrLine": "396 Chang Jung Road, Sec.1", |
|
"settlement": "Kway Jen, Tainan", |
|
"country": "Taiwan, R.O.C" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Jen-Tzung", |
|
"middle": [], |
|
"last": "Chien", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Jung Christian University", |
|
"location": { |
|
"addrLine": "396 Chang Jung Road, Sec.1", |
|
"settlement": "Kway Jen, Tainan", |
|
"country": "Taiwan, R.O.C" |
|
} |
|
}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "This paper describes a project that aims to create a Mandarin speech database for the automobile setting (TAICAR). A group of researchers from several universities and research institutes in Taiwan have participated in the project. The goal is to generate a corpus for the development and testing of various speech-processing techniques. There are six recording sites in this project. Various words, sentences, and spontaneously queries uttered in the vehicular navigation setting have been collected in this project. A preliminary corpus of utterances from 192 speakers was created from utterances generated in different vehicles. The database contains more than 163,000 files, occupying 16.8 gigabytes of disk space.", |
|
"pdf_parse": { |
|
"paper_id": "O05-3005", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "This paper describes a project that aims to create a Mandarin speech database for the automobile setting (TAICAR). A group of researchers from several universities and research institutes in Taiwan have participated in the project. The goal is to generate a corpus for the development and testing of various speech-processing techniques. There are six recording sites in this project. Various words, sentences, and spontaneously queries uttered in the vehicular navigation setting have been collected in this project. A preliminary corpus of utterances from 192 speakers was created from utterances generated in different vehicles. The database contains more than 163,000 files, occupying 16.8 gigabytes of disk space.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Driver information systems are becoming increasingly complex as more and more functions are integrated into modern cars. Speech-enabled functions will enhance the safety and convenience of operating for future vehicles. To realize such functions, in-car speech processing techniques need to be built and tested first. Thus, it is necessary to collect an in-car speech database. Although many speech corpora [Tapisa et al. 1994] , [Roach et al. 1996] , [Kudo et al. 1994] , [Bernstein et al. 1994] have been created to improve speech-processing effectiveness, few in-car speech databases have been reported.", |
|
"cite_spans": [ |
|
{ |
|
"start": 407, |
|
"end": 427, |
|
"text": "[Tapisa et al. 1994]", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 430, |
|
"end": 449, |
|
"text": "[Roach et al. 1996]", |
|
"ref_id": "BIBREF7" |
|
}, |
|
{ |
|
"start": 452, |
|
"end": 470, |
|
"text": "[Kudo et al. 1994]", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 473, |
|
"end": 496, |
|
"text": "[Bernstein et al. 1994]", |
|
"ref_id": "BIBREF0" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "In-car speech corpora review", |
|
"sec_num": "1.1" |
|
}, |
|
{ |
|
"text": "Researches on speech processing in the vehicular environment, including works on speech recognition, noise reduction and speaker adaptation, have been published at numerous conferences, for example, the International Workshop on Hand-Free Speech Communication, which was held in 2001 in Kyoto, Japan; the biannual European Conference on Speech Communication and Technology (EuroSpeech) ; and the International Conference on Spoken Language Processing (ICSLP) . To our knowledge, several research organizations have carried out in-car speech database collection. In Japan, professor Itakura at CIAIR collected multimedia data, such as audio, video, and auxiliary vehicle information, from dialogues spoken in moving cars [Itakura 2001] . The system was built in a Data Collection Vehicle (DCV) supporting the synchronous recording of multi-channel audio and video data through microphones and cameras. In Europe, researchers in countries such as France, Germany, Britain, and Spain joined in a cooperative project, SpeechDat to collect an in-car speech database for multi-lingual speech processing purposes. The resulting SpeechDat-Car database contains speech data recorded from three microphones and one cellular phone. A similar project has also been reported by Langmann and his colleagues. [Langmann et al. 1998 ]. Researchers at the University of Illinois in Champaign-Urbana designed a project whose purpose was to collect multi-channel database consisting of both speech and video data. One hundred speakers participated in the project, and a total of 59,000 utterances were collected [Lee et al. 2004] . Table 1 shows a brief comparison of some existing in-car speech corpora and the TAICAR corpus. ", |
|
"cite_spans": [ |
|
{ |
|
"start": 344, |
|
"end": 385, |
|
"text": "Communication and Technology (EuroSpeech)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 451, |
|
"end": 458, |
|
"text": "(ICSLP)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 720, |
|
"end": 734, |
|
"text": "[Itakura 2001]", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 1294, |
|
"end": 1315, |
|
"text": "[Langmann et al. 1998", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 1592, |
|
"end": 1609, |
|
"text": "[Lee et al. 2004]", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 1612, |
|
"end": 1619, |
|
"text": "Table 1", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "In-car speech corpora review", |
|
"sec_num": "1.1" |
|
}, |
|
{ |
|
"text": "A group of researchers in the field of speech processing in Taiwan initiated an in-car speech collection project called TAICAR (Taiwan in-CAR speech database). The goal is to generate an in-car speech database to be applied to various noisy speech processing researches. In order to generate the corpus rapidly and usefully, some considerations with regard to setting up the data collection procedure were deemed important. These considerations are described below.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation and Setup", |
|
"sec_num": "1.2" |
|
}, |
|
{ |
|
"text": "The philosophy behind the TaiCar corpus collection procedure is to use convenient and readily available equipment to collect speech and environmental noise in various vehicles. The following are the ten considerations deemed important.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Setup of the TaiCar project", |
|
"sec_num": "1.3" |
|
}, |
|
{ |
|
"text": "1. The platform for in-car speech collection should be a notebook PC. 8. In addition to the speech data, the corpus should also include environmental noises.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Setup of the TaiCar project", |
|
"sec_num": "1.3" |
|
}, |
|
{ |
|
"text": "9. The database should reflect two real-world road conditions of the real world. The vehicle should be routed through a downtown area and along a highway during a recording session.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Setup of the TaiCar project", |
|
"sec_num": "1.3" |
|
}, |
|
{ |
|
"text": "10. The database should contain some spontaneous sentences to facilitate research on mobile dialogue systems.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Setup of the TaiCar project", |
|
"sec_num": "1.3" |
|
}, |
|
{ |
|
"text": "This paper is organized as follows. Section 2 describes the recording procedure. Section 3 presents the annotating procedure. Preliminary results of the TaiCar project are given in Section 4, and Section 5 gives a conclusion.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Setup of the TaiCar project", |
|
"sec_num": "1.3" |
|
}, |
|
{ |
|
"text": "In this project, six recording sites at universities and research institutions have been set up so far across Taiwan. Each site uses a notebook PC equipped with a PCMCIA multi-channel signal-recording card as the recording platform. A pre-amplification circuit amplifies the input signals, which go to the recording card from the microphones. Six microphones are placed in the vehicle. A microphone array with four omni-microphones is placed on the sun visors. The distance between the microphones is 30 cm. Another microphone is bound above the notebook PC placed on the lap of the speaker. Due to safety considerations, the speaker should be the navigator instead of the driver. The last microphone, a unidirectional anti-noise one, is worn on the head of the speaker. The reason for using such a good microphone is to provide nearly clean speech for reference purposes. The hardware elements are described in detail below: 1. A DAQP PCMCIA multi-channel signal recording card capable of recording up to 16 channels of signal is plugged into the notebook PC as the recording interface.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data collection system", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "2. Four omni-directional microphones form a linear microphone array (channels 0-3).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data collection system", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "3. One omni-directional microphone is placed in front of the speaker (channel 4).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data collection system", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "4. One unidirectional microphone is worn on the head of the speaker (channel 5).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data collection system", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "5. A pre-amplification circuit is utilized before the speech signal is fed to the PCMCIA card. Figure 1 shows the configuration and the positioning of the microphone array, the navigator, and the pre-amplification circuit. ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 95, |
|
"end": 103, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Data collection system", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "During the recording process, the notebook PC is placed on the lap of the navigator. The material to be uttered is shown on the screen in prompts so that the speaker can follow. A sample screenshot captured during the recording procedure is shown in Figure 2 . ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 250, |
|
"end": 258, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Speech Database Created in Taiwan", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "For each utterance, six speech files are recorded. The files, saved in the MS-Windows file format for audio waveforms, are composed of two parts: a file header and sampled data. The file header contains the following information about the speech: 1) the number of channels, which indicates whether the speech was recorded in mono or stereo; 2) the number of samples recorded per second; 3) the number of bits per sample; and 4) the size of the speech data. The sampled data of speech signals are in the binary format. The files retain the waveforms of the recorded utterances as well as the preceding and following silence.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speech Files Format", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Unlike several existing speech databases, for example, MAT [Wang 1997 ], the transcribed Chinese characters are stored in separate files using Big-5 code. This makes it convenient to preview these files using common text processing programs under most operating systems.", |
|
"cite_spans": [ |
|
{ |
|
"start": 59, |
|
"end": 69, |
|
"text": "[Wang 1997", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speech Files Format", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "The TAICAR database material contains two parts. The first part is used to collect the reading speech of the speakers. It is generated by following the philosophy of the creation of MAT-2400 database material [Wang 1997 ]. The framework for this material was created by Dr. Tseng of Academia Sinica [Tseng 1995] . The materials were extracted from two text corpora consisting of 77,324 lexical entries and 5,353 sentences. The material contains 407 base-syllables in Mandarin Chinese without tones; 1,062 words with two to four syllables; and 200 numbers in five different contexts, including digital sequences, dates, time, prices, and car license plate numbers.", |
|
"cite_spans": [ |
|
{ |
|
"start": 209, |
|
"end": 219, |
|
"text": "[Wang 1997", |
|
"ref_id": "BIBREF10" |
|
}, |
|
{ |
|
"start": 299, |
|
"end": 311, |
|
"text": "[Tseng 1995]", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Corpus Design", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "The second part consists of spontaneous FAQ's (Frequently Asked Questions) collected from the general public in Taiwan. This material was generated by asking them several questions. The scenario questions were given to ordinary citizens, and their answers were transcribed and used as the material for the spontaneous FAQ's. The scenario questions include a description of seven query domains containing questions which are usually asked while driving a car. The seven query domains and some collected FAQ's are listed in Table 2 . The collected FAQ's were randomly chosen to be included in the TAICAR prompt sheets. Each prompt sheet contains 10 FAQ's that the speaker utters spontaneously.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 522, |
|
"end": 529, |
|
"text": "Table 2", |
|
"ref_id": "TABREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Corpus Design", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "The prompt sheets are designed to serve as guides for the speaker to follow while uttering speech. The prompt sheet contains two parts of the aforementioned materials--the spontaneous speech and FAQ's. A total of 72 items are listed on a prompt sheet. The items are: ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Prompt Sheet", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "The prompt sheet is designed to contain as many syllable and phonetic combinations as possible. The FAQ's are also included on the prompt sheet since they are useful for research of vehicular dialogue systems. An example of a prompt sheet is shown in Appendix A.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speech Database Created in Taiwan", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "For a speech corpus to be useful, various phenomena of speaker behaviour and the deficiencies of the speech files should be annotated correctly. Since the annotation of a speech database is a labour consuming task, the tagging procedure for the TAICAR database was designed to be as convenient as possible. In the annotation phase, the annotators check whether the speech files are intelligible and whether the auto-transcribed syllables match the speaker's utterances, and they mark the starting and ending points of the speech. Figure 3 shows a screenshot of the annotation process. If the starting or ending point of an utterance does not match the syllables, the annotator should mark another boundary of the utterance and correct both the text content and phonetic syllables in the database.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 530, |
|
"end": 538, |
|
"text": "Figure 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "The Annotating Process", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "The TAICAR project was carried out between 2002 and 2003. According to the initial plan, researchers at each recording site would record the speech of 40 speakers and annotate the utterances. However, for technical and financial reasons, researchers at some sites did not complete these tasks. In all, 192 speakers at the six recording sites participated in this project. The result was an in-car speech database consisting of utterances recorded in both downtown and highway environments. Since it is hard to accurately read long sentences on a screen while driving, utterances consisting of FAQ sentences were collected at only one site. Some statistics for the resulting database are shown in Table 3 . Note that the number of files or contents is for 192 speakers driving along two different routes with six recording-channels. Figure 4 shows the waveforms of the utterance \"EQ7673\" from channel 0 to channel 5. As mentioned in Section 2, the microphone for channel 5 is unidirectional and anti-noise. It is adopted to record the reference signal for calculating the signal-to-noise ratio (SNR) and the time shift for other channels. The SNR can be computed as follows:", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 696, |
|
"end": 703, |
|
"text": "Table 3", |
|
"ref_id": "TABREF3" |
|
}, |
|
{ |
|
"start": 832, |
|
"end": 840, |
|
"text": "Figure 4", |
|
"ref_id": "FIGREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Preliminary Data Collection Result", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "10 [ ] 10 log [ ] E speech SNR E noise = \u22c5 , (in dB)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Preliminary Data Collection Result", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "where E[x] stands for the energy of signal x.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Preliminary Data Collection Result", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "To estimate the SNR for M 5 , the speech region is detected first. Then the noise can be estimated from the non-speech part. Based on the estimated noise level, the average SNR in the speech region can be determined. By aligning the signal of M 5 with the signals of M 0~M4 , one can locate the speech regions in M 0~M4 . Then the noise level and SNRs for M 0~M4 can be computed. The SNRs for different routes measured in the downtown and highway environments are reported in Table 4 . To calculate the time shift between channel k (0\u2264 k \u22644) and channel 5, the configuration of all six microphones should be considered first, as shown in Figure 5 . The microphone for channel k is M k . The distance between M i and M j is D i,j. The distances for D 3,5 and D 0,5 are predefined as 40 cm and 60 cm, respectively. The distance between the microphones in the microphone array is 30 cm, i.e. D 1,2 = D 2,3 = D 3,4 = 30 cm. Applying the Pythagorean Theorem, we can calculate the distances D 1,5 , D 2,5 , and D 4,5 obtaining 72, 50, and 50 cm, respectively. Because the speed of sound is 32,000 cm/sec, the time shift between M i and M j (0\u2264i,j\u22644), denoted as T i,j , can be determined. Since M 5 is placed in front of the mouth of the speaker, it can be regarded as the original source of the utterance. The time shift for each channel can be determined as T 0,5 =0.00187, T 1,5 =0.00156, T 2,5 =0.00125, T 3,5 =0.00156, and T 4,5 =0.00221. ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 476, |
|
"end": 483, |
|
"text": "Table 4", |
|
"ref_id": "TABREF4" |
|
}, |
|
{ |
|
"start": 638, |
|
"end": 646, |
|
"text": "Figure 5", |
|
"ref_id": "FIGREF5" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Preliminary Data Collection Result", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "This paper has described the TAICAR project that aims to create a Mandarin Chinese speech database based on the in-car environment in Taiwan. The preliminary result is a 192-speaker speech database containing 145.8 hours of utterances and environmental noises recorded in various types of automobiles. So far, two works have adopted the TaiCar corpus studies on speech enhancement in car noise environment [Yang et al. 2004] , [Wang et al. 2004] and have", |
|
"cite_spans": [ |
|
{ |
|
"start": 406, |
|
"end": 424, |
|
"text": "[Yang et al. 2004]", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 427, |
|
"end": 445, |
|
"text": "[Wang et al. 2004]", |
|
"ref_id": "BIBREF11" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "shown that the use of this corpus is of fundamental importance for the testing of in-car noise reduction technology. The database can also be used to develop various in-car speech processing techniques, such as speech source separation, active speech detection, channel equalization, and robust noisy speech recognition.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speech Database Created in Taiwan", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "The authors would like to express their appreciation to the researchers from National Taiwan University, National Chiao-Tung University, National Tsing-Hua University, National Cheng-Kung University, Industrial Technology Research Institute (ITRI), and Chunghwa Telecom Laboratories (CTL). Without their help, the TaiCar project would not have been possible.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgement", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "MACROPHONE: An American English Telephone Speech Corpus for Polyphone Project", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Bernstein", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Taussig", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Godfrey", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "Proceedings of the IEEE International Conference on Acoustics Speech and Signal Processing", |
|
"volume": "I", |
|
"issue": "", |
|
"pages": "81--84", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Bernstein, J., K. Taussig and J. Godfrey, \"MACROPHONE: An American English Telephone Speech Corpus for Polyphone Project,\" In Proceedings of the IEEE International Conference on Acoustics Speech and Signal Processing, 1994, Adelaide, Australia, Vol. I, pp. 81-84.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "SpeechDat-Car: Towards a collection of speech databases for automotive environments", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Heuvel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Bonafonte", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Boudy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Dufour", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Lockwood", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Moreno", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Richard", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "Proceedings of the Nokia-COST249 Workshop on Robust Methods for Speech Recognition in Adverse Conditions", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "135--138", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Heuvel, H., A. Bonafonte , J. Boudy, S. Dufour, P. Lockwood, A. Moreno and G. Richard, \"SpeechDat-Car: Towards a collection of speech databases for automotive environments,\" In Proceedings of the Nokia-COST249 Workshop on Robust Methods for Speech Recognition in Adverse Conditions, 1999, Tampere, Finland, pp. 135-138.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "The SpeechDat-Car multilingual speech databases for in-car applications: Some first validation results", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Heuvel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Boudy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Comeyne", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Euler", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Moreno", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Richard", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "Proceedings of 6 th European Conference on Speech Communication and Technology", |
|
"volume": "5", |
|
"issue": "", |
|
"pages": "2279--2282", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Heuvel, H., J. Boudy, R. Comeyne, S. Euler, A. Moreno and G. Richard, \"The SpeechDat-Car multilingual speech databases for in-car applications: Some first validation results,\" In Proceedings of 6 th European Conference on Speech Communication and Technology, 1999, Budapest, Hungary, Vol.5, pp. 2279-2282.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Multi-Media Data Collection for In-Car Speech Communication -Ongoing Data Collection and Preliminary Results", |
|
"authors": [ |
|
{ |
|
"first": "F", |
|
"middle": [], |
|
"last": "Itakura", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Proceedings of International Workshop on Hand-Free Speech communication", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--5", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Itakura, F., \"Multi-Media Data Collection for In-Car Speech Communication -Ongoing Data Collection and Preliminary Results,\" In Proceedings of International Workshop on Hand-Free Speech communication, 2001, Kyoto, Japan, pp. 1-5.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "The Database Collection of Voice Across Japan (VAJ) Project", |
|
"authors": [ |
|
{ |
|
"first": "I", |
|
"middle": [], |
|
"last": "Kudo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Nakama", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Arai", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Fujimura", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "Proceedings of 2 nd International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1799--1802", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kudo, I., T. Nakama, N. Arai and N. Fujimura, \"The Database Collection of Voice Across Japan (VAJ) Project,\" In Proceedings of 2 nd International Conference on Spoken Language Processing, 1994, Yokohama, Japan, pp.1799-1802.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "CSDC, the MoTiV Car Speech Data Collection", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Langmann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [ |
|
"R" |
|
], |
|
"last": "Pfitzinger", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Schneider", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Grudszus", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Fischer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Westphal", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Crull", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "U", |
|
"middle": [], |
|
"last": "Jekosch", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of 1 st International Conference on Language, Resources and Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1107--1110", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Langmann, D., H.R. Pfitzinger, T. Schneider, R. Grudszus, A. Fischer, M. Westphal, T. Crull and U. Jekosch, \"CSDC, the MoTiV Car Speech Data Collection,\" In Proceedings of 1 st International Conference on Language, Resources and Evaluation, 1998, Granada, Spain, pp. 1107-1110.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "AVICAR: Audio-Visual Speech Corpus in a Car Environment", |
|
"authors": [ |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H.-J", |
|
"middle": [], |
|
"last": "Mark", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Goudeseune", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Kamdar", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Borys", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Huang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of 8 th International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lee, B., H.-J. Mark, C. Goudeseune, S. Kamdar, S. Borys, M. Liu and T. Huang, \"AVICAR: Audio-Visual Speech Corpus in a Car Environment,\" In Proceedings of 8 th International Conference on Spoken Language Processing, 2004, Jeju Island, Korea.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "BABEL: An Eastern European Multi-language Database", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Roach", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Arnfield", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Barry", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Baltova", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Boldea", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Fourcin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Gonet", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Gubrynowicz", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "E", |
|
"middle": [], |
|
"last": "Hallum", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Lamel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Marasek", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Marchal", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "E", |
|
"middle": [], |
|
"last": "Meister", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Vicsi", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1996, |
|
"venue": "Proceedings of 4 th International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1892--1893", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Roach, P., S. Arnfield, W. Barry, J. Baltova, M. Boldea, A. Fourcin, W. Gonet, R. Gubrynowicz, E. Hallum,L. Lamel, K. Marasek, A. Marchal, E. Meister and K. Vicsi, \"BABEL: An Eastern European Multi-language Database,\" In Proceedings of 4 th International Conference on Spoken Language Processing, 1996, Philadelphia, USA, pp. 1892-1893.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "The VESTEL Telephone Speech Database", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Tapias", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Acero", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Esteve", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Torrecilia", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "Proceedings of 3 rd International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1811--1814", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Tapias, D., A. Acero, J. Esteve and J.C. Torrecilia, \"The VESTEL Telephone Speech Database,\" In Proceedings of 3 rd International Conference on Spoken Language Processing, 1994, Yokohama, Japan, pp.1811-1814.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "A Phonetically Oriented Speech Database for Mandarin Chinese", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Tseng", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1995, |
|
"venue": "Proceedings of the 13 th International Congress on Phonetic Sciences", |
|
"volume": "3", |
|
"issue": "", |
|
"pages": "326--329", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Tseng, C.Y. \"A Phonetically Oriented Speech Database for Mandarin Chinese,\" In Proceedings of the 13 th International Congress on Phonetic Sciences, 1995, Stockholm, Sweden, Vol. 3, pp.326-329.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "MAT -A Project to Collect Mandarin Speech Data Through Telephone Networks in Taiwan", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Wang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1997, |
|
"venue": "Computational Linguistics and Chinese Language Processing", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "73--90", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wang, H.C., \"MAT -A Project to Collect Mandarin Speech Data Through Telephone Networks in Taiwan,\" Computational Linguistics and Chinese Language Processing, 2(1), 1997, pp. 73-90.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Using Perceptual Wavelet Decomposition and Subspace Tracking for Noise Removal in Car Environment", |
|
"authors": [ |
|
{ |
|
"first": "J.-F", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C.-H", |
|
"middle": [], |
|
"last": "Yang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K.-H", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of ROCLING XVI: Conference on Computational Linguistics and Speech Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wang, J.-F., C.-H. Yang and K.-H. Chang, \"Using Perceptual Wavelet Decomposition and Subspace Tracking for Noise Removal in Car Environment,\" In Proceedings of ROCLING XVI: Conference on Computational Linguistics and Speech Processing, 2004, Taipei, Taiwan.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Subspace Tracking for Speech Enhancement in Car Noise Environments", |
|
"authors": [ |
|
{ |
|
"first": "C.-H", |
|
"middle": [], |
|
"last": "Yang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J.-F", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K.-H", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of International Conference on Acoustic, Speech, and Signal Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yang, C.-H., J.-F. Wang and K.-H. Chang, \"Subspace Tracking for Speech Enhancement in Car Noise Environments,\" In Proceedings of International Conference on Acoustic, Speech, and Signal Processing, 2004, Quebec, Canada.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "(a) The configuration of TAICAR recording system. The distance between the microphones in the array is 30 cm. (b) The microphone array attached to the sun visor above; (c) the positions of the speaker and recording notebook PC; (d) the amplification circuit board for multi-channel recording.", |
|
"num": null |
|
}, |
|
"FIGREF1": { |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "A screenshot from the TAICAR database recording procedure", |
|
"num": null |
|
}, |
|
"FIGREF2": { |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "numbers spoken in different ways (No's. 10-14); 12 isolated Mandarin syllables (No's. 15-26); 45 isolated words (No's. 27-56, 67-82); 10 FAQ sentences (No's. 57-66).", |
|
"num": null |
|
}, |
|
"FIGREF3": { |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "waveform to mark the starting/ending point of the speech. (11) Update the database when tagging is finished.", |
|
"num": null |
|
}, |
|
"FIGREF4": { |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "Speech waveforms of the utterance \"EQ7637\" (in Mandarin), from channel 0 to channel 5.", |
|
"num": null |
|
}, |
|
"FIGREF5": { |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "The detailed configuration and distances between the six microphones", |
|
"num": null |
|
}, |
|
"TABREF0": { |
|
"type_str": "table", |
|
"content": "<table><tr><td>Corpus name (year)</td><td>CSDC-MoTiV (1998)</td><td>SpeechDat-Car (1999)</td><td>CU-Move (2000)</td><td>CIAIR-HCC (2001)</td><td>CMU (2001)</td><td>AVICAR (2004)</td><td>TAICAR (2004)</td></tr><tr><td>Country</td><td>Germany</td><td>Europe</td><td>USA</td><td>Japan</td><td>USA</td><td>USA</td><td>Taiwan</td></tr><tr><td># of People</td><td>641</td><td>N/A</td><td>N/A</td><td>ongoing</td><td>43</td><td>100</td><td>192</td></tr><tr><td>Microphone</td><td>Array</td><td>Array</td><td>Array</td><td>Mesh</td><td>Array</td><td>Array</td><td>Array</td></tr><tr><td>Content</td><td>Digits; Commands</td><td>Multi-lingual</td><td>Digits; Commands</td><td>Digits; Words; Sentences</td><td>Short words</td><td>Digits; Letters; Sentences</td><td>Digits; Words; FAQs</td></tr><tr><td>Need Specific Car</td><td>No</td><td>No</td><td>No</td><td>Yes</td><td>No</td><td>No</td><td>No</td></tr></table>", |
|
"html": null, |
|
"num": null, |
|
"text": "" |
|
}, |
|
"TABREF2": { |
|
"type_str": "table", |
|
"content": "<table><tr><td>Domain</td><td>Scenario</td><td>Collected FAQ</td></tr><tr><td>Food</td><td>\u2666 You are hungry and looking for a restaurant.</td><td/></tr><tr><td/><td>\u2666 You are looking for a place</td><td>\u2212 I would like to know the location of the</td></tr><tr><td>Lodging</td><td>to stay.</td><td>Hilton Hotel.</td></tr><tr><td/><td/><td>\u2212 Show me the nearby hotels.</td></tr><tr><td>Navigation</td><td>\u2666 You want to know how to get to a destination.</td><td>\u2212 How do I go to CKS airport? \u2212 Where is City Hall?</td></tr><tr><td colspan=\"2\">Entertainment \u2666 You want to have fun.</td><td>\u2212 How do I get to the nearest theater?</td></tr><tr><td/><td>\u2666 Weather conditions.</td><td>\u2212 What's the temperature in Taipei?</td></tr><tr><td>Others</td><td>\u2666 Other information one wants to know while driving.</td><td>\u2212 Is there any museum nearby? \u2212 Turn the CD player on.</td></tr></table>", |
|
"html": null, |
|
"num": null, |
|
"text": "" |
|
}, |
|
"TABREF3": { |
|
"type_str": "table", |
|
"content": "<table><tr><td/><td>Total:</td><td>192</td></tr><tr><td/><td>Male:</td><td>115 (59.8%)</td></tr><tr><td>Speakers</td><td>Female: Age:</td><td>77 (40.2%) from 19 to 58, mostly 20~30 (71.3%)</td></tr><tr><td/><td>Education:</td><td>most has BS degrees (89.6%)</td></tr><tr><td/><td colspan=\"2\">Daily Language: Taiwanese (64.6%)</td></tr><tr><td>Car</td><td colspan=\"2\">Type: Engine capacity: below 2.0L: (57.8%); 2.0~3.0L (37.5%) mostly sedans (71.3%)</td></tr><tr><td/><td>6 DVDs</td><td/></tr><tr><td>Speech data</td><td>163,890 files</td><td/></tr><tr><td>amount</td><td>16.8 gigabytes</td><td/></tr><tr><td/><td>145.8 hours</td><td/></tr><tr><td/><td>16,128 digits</td><td/></tr><tr><td/><td colspan=\"2\">4,608 English letters</td></tr><tr><td>Database content</td><td colspan=\"2\">27,648 Isolated syllables</td></tr><tr><td/><td colspan=\"2\">101,376 Words with two-four characters</td></tr><tr><td/><td>960 FAQ's</td><td/></tr></table>", |
|
"html": null, |
|
"num": null, |
|
"text": "" |
|
}, |
|
"TABREF4": { |
|
"type_str": "table", |
|
"content": "<table><tr><td/><td colspan=\"6\">Channel 0 Channel 1 Channel 2 Channel 3 Channel 4 Channel 5</td></tr><tr><td>Highway</td><td>-2.8550</td><td>-2.4770</td><td>-2.7171</td><td>-2.5318</td><td>-2.6763</td><td>11.4040</td></tr><tr><td colspan=\"2\">Downtown -2.6187</td><td>-2.2637</td><td>-2.0714</td><td>-2.4655</td><td>-2.5261</td><td>11.2245</td></tr></table>", |
|
"html": null, |
|
"num": null, |
|
"text": "" |
|
} |
|
} |
|
} |
|
} |