<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://www.hlt.inesc-id.pt/wiki/index.php?action=history&amp;feed=atom&amp;title=BN_Speech_Recognition</id>
	<title>BN Speech Recognition - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://www.hlt.inesc-id.pt/wiki/index.php?action=history&amp;feed=atom&amp;title=BN_Speech_Recognition"/>
	<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=BN_Speech_Recognition&amp;action=history"/>
	<updated>2026-05-16T14:08:19Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.41.0</generator>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=BN_Speech_Recognition&amp;diff=2792&amp;oldid=prev</id>
		<title>Meinedo at 11:16, 3 July 2006</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=BN_Speech_Recognition&amp;diff=2792&amp;oldid=prev"/>
		<updated>2006-07-03T11:16:43Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;Automatic Speech Recognition (ASR) can be described as the process of&lt;br /&gt;
machine transcriptioning into text the words uttered by human&lt;br /&gt;
speakers. If one wants to perform a query in a database of BN data it&lt;br /&gt;
is necessary to have a textual transcription or at least a detailed&lt;br /&gt;
description of the news content.&lt;br /&gt;
&lt;br /&gt;
State of the art ASR systems use hierarchical models to recognize speech. The input audio is broken into small chunks and the &amp;quot;acoustic model&amp;quot; of the ASR system attribute to each chunk a probability or likelihood of occurrence of a&lt;br /&gt;
basic sound (phoneme). Then a search guided by two models determines&lt;br /&gt;
the final sequence of words. The two models are the &amp;quot;lexicon&amp;quot;&lt;br /&gt;
which models the sequence of phonemes that form words and the &amp;quot;language model&amp;quot; which models the sequence of words of the language.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Image:audimus.block_diagram.png|center]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In the case of a BN task, training all these models involves large&lt;br /&gt;
amounts of data, both audio (for the &amp;quot;acoustic model&amp;quot;) and text&lt;br /&gt;
(for the &amp;quot;language model&amp;quot;). Coping with large amounts of acoustic&lt;br /&gt;
data is an issue which L2F is addressing, more&lt;br /&gt;
specifically developing efficient training methods for the &amp;quot;acoustic model&amp;quot;.&lt;/div&gt;</summary>
		<author><name>Meinedo</name></author>
	</entry>
</feed>