<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://www.hlt.inesc-id.pt/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Alberto</id>
	<title>HLT@INESC-ID - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://www.hlt.inesc-id.pt/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Alberto"/>
	<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/w/Special:Contributions/Alberto"/>
	<updated>2026-04-09T13:25:56Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.41.0</generator>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8608</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8608"/>
		<updated>2024-02-16T09:43:16Z</updated>

		<summary type="html">&lt;p&gt;Alberto: /* Research Interests */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;Inesc-id what='card' id='5b544c6b-f1bb-4007-b82a-414f2d1a24e2'&amp;gt;&amp;lt;/Inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree and the PhD degree on 2002 and 2007 respectively, both from the [http://www.upc.edu Technical University of Catalonia] (UPC). Currently, he is an Associate Professor at the [https://dei.tecnico.ulisboa.pt Department of Computer Science and Engineering (DEI)] of [http://www.ist.utl.pt Instituto Superior Técnico (IST)] and researcher at INESC-ID. He is the coordinator of the [http://www.hlt.inesc-id.pt Human LanguageTechnologies laboratory (HLT)] at INESC-ID and the deputy coordinator of the [https://fenix.tecnico.ulisboa.pt/cursos/meic-a Master in Computer Science and Engineering of IST]. He is also an IEEE Senior member, an ACM member, and an ISCA member.&lt;br /&gt;
&lt;br /&gt;
Alberto Abad has developed his research career in the area of human language technologies for more than 15 years. His research interests include robust speech recognition, speaker and language characterization, applied machine learning, health-care applications, and privacy-preserving speech processing. During these years, Alberto Abad has co-authored over 100 peer-reviewed publications in top-tier conferences and scientific journals of the field. He has been Principal Investigator of FCT funded project VITHEA and INESC-ID representative in the DIRHA European project, Biovisualspeech CMU-Portugal project, and TAPAS ITN. He is currently INESC-ID/IST PI at [https://www.inesc-id.pt/accelerat-ai/ Accelerat.ai project]. He has also been team member in various Spanish and Portuguese national and international projects, and team leader in international technological evaluation challenges achieving excellent results. He has supervised and co-supervised more than 25 M.Sc. thesis and 1 Ph.D. thesis (+5 on-going) and lectured in courses at both UPC and IST on a variety of topics, including signal processing, machine learning, programming principles, object oriented programming and compilers. Alberto Abad has been actively involved in research community activities, including conference organization, and closely collaborated with local industry and start-up companies.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in many international technological evaluation challenges achieving excellent results. He  also collaborated in several courses of the [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon. In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where was lecturer of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) during the academic years 2013- 2014 and 2014-2015. In February 2015, Alberto Abad won a position as Assistant Professor at the [https://fenix.tecnico.ulisboa.pt/departamentos/dei Department of Computer Science and Engineering (DEI)] of the IST.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language characterization&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
* Privacy-preserving speech processing&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV2021.pdf]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;Inesc-id what='publications' id='5b544c6b-f1bb-4007-b82a-414f2d1a24e2'&amp;gt;&amp;lt;/Inesc-id&amp;gt;&lt;br /&gt;
&amp;lt;Inesc-id what='advising' id='5b544c6b-f1bb-4007-b82a-414f2d1a24e2'&amp;gt;&amp;lt;/Inesc-id&amp;gt;&lt;br /&gt;
&amp;lt;Inesc-id what='projects' id='5b544c6b-f1bb-4007-b82a-414f2d1a24e2'&amp;gt;&amp;lt;/Inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[category: People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8607</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8607"/>
		<updated>2024-02-16T09:42:12Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;Inesc-id what='card' id='5b544c6b-f1bb-4007-b82a-414f2d1a24e2'&amp;gt;&amp;lt;/Inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree and the PhD degree on 2002 and 2007 respectively, both from the [http://www.upc.edu Technical University of Catalonia] (UPC). Currently, he is an Associate Professor at the [https://dei.tecnico.ulisboa.pt Department of Computer Science and Engineering (DEI)] of [http://www.ist.utl.pt Instituto Superior Técnico (IST)] and researcher at INESC-ID. He is the coordinator of the [http://www.hlt.inesc-id.pt Human LanguageTechnologies laboratory (HLT)] at INESC-ID and the deputy coordinator of the [https://fenix.tecnico.ulisboa.pt/cursos/meic-a Master in Computer Science and Engineering of IST]. He is also an IEEE Senior member, an ACM member, and an ISCA member.&lt;br /&gt;
&lt;br /&gt;
Alberto Abad has developed his research career in the area of human language technologies for more than 15 years. His research interests include robust speech recognition, speaker and language characterization, applied machine learning, health-care applications, and privacy-preserving speech processing. During these years, Alberto Abad has co-authored over 100 peer-reviewed publications in top-tier conferences and scientific journals of the field. He has been Principal Investigator of FCT funded project VITHEA and INESC-ID representative in the DIRHA European project, Biovisualspeech CMU-Portugal project, and TAPAS ITN. He is currently INESC-ID/IST PI at [https://www.inesc-id.pt/accelerat-ai/ Accelerat.ai project]. He has also been team member in various Spanish and Portuguese national and international projects, and team leader in international technological evaluation challenges achieving excellent results. He has supervised and co-supervised more than 25 M.Sc. thesis and 1 Ph.D. thesis (+5 on-going) and lectured in courses at both UPC and IST on a variety of topics, including signal processing, machine learning, programming principles, object oriented programming and compilers. Alberto Abad has been actively involved in research community activities, including conference organization, and closely collaborated with local industry and start-up companies.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in many international technological evaluation challenges achieving excellent results. He  also collaborated in several courses of the [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon. In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where was lecturer of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) during the academic years 2013- 2014 and 2014-2015. In February 2015, Alberto Abad won a position as Assistant Professor at the [https://fenix.tecnico.ulisboa.pt/departamentos/dei Department of Computer Science and Engineering (DEI)] of the IST.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV2021.pdf]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;Inesc-id what='publications' id='5b544c6b-f1bb-4007-b82a-414f2d1a24e2'&amp;gt;&amp;lt;/Inesc-id&amp;gt;&lt;br /&gt;
&amp;lt;Inesc-id what='advising' id='5b544c6b-f1bb-4007-b82a-414f2d1a24e2'&amp;gt;&amp;lt;/Inesc-id&amp;gt;&lt;br /&gt;
&amp;lt;Inesc-id what='projects' id='5b544c6b-f1bb-4007-b82a-414f2d1a24e2'&amp;gt;&amp;lt;/Inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[category: People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8414</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8414"/>
		<updated>2022-09-09T08:23:02Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree and the PhD degree on 2002 and 2007 respectively, both from the [http://www.upc.edu Technical University of Catalonia] (UPC). Currently, he is an Associate Professor at the [https://dei.tecnico.ulisboa.pt Department of Computer Science and Engineering (DEI)] of [http://www.ist.utl.pt Instituto Superior Técnico (IST)] and researcher at INESC-ID. He is the coordinator of the [http://www.hlt.inesc-id.pt Human LanguageTechnologies laboratory (HLT)] at INESC-ID and the executive coordinator of the scientific area of [https://dei.tecnico.ulisboa.pt/sobre-o-departamento/areas-cientificas/ac-metodologia-e-tecnologias-de-programacao Programming Methodology and Technology (MTP)] at DEI. He is also an IEEE Senior member, an ACM member, an ISCA member, and an Associated Editor for the IEEE/ACM Transactions on Audio, Speech and Language Processing journal.&lt;br /&gt;
&lt;br /&gt;
Alberto Abad has developed his research career in the area of human language technologies for more than 15 years. His research interests include robust speech recognition, speaker and language characterization, applied machine learning, health-care applications, and privacy-preserving speech processing. During these years, Alberto Abad has co-authored over 100 peer-reviewed publications in top-tier conferences and scientific journals of the field. He has been Principal Investigator of FCT funded project VITHEA and INESC-ID representative in the DIRHA European project, Biovisualspeech CMU-Portugal project, and TAPAS ITN. He has also been team member in various Spanish and Portuguese national and international projects, and team leader in international technological evaluation challenges achieving excellent results. He has supervised and co-supervised more than 25 M.Sc. thesis and 1 Ph.D. thesis (+5 on-going) and lectured in courses at both UPC and IST on a variety of topics, including signal processing, machine learning, programming principles, object oriented programming and compilers. Alberto Abad has been actively involved in research community activities, including conference organization, and closely collaborated with local industry and start-up companies.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in many international technological evaluation challenges achieving excellent results. He  also collaborated in several courses of the [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon. In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where was lecturer of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) during the academic years 2013- 2014 and 2014-2015. In February 2015, Alberto Abad won a position as Assistant Professor at the [https://fenix.tecnico.ulisboa.pt/departamentos/dei Department of Computer Science and Engineering (DEI)] of the IST.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV2021.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8413</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8413"/>
		<updated>2022-09-09T08:12:29Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad is an Associate Professor at the [https://dei.tecnico.ulisboa.pt Department of Computer Science and Engineering (DEI)] of [http://www.ist.utl.pt Instituto Superior Técnico (IST)] and researcher at INESC-ID. He is the coordinator of the [http://www.hlt.inesc-id.pt Human LanguageTechnologies laboratory (HLT)] at INESC-ID and the executive coordinator of the scientific area of [https://dei.tecnico.ulisboa.pt/sobre-o-departamento/areas-cientificas/ac-metodologia-e-tecnologias-de-programacao Programming Methodology and Technology (MTP)] at DEI. He is also an IEEE Senior member, an ACM member, an ISCA member, and an Associated Editor for the IEEE/ACM Transactions on Audio, Speech and Language Processing journal.&lt;br /&gt;
&lt;br /&gt;
Alberto Abad has developed his research career in the area of human language technologies for more than 15 years. His research interests include robust speech recognition, speaker and language characterization, applied machine learning, health-care applications, and privacy-preserving speech processing. During these years, Alberto Abad has co-authored over 100 peer-reviewed publications in top-tier conferences and scientific journals of the field. He has been Principal Investigator of FCT funded project VITHEA and INESC-ID representative in the DIRHA European project, Biovisualspeech CMU-Portugal project, and TAPAS ITN. He has also been team member in various Spanish and Portuguese national and international projects, and team leader in international technological evaluation challenges achieving excellent results. He has supervised and co-supervised more than 25 M.Sc. thesis and 1 Ph.D. thesis (+5 on-going) and lectured in courses at both [http://www.upc.edu Technical University of Catalonia] (UPC) and IST on a variety of topics, including signal processing, machine learning, programming principles, object oriented programming and compilers. Alberto Abad has been actively involved in research community activities, including conference organization, and closely collaborated with local industry and start-up companies.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in many international technological evaluation challenges achieving excellent results. He  also collaborated in several courses of the [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon. In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where was lecturer of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) during the academic years 2013- 2014 and 2014-2015. In February 2015, Alberto Abad won a position as Assistant Professor at the [https://fenix.tecnico.ulisboa.pt/departamentos/dei Department of Computer Science and Engineering (DEI)] of the IST.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8134</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=8134"/>
		<updated>2020-03-10T23:31:37Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in many international technological evaluation challenges achieving excellent results. He  also collaborated in several courses of the [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon. In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where was lecturer of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) during the academic years 2013- 2014 and 2014-2015. In February 2015, Alberto Abad won a position as Assistant Professor at the [https://fenix.tecnico.ulisboa.pt/departamentos/dei Department of Computer Science and Engineering (DEI)] of the IST.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Tipo-passe-alberto.png&amp;diff=7907</id>
		<title>File:Tipo-passe-alberto.png</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Tipo-passe-alberto.png&amp;diff=7907"/>
		<updated>2017-02-01T11:24:57Z</updated>

		<summary type="html">&lt;p&gt;Alberto: Alberto uploaded a new version of File:Tipo-passe-alberto.png&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Tipo-passe-alberto.png&amp;diff=7906</id>
		<title>File:Tipo-passe-alberto.png</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Tipo-passe-alberto.png&amp;diff=7906"/>
		<updated>2017-02-01T11:18:10Z</updated>

		<summary type="html">&lt;p&gt;Alberto: Alberto uploaded a new version of File:Tipo-passe-alberto.png&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7773</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7773"/>
		<updated>2015-11-03T11:57:06Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad Gareta&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in many international technological evaluation challenges achieving excellent results. He  also collaborated in several courses of the [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon. In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where was lecturer of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) during the academic years 2013- 2014 and 2014-2015. In February 2015, Alberto Abad won a position as Assistant Professor at the [https://fenix.tecnico.ulisboa.pt/departamentos/dei Department of Computer Science and Engineering (DEI)] of the IST.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7374</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7374"/>
		<updated>2014-09-30T10:36:34Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad Gareta&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in many international technological evaluation challenges achieving excellent results. He  also collaborated in several courses of the [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon. In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where he has been lecturer of the course ''Arquitectura de Computadores'' of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) since the academic year 2013-2014.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7373</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7373"/>
		<updated>2014-09-30T10:36:16Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad Gareta&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in many international technological evaluation challenges achieving excellent results. He  also collaborated in several courses of the [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon.&lt;br /&gt;
&lt;br /&gt;
In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where he has been lecturer of the course ''Arquitectura de Computadores'' of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) since the academic year 2013-2014.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7372</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7372"/>
		<updated>2014-09-30T10:34:01Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad Gareta&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [http://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [http://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a ''Programa Ciência'' contract for 5 years. During that period at INESC-ID, Alberto Abad contributed to consolidate group's research lines and to open new strategic ones. He has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in several international technological evaluation challenges achieving excellent results. He  also collaborated in courses of the Electrical and Computer Engineering degree  of [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon.&lt;br /&gt;
&lt;br /&gt;
In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where he has been lecturer of the course ''Arquitectura de Computadores'' of the [http://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) since the academic year 2013-2014.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7371</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7371"/>
		<updated>2014-09-30T10:32:18Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad Gareta&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the [http://www.etsetb.upc.edu/ Telecommunication Engineering School of Barcelona] and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [https://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [https://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a &amp;quot;Programa Ciência&amp;quot; contract for 5 years. During that period at INESC-ID, Alberto Abad developed both basic research and managerial activities that contributed to consolidate group's research lines and to open new strategic ones. Alberto Abad has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in several international technological evaluation challenges achieving excellent results. He  also collaborated in courses of the Electrical and Computer Engineering degree  of [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon.&lt;br /&gt;
&lt;br /&gt;
In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where he has been lecturer of the course &amp;quot;Arquitectura de Computadores&amp;quot; of the [https://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) since the academic year 2013-2014.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7370</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7370"/>
		<updated>2014-09-30T10:31:01Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad Gareta&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the Telecommunication Engineering School of Barcelona and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the [https://www.l2f.inesc-id.pt Spoken Language Systems Lab (L2F)] of [https://www.inesc-id.pt/ INESC-ID] and in July 2008 he obtained a &amp;quot;Programa Ciência&amp;quot; contract for 5 years. During that period at INESC-ID, Alberto Abad developed both basic research and managerial activities that contributed to consolidate group's research lines and to open new strategic ones. Alberto Abad has been Principal Investigator of the FCT funded project [http://www.vithea.org VITHEA] and INESC-ID representative in the  [http://dirha.fbk.eu DIRHA] European project, besides team member in several national and international projects, and team leader in several international technological evaluation challenges achieving excellent results. He  also collaborated in courses of the Electrical and Computer Engineering degree  of [http://www.ist.utl.pt Instituto Superior Técnico (IST)] of the University of Lisbon.&lt;br /&gt;
&lt;br /&gt;
In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where he has been lecturer of the course &amp;quot;Arquitectura de Computadores&amp;quot; of the [https://fenix.tecnico.ulisboa.pt/cursos/meec Mestrado Integrado em Engenharia Electrotécnica e de Computadores] (MEEC) since the academic year 2013-2014.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7369</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7369"/>
		<updated>2014-09-30T10:26:06Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad Gareta&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he was a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments. At UPC he participated in several European and National Spanish projects. He was also lecturer in the Telecommunication Engineering School of Barcelona and collaborating lecturer in the doctorate program of the Department of Signal Theory and Communications.&lt;br /&gt;
&lt;br /&gt;
In February 2007, Alberto Abad joined the Spoken Language Systems Lab (L2F) of INESC-ID and in July 2008 he obtained a &amp;quot;Programa Ciência&amp;quot; contract for 5 years. During that period at INESC-ID, Alberto Abad developed both basic research and managerial activities that contributed to consolidate group's research lines and to open new strategic ones. Alberto Abad has been Principal Investigator of the FCT funded project -VITHEA- and INESC-ID representative in the  DIRHA European project, besides team member in several national and international projects, and team leader in several international technological evaluation challenges achieving excellent results. He  also collaborated in courses of the Electrical and Computer Engineering degree  of Instituto Superior Técnico (IST) of the University of Lisbon.&lt;br /&gt;
&lt;br /&gt;
In August 2013, Alberto Abad was hired as an Invited Researcher and Professor at the IST, where he has been lecturer of the course &amp;quot;Arquitectura de Computadores&amp;quot; of the &amp;quot;Mestrado Integrado em Engenharia Electrotécnica e de Computadores&amp;quot; (MEEC) since the academic year 2013-2014.&lt;br /&gt;
&lt;br /&gt;
His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis, multimedia, and health-care applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Robust Speech Recognition&lt;br /&gt;
* Speaker and language recognition&lt;br /&gt;
* Speech and language technology for health-care applications&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/CV_new_latex.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7198</id>
		<title>Alberto Abad Gareta</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Alberto_Abad_Gareta&amp;diff=7198"/>
		<updated>2013-12-19T10:33:36Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{infobox|name=Alberto Abad Gareta&lt;br /&gt;
|username=alberto&lt;br /&gt;
|contact=alberto.abad&lt;br /&gt;
|phone=+351-213-100-217&lt;br /&gt;
|fax=+351-213-145-843&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
Alberto Abad obtained the Telecommunication Engineering degree on 2002 from [http://www.upc.edu Technical University of Catalonia] (UPC). Between 2002 and 2006 he has been a PhD student at the [http://www.talp.upc.edu TALP Research Centre] of the [http://www.tsc.upc.edu/ Department of Signal Theory and Communications] of the Technical University of Catalonia, where he was supported by a Catalan Government grant. His Thesis was related with multi-microphone approaches to speech processing in smart-room environments.&lt;br /&gt;
&lt;br /&gt;
At UPC he participated in several European and National Spanish projects. He was also lecturer of a laboratory course about signal processing applications in the Telecommunication Engineering School of Barcelona during the academic years 2004-2005 and 2005-2006, and collaborating lecturer in the doctorate course of the Department of Signal Theory and Communications &amp;quot;Speech and Language Technology and Applications&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
He is currently a researcher in the Spoken Language Systems Lab (L2F) of INESC-ID. He has participated in the TECNOVOZ project developing robust automatic speech recognition for telephone based applications and he is currently involved in the on-going Vidivideo and PostPor projects. His current research interests include robust speech recognition, speaker  and language identification, computational acoustic scene analysis and multimodal audio-visual applications.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;inesc-id what='person' id='905'&amp;gt;&amp;lt;/inesc-id&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Research Interests ==&lt;br /&gt;
&lt;br /&gt;
* Multi-microphone processing&lt;br /&gt;
* Automatic Speech Recognition&lt;br /&gt;
&lt;br /&gt;
== Resume ==&lt;br /&gt;
PDF Version [http://www.l2f.inesc-id.pt/~alberto/cv07.pdf]&lt;br /&gt;
&lt;br /&gt;
[[category:People]]&lt;br /&gt;
[[category:Researchers]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7164</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7164"/>
		<updated>2013-10-17T00:18:58Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from [http://www.l2f.inesc-id.pt/~alberto/STDfusion/STDfusion.v1.tgz here].&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluation scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in the previous step.&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
:* '''Warning''' The [https://sites.google.com/site/bosaristoolkit/ Bosaris toolkit] needs to be installed. Change the variable $BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters $P_TARGET, $C_MISS, $C_FA &lt;br /&gt;
::*  $TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
: This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the individual systems and for the fusion system obtained following this example are:&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 ||  0.2355  / 0.2320&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7163</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7163"/>
		<updated>2013-10-17T00:04:34Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from [[Media:STDfusion.tgz | here]].&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluation scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in the previous step.&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
:* '''Warning''' The [https://sites.google.com/site/bosaristoolkit/ Bosaris toolkit] needs to be installed. Change the variable $BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters $P_TARGET, $C_MISS, $C_FA &lt;br /&gt;
::*  $TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
: This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the individual systems and for the fusion system obtained following this example are:&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 ||  0.2355  / 0.2320&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7162</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7162"/>
		<updated>2013-10-17T00:03:30Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from [[Media:STDfusion.tgz | here]].&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluation scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in the previous step.&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
:* '''Warning''' The [https://sites.google.com/site/bosaristoolkit/ Bosaris toolkit] needs to be installed. Change the variable $BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters $P_TARGET, $C_MISS, $C_FA &lt;br /&gt;
::*  $TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
:: This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the individual systems and for the fusion system obtained following this example are:&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7161</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7161"/>
		<updated>2013-10-17T00:00:14Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from [[Media:STDfusion.tgz | here]].&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluation scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in the previous step.&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The [https://sites.google.com/site/bosaristoolkit/ Bosaris toolkit] needs to be installed. Change the variable $BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters $P_TARGET, $C_MISS, $C_FA &lt;br /&gt;
&lt;br /&gt;
::*  $TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the individual systems and for the fusion system obtained following this example are:&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7160</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7160"/>
		<updated>2013-10-16T23:59:15Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from [[Media:STDfusion.tgz | here]].&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluation scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in the previous step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The [https://sites.google.com/site/bosaristoolkit/ Bosaris toolkit] needs to be installed. Change the variable $BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters $P_TARGET, $C_MISS, $C_FA &lt;br /&gt;
&lt;br /&gt;
::*  $TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the individual systems and for the fusion system obtained following this example are:&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7159</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7159"/>
		<updated>2013-10-16T23:53:57Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from [[Media:STDfusion.tgz | here]].&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The [https://sites.google.com/site/bosaristoolkit/ Bosaris toolkit] needs to be installed. Change the variable $BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters $P_TARGET, $C_MISS, $C_FA &lt;br /&gt;
&lt;br /&gt;
::*  $TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the individual systems and for the fusion system obtained following this example are:&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7158</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7158"/>
		<updated>2013-10-16T23:52:25Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from [[Media:STDfusion.tgz | here]].&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The [https://sites.google.com/site/bosaristoolkit/ Bosaris toolkit] needs to be installed. Change the variable $BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters $P_TARGET, $C_MISS, $C_FA &lt;br /&gt;
&lt;br /&gt;
::*  $TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7157</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7157"/>
		<updated>2013-10-16T23:51:22Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from [[Media:STDfusion.tgz | here]].&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the $MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable $BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters $P_TARGET, $C_MISS, $C_FA &lt;br /&gt;
&lt;br /&gt;
::*  $TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7156</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7156"/>
		<updated>2013-10-16T23:49:01Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the [https://www.l2f.inesc-id.pt L2F] and the [http://gtts.ehu.es/gtts/ GTTS] groups in the context of their research activities in the topic of query-by-example STD. The method has been succesfully tested in the SWS2012 and SWS2013 tasks of  [http://www.multimediaeval.org/mediaeval2013/sws2013/ Mediaeval Evaluation]. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters P_TARGET, C_MISS, C_FA &lt;br /&gt;
&lt;br /&gt;
::*  TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7155</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7155"/>
		<updated>2013-10-16T23:24:47Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o dev_scores_4fusion.txt \&lt;br /&gt;
          -t num_queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_scores_4fusion.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''num_queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o eval_scores_4fusion.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh dev_scores_4fusion.txt eval_scores_4fusion.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters P_TARGET, C_MISS, C_FA &lt;br /&gt;
&lt;br /&gt;
::*  TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the dev_scores_4fusion.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7154</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7154"/>
		<updated>2013-10-16T23:16:24Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' DEV_SCORES_4FUSION.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o EVAL_SCORES_4FUSION.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters P_TARGET, C_MISS, C_FA &lt;br /&gt;
&lt;br /&gt;
::*  TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the DEV_SCORES_4FUSION.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
{| border=1 cellspacing=0 align=center cellpadding=5px width=50%&lt;br /&gt;
  |+ '''Reference results for Mediaeval SWS 2013''' &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7153</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7153"/>
		<updated>2013-10-16T23:13:46Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' DEV_SCORES_4FUSION.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
      &lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o EVAL_SCORES_4FUSION.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters P_TARGET, C_MISS, C_FA &lt;br /&gt;
&lt;br /&gt;
::*  TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the DEV_SCORES_4FUSION.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=1&lt;br /&gt;
  |+ Reference results for Mediaeval SWS 2013 &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7152</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7152"/>
		<updated>2013-10-16T23:12:12Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' DEV_SCORES_4FUSION.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
      &lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o EVAL_SCORES_4FUSION.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters P_TARGET, C_MISS, C_FA &lt;br /&gt;
&lt;br /&gt;
::*  TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the DEV_SCORES_4FUSION.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
{| border=1&lt;br /&gt;
  |+ Reference sample results at Mediaeval 2013 SWS &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7151</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7151"/>
		<updated>2013-10-16T23:11:37Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' DEV_SCORES_4FUSION.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
      &lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o EVAL_SCORES_4FUSION.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters P_TARGET, C_MISS, C_FA &lt;br /&gt;
&lt;br /&gt;
::*  TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the DEV_SCORES_4FUSION.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
{| border=1&lt;br /&gt;
  |+ TWV &lt;br /&gt;
|-&lt;br /&gt;
  ! !! dev (mtwv/atwv) !! eval  (mtwv/atwv) &lt;br /&gt;
|-&lt;br /&gt;
  |   akws-br   ||    0.1571 / 0.1408 || 0.1441  / 0.1271 &lt;br /&gt;
|-&lt;br /&gt;
  |   dtw-br   ||   0.2066  / 0.2012 ||  0.1654  /  0.1581 &lt;br /&gt;
|-&lt;br /&gt;
  | fusion || 0.2731  / 0.2713 || &lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7150</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7150"/>
		<updated>2013-10-16T22:56:22Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' DEV_SCORES_4FUSION.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
      &lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o EVAL_SCORES_4FUSION.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters P_TARGET, C_MISS, C_FA &lt;br /&gt;
&lt;br /&gt;
::*  TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the DEV_SCORES_4FUSION.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	   akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7149</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7149"/>
		<updated>2013-10-16T22:55:46Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP1'''&amp;lt;/big&amp;gt; Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' DEV_SCORES_4FUSION.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
      &lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP2'''&amp;lt;/big&amp;gt; Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o EVAL_SCORES_4FUSION.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
     &lt;br /&gt;
&amp;lt;big&amp;gt;'''STEP3'''&amp;lt;/big&amp;gt; Train fusion parameters using thedevelopment scores and apply it to the evaluation scores &lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The Bosaris toolkit needs to be installed. Change the variable BOSARIS to the path where this the Bosaris toolkit is installed.&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The script contains 4 internal variables that are specific to the Mediaeval SWS2013 evaluation: &lt;br /&gt;
::* The cost parameters P_TARGET, C_MISS, C_FA &lt;br /&gt;
&lt;br /&gt;
::*  TSPEECH - The total duration (in seconds) of the data collection &lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' The call to the fusion script can take several minutes. &lt;br /&gt;
&lt;br /&gt;
:* This call permits learning from the DEV_SCORES_4FUSION.txt (with groundtruh information) the calibration and fusion parameters that are applied both to the development and to the evaluation scores. It produces 3 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' dev_fusion.scores - file containing the development well-calibrated fusion scores&lt;br /&gt;
::'''2.''' eval_fusion.scores - file containing the evaluation well-calibrated fusion scores&lt;br /&gt;
::'''3.''' fuse_params.txt - file containig thefusion parameters &lt;br /&gt;
&lt;br /&gt;
::The format of the score output files is as follows (one row per candidate detection):&lt;br /&gt;
&lt;br /&gt;
              &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
::The &amp;lt;decision&amp;gt; field is 0 or 1 depending if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
* To convert the score files to the SWS2013 stdslist.xml format type the following commands:&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh dev_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
          ./bin/raw2stdslist.sh eval_fusion.scores \&lt;br /&gt;
          ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
:* This script can take a third parameter, a threshold value  to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;big&amp;gt;'''Reference results'''&amp;lt;big&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* In the folder ./sample_results/ you can find each one of the intermediate files procuced in this example&lt;br /&gt;
&lt;br /&gt;
* The TWV results obtained with the sample score files provided for the Mediaeval SWS2013 task are for each one of the indovidual systems and for the fusion system obtained following this example are :&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	   akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7148</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7148"/>
		<updated>2013-10-16T22:31:56Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
'''STEP1''' Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' DEV_SCORES_4FUSION.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
            &lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
       Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
        &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
        -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
        -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
        -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
        -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
        -g add ground-truth information to the outputfile&lt;br /&gt;
        -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
        -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
        -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
        -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
        -h help &lt;br /&gt;
&lt;br /&gt;
        NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
'''STEP2'''  Prepare the evaluations scores for fusion&lt;br /&gt;
&lt;br /&gt;
* Type  the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm \&lt;br /&gt;
          -o EVAL_SCORES_4FUSION.txt \&lt;br /&gt;
          -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/eval/akws_br-evalterms.stdlist.xml  \&lt;br /&gt;
          ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:*'''Warning''' It is fundamental to provide the stdlist.xml input  files in the same order used in theprevious step.&lt;br /&gt;
&lt;br /&gt;
:*Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7147</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7147"/>
		<updated>2013-10-16T22:26:23Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the&lt;br /&gt;
           groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
'''STEP1''' Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
:* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
:* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
::'''1.''' DEV_SCORES_4FUSION.txt - It contains the scores ready to be used for the following fusion stage in the following general format (one row per detection candidate):&lt;br /&gt;
&lt;br /&gt;
                  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;sc1&amp;gt; &amp;lt;sc2&amp;gt; ... &amp;lt;scN&amp;gt; &amp;lt;label&amp;gt;&lt;br /&gt;
            &lt;br /&gt;
::*Notice that now all systems produce a score for all candidate detections (the score matrix is full). &lt;br /&gt;
&lt;br /&gt;
::*  If the -g option is selected (like in this case) the &amp;lt;label&amp;gt; column contains 0s and 1s for the false and true trials respectively (derived from de rttm file). If the -g option is not selected, the last column simply contains a column with 1s.&lt;br /&gt;
&lt;br /&gt;
::'''2.'''queries_in_data.ref - This second (optional) output file contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
*You can have a look to the general usage of this script calling it without arguments:&lt;br /&gt;
  &lt;br /&gt;
    Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
     &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
     -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
     -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
     -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
     -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
     -g add ground-truth information to the outputfile&lt;br /&gt;
     -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
     -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
     -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
     -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
     -h help &lt;br /&gt;
&lt;br /&gt;
     NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
  STEP2 - Prepare the eval scores for fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
    ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm -o EVAL_SCORES_4FUSION.txt -z qnorm -m 1 -n 1 ./scores/eval/akws_br-evalterms.stdlist.xml  ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
    Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
    It is also FUNDAMENTAL to provide the stdlist.xml score files in the same order used in the preparation of the development scores.&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7146</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7146"/>
		<updated>2013-10-16T22:18:32Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the&lt;br /&gt;
           groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
'''STEP1''' Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
* DEV_SCORES_4FUSION.txt: It contains the scores ready to be used for the following fusion stage in the following formar:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  Notice that all systems have produced a score for all candidate detections. &lt;br /&gt;
&lt;br /&gt;
  If the -g option is selected (like in this case) the last column will contain the 0s and 1s for the false and true trials respectively derived from de rttm. If the -g option is not selected, the last column simply contains a column of 1s.&lt;br /&gt;
&lt;br /&gt;
  The second (optional) output file is the queries_in_data.ref. This file simply contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
  You can have a look to the general usage of this script typing without arguments:&lt;br /&gt;
  &lt;br /&gt;
    Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
     &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
     -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
     -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
     -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
     -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
     -g add ground-truth information to the outputfile&lt;br /&gt;
     -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
     -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
     -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
     -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
     -h help &lt;br /&gt;
&lt;br /&gt;
     NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
  STEP2 - Prepare the eval scores for fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
    ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm -o EVAL_SCORES_4FUSION.txt -z qnorm -m 1 -n 1 ./scores/eval/akws_br-evalterms.stdlist.xml  ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
    Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
    It is also FUNDAMENTAL to provide the stdlist.xml score files in the same order used in the preparation of the development scores.&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7145</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7145"/>
		<updated>2013-10-16T22:14:49Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the&lt;br /&gt;
           groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
'''STEP1''' Prepare the development scores for learning the calibration/fusion&lt;br /&gt;
&lt;br /&gt;
* Type the following command:&lt;br /&gt;
&lt;br /&gt;
          ./bin/PrepareForFusion.sh \&lt;br /&gt;
          -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
          -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
          -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
          -t queries_in_data.ref \&lt;br /&gt;
          -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
          ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
          ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
* '''Warning''' Before running the script, you will have to change the MATLAB_BIN variable to be the path where your Matlab binary is actually installed.&lt;br /&gt;
&lt;br /&gt;
* This process  will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
1. DEV_SCORES_4FUSION.txt contains the scores ready to be used for the following fusion stage in the following formar:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  Notice that all systems have produced a score for all candidate detections. &lt;br /&gt;
&lt;br /&gt;
  If the -g option is selected (like in this case) the last column will contain the 0s and 1s for the false and true trials respectively derived from de rttm. If the -g option is not selected, the last column simply contains a column of 1s.&lt;br /&gt;
&lt;br /&gt;
  The second (optional) output file is the queries_in_data.ref. This file simply contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
  You can have a look to the general usage of this script typing without arguments:&lt;br /&gt;
  &lt;br /&gt;
    Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
     &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
     -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
     -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
     -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
     -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
     -g add ground-truth information to the outputfile&lt;br /&gt;
     -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
     -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
     -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
     -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
     -h help &lt;br /&gt;
&lt;br /&gt;
     NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
  STEP2 - Prepare the eval scores for fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
    ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm -o EVAL_SCORES_4FUSION.txt -z qnorm -m 1 -n 1 ./scores/eval/akws_br-evalterms.stdlist.xml  ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
    Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
    It is also FUNDAMENTAL to provide the stdlist.xml score files in the same order used in the preparation of the development scores.&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7144</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7144"/>
		<updated>2013-10-16T22:09:36Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - Contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - Contains the different scripts necessary for calibration and fusion&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the&lt;br /&gt;
           groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - Used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - Used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation scores&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the SWS evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - Contains the Mediaeval 2013 atwv scoring package&lt;br /&gt;
         ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample (intermediate) results that are obtained if you run the  instructions bellow&lt;br /&gt;
&lt;br /&gt;
== Example of use ==&lt;br /&gt;
&lt;br /&gt;
'''STEP1''' Prepare the development scores for training the fusion parameters&lt;br /&gt;
&lt;br /&gt;
* Typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/PrepareForFusion.sh \&lt;br /&gt;
  -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml \&lt;br /&gt;
  -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm \&lt;br /&gt;
  -o DEV_SCORES_4FUSION.txt \&lt;br /&gt;
  -t queries_in_data.ref \&lt;br /&gt;
  -g -z qnorm -m 1 -n 1 \&lt;br /&gt;
  ./scores/dev/akws_br-devterms.stdlist.xml \&lt;br /&gt;
  ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
  Before running the script, change the MATLAB_BIN variable to be the path to your Matlab binary.&lt;br /&gt;
&lt;br /&gt;
  This script call will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
  The first one is the DEV_SCORES_4FUSION.txt that contains the scores ready to be used for the following stage. The format is something like:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  Notice that all systems have produced a score for all candidate detections. &lt;br /&gt;
&lt;br /&gt;
  If the -g option is selected (like in this case) the last column will contain the 0s and 1s for the false and true trials respectively derived from de rttm. If the -g option is not selected, the last column simply contains a column of 1s.&lt;br /&gt;
&lt;br /&gt;
  The second (optional) output file is the queries_in_data.ref. This file simply contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
  You can have a look to the general usage of this script typing without arguments:&lt;br /&gt;
  &lt;br /&gt;
    Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
     &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
     -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
     -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
     -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
     -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
     -g add ground-truth information to the outputfile&lt;br /&gt;
     -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
     -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
     -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
     -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
     -h help &lt;br /&gt;
&lt;br /&gt;
     NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
  STEP2 - Prepare the eval scores for fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
    ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm -o EVAL_SCORES_4FUSION.txt -z qnorm -m 1 -n 1 ./scores/eval/akws_br-evalterms.stdlist.xml  ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
    Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
    It is also FUNDAMENTAL to provide the stdlist.xml score files in the same order used in the preparation of the development scores.&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7143</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7143"/>
		<updated>2013-10-16T22:01:38Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems as described in:&lt;br /&gt;
&lt;br /&gt;
           A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
           On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
           In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here.&lt;br /&gt;
&lt;br /&gt;
== Package contents ==&lt;br /&gt;
&lt;br /&gt;
* The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
         README.txt - It contains pretty much the same information of this wiki site&lt;br /&gt;
         ./bin/ - This directory contains the different scripts necessary for calibration and fusion:&lt;br /&gt;
         ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the&lt;br /&gt;
                                                    groundtruh for the input systems&lt;br /&gt;
         ./bin/align_score_files.pl - This script is used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/create_groundtruth_centerdistance.pl - This script is used by ./bin/PrepareForFusion.sh&lt;br /&gt;
         ./bin/heuristicScoring.pl - This script is used by ./bin/PrepareForFusion.sh &lt;br /&gt;
         ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation set&lt;br /&gt;
         ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the evaluation&lt;br /&gt;
         ./scoring_atwv_sws2013/ - This directory contains the Mediaeval 2013 official scoring package (only for atwv)&lt;br /&gt;
        ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
         ./sample_results/ - Contains the sample results that should be obtained if you run the scripts following the instructions bellow&lt;br /&gt;
&lt;br /&gt;
  STEP BY STEP SAMPLE INSTRUCTIONS&lt;br /&gt;
&lt;br /&gt;
  STEP1 - Prepare the dev scores for training fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm -o DEV_SCORES_4FUSION.txt -t queries_in_data.ref -g -z qnorm -m 1 -n 1 ./scores/dev/akws_br-devterms.stdlist.xml ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
  Before running the script, change the MATLAB_BIN variable to be the path to your Matlab binary.&lt;br /&gt;
&lt;br /&gt;
  This script call will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
  The first one is the DEV_SCORES_4FUSION.txt that contains the scores ready to be used for the following stage. The format is something like:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  Notice that all systems have produced a score for all candidate detections. &lt;br /&gt;
&lt;br /&gt;
  If the -g option is selected (like in this case) the last column will contain the 0s and 1s for the false and true trials respectively derived from de rttm. If the -g option is not selected, the last column simply contains a column of 1s.&lt;br /&gt;
&lt;br /&gt;
  The second (optional) output file is the queries_in_data.ref. This file simply contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
  You can have a look to the general usage of this script typing without arguments:&lt;br /&gt;
  &lt;br /&gt;
    Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
     &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
     -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
     -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
     -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
     -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
     -g add ground-truth information to the outputfile&lt;br /&gt;
     -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
     -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
     -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
     -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
     -h help &lt;br /&gt;
&lt;br /&gt;
     NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
  STEP2 - Prepare the eval scores for fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
    ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm -o EVAL_SCORES_4FUSION.txt -z qnorm -m 1 -n 1 ./scores/eval/akws_br-evalterms.stdlist.xml  ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
    Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
    It is also FUNDAMENTAL to provide the stdlist.xml score files in the same order used in the preparation of the development scores.&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7142</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7142"/>
		<updated>2013-10-16T21:53:57Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
* In this wiki we make available the necessary code and the basic instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection (STD) systems. &lt;br /&gt;
&lt;br /&gt;
* The proposed method is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in the Mediaeval Spoken Web Search (SWS) task. The method has been succesfully tested in the SWS2012 and SWS2013 tasks. &lt;br /&gt;
&lt;br /&gt;
* Unfortunately, we did not have time to consolidate (and clean up) the code and what we are making available is very much the same set of scripts that we developed during the first experiments. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some external dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
* The package can be downloaded from here. Please cite the following work if you find it useful:&lt;br /&gt;
&lt;br /&gt;
     A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
     On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
     In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
  README.txt - This file&lt;br /&gt;
  ./bin/ - This directory contains the different scripts necessary for calibration and fusion:&lt;br /&gt;
  ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for a vaiable number of input devices&lt;br /&gt;
  ./bin/align_score_files.pl - This script is used by ./bin/PrepareForFusion.sh&lt;br /&gt;
  ./bin/create_groundtruth_centerdistance.pl - This script is used by ./bin/PrepareForFusion.sh&lt;br /&gt;
  ./bin/heuristicScoring.pl - This script is used by ./bin/PrepareForFusion.sh &lt;br /&gt;
  ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation set&lt;br /&gt;
  ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the evaluation&lt;br /&gt;
  ./scoring_atwv_sws2013/ - This directory contains the Mediaeval 2013 official scoring package (only for atwv)&lt;br /&gt;
  ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
  ./sample_results/ - Contains the sample results that should be obtained if you run the scripts following the instructions bellow&lt;br /&gt;
&lt;br /&gt;
  STEP BY STEP SAMPLE INSTRUCTIONS&lt;br /&gt;
&lt;br /&gt;
  STEP1 - Prepare the dev scores for training fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm -o DEV_SCORES_4FUSION.txt -t queries_in_data.ref -g -z qnorm -m 1 -n 1 ./scores/dev/akws_br-devterms.stdlist.xml ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
  Before running the script, change the MATLAB_BIN variable to be the path to your Matlab binary.&lt;br /&gt;
&lt;br /&gt;
  This script call will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
  The first one is the DEV_SCORES_4FUSION.txt that contains the scores ready to be used for the following stage. The format is something like:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  Notice that all systems have produced a score for all candidate detections. &lt;br /&gt;
&lt;br /&gt;
  If the -g option is selected (like in this case) the last column will contain the 0s and 1s for the false and true trials respectively derived from de rttm. If the -g option is not selected, the last column simply contains a column of 1s.&lt;br /&gt;
&lt;br /&gt;
  The second (optional) output file is the queries_in_data.ref. This file simply contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
  You can have a look to the general usage of this script typing without arguments:&lt;br /&gt;
  &lt;br /&gt;
    Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
     &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
     -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
     -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
     -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
     -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
     -g add ground-truth information to the outputfile&lt;br /&gt;
     -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
     -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
     -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
     -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
     -h help &lt;br /&gt;
&lt;br /&gt;
     NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
  STEP2 - Prepare the eval scores for fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
    ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm -o EVAL_SCORES_4FUSION.txt -z qnorm -m 1 -n 1 ./scores/eval/akws_br-evalterms.stdlist.xml  ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
    Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
    It is also FUNDAMENTAL to provide the stdlist.xml score files in the same order used in the preparation of the development scores.&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7141</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7141"/>
		<updated>2013-10-16T21:42:43Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction&lt;br /&gt;
&lt;br /&gt;
In this wiki we make available the necessary code and the instructions to carry out discriminative calibration and fusion of heterogeneous spoken term detection systems. This approach is the result of the collaboration between the L2F and the GTTS groups in the context of their activities in  and described in:&lt;br /&gt;
&lt;br /&gt;
     A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. &lt;br /&gt;
     On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. &lt;br /&gt;
     In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The proposed approach has been succesfully tested for the query-by-example spoken search task of Mediaeval 2012 and Mediaeval 2013. Sample Mediaeval 2013 system results are provided to test the package.&lt;br /&gt;
&lt;br /&gt;
Unfortunately, we did not have time to consolidate (and clean) the code and what we are making available is very much the same set of scripts that we developed during the preparation of the experiments reported in the Interspeech paper. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
  The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
  README.txt - This file&lt;br /&gt;
  ./bin/ - This directory contains the different scripts necessary for calibration and fusion:&lt;br /&gt;
  ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for a vaiable number of input devices&lt;br /&gt;
  ./bin/align_score_files.pl - This script is used by ./bin/PrepareForFusion.sh&lt;br /&gt;
  ./bin/create_groundtruth_centerdistance.pl - This script is used by ./bin/PrepareForFusion.sh&lt;br /&gt;
  ./bin/heuristicScoring.pl - This script is used by ./bin/PrepareForFusion.sh &lt;br /&gt;
  ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation set&lt;br /&gt;
  ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the evaluation&lt;br /&gt;
  ./scoring_atwv_sws2013/ - This directory contains the Mediaeval 2013 official scoring package (only for atwv)&lt;br /&gt;
  ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
  ./sample_results/ - Contains the sample results that should be obtained if you run the scripts following the instructions bellow&lt;br /&gt;
&lt;br /&gt;
  STEP BY STEP SAMPLE INSTRUCTIONS&lt;br /&gt;
&lt;br /&gt;
  STEP1 - Prepare the dev scores for training fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm -o DEV_SCORES_4FUSION.txt -t queries_in_data.ref -g -z qnorm -m 1 -n 1 ./scores/dev/akws_br-devterms.stdlist.xml ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
  Before running the script, change the MATLAB_BIN variable to be the path to your Matlab binary.&lt;br /&gt;
&lt;br /&gt;
  This script call will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
  The first one is the DEV_SCORES_4FUSION.txt that contains the scores ready to be used for the following stage. The format is something like:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  Notice that all systems have produced a score for all candidate detections. &lt;br /&gt;
&lt;br /&gt;
  If the -g option is selected (like in this case) the last column will contain the 0s and 1s for the false and true trials respectively derived from de rttm. If the -g option is not selected, the last column simply contains a column of 1s.&lt;br /&gt;
&lt;br /&gt;
  The second (optional) output file is the queries_in_data.ref. This file simply contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
  You can have a look to the general usage of this script typing without arguments:&lt;br /&gt;
  &lt;br /&gt;
    Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
     &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
     -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
     -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
     -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
     -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
     -g add ground-truth information to the outputfile&lt;br /&gt;
     -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
     -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
     -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
     -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
     -h help &lt;br /&gt;
&lt;br /&gt;
     NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
  STEP2 - Prepare the eval scores for fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
    ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm -o EVAL_SCORES_4FUSION.txt -z qnorm -m 1 -n 1 ./scores/eval/akws_br-evalterms.stdlist.xml  ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
    Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
    It is also FUNDAMENTAL to provide the stdlist.xml score files in the same order used in the preparation of the development scores.&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7140</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7140"/>
		<updated>2013-10-16T21:37:08Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This package contains the code necessary to run descriminative calibration and fusion of spoken term detection systems resulting from the collaboration between the L2F and the GTTS group of the Unversity of the Basque Country and described in:&lt;br /&gt;
&lt;br /&gt;
 A. Abad, L. J. Rodriguez Fuentes, M. Penagarikano, A. Varona, M. Diez, and G. Bordel. On the Calibration and Fusion of Heterogeneous Spoken Term Detection Systems. In Interspeech 2013, August 25-29 2013&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The proposed approach has been succesfully tested for the query-by-example spoken search task of Mediaeval 2012 and Mediaeval 2013. Sample Mediaeval 2013 system results are provided to test the package.&lt;br /&gt;
&lt;br /&gt;
Unfortunately, we did not have time to consolidate (and clean) the code and what we are making available is very much the same set of scripts that we developed during the preparation of the experiments reported in the Interspeech paper. Thus, the code is written in different pieces in different coding languages (bash, perl, matlab, etc.) and it has some dependencies. Anyway, we expect that it can be still useful for researchers that want to try to fuse their STD systems.&lt;br /&gt;
&lt;br /&gt;
  The package contains the following files and directories:&lt;br /&gt;
&lt;br /&gt;
  README.txt - This file&lt;br /&gt;
  ./bin/ - This directory contains the different scripts necessary for calibration and fusion:&lt;br /&gt;
  ./bin/PrepareForFusion.sh  - Main script than normalizes, aligns, hypotesizes missing scores and creates the groundtruh for a vaiable number of input devices&lt;br /&gt;
  ./bin/align_score_files.pl - This script is used by ./bin/PrepareForFusion.sh&lt;br /&gt;
  ./bin/create_groundtruth_centerdistance.pl - This script is used by ./bin/PrepareForFusion.sh&lt;br /&gt;
  ./bin/heuristicScoring.pl - This script is used by ./bin/PrepareForFusion.sh &lt;br /&gt;
  ./bin/fusion.sh  - This script takes the output of ./bin/PrepareForFusion.sh to learn the fusion parameters and apply them to the evaluation set&lt;br /&gt;
  ./bin/raw2stdslist.sh - Converts scores from a raw (internal) format to the stdlist xml format of the evaluation&lt;br /&gt;
  ./scoring_atwv_sws2013/ - This directory contains the Mediaeval 2013 official scoring package (only for atwv)&lt;br /&gt;
  ./scores/ - Contains sample dev (./scores/dev/akws_br-devterms.stdlist.xml, ./scores/dev/dtw_br-devterms.stdlist.xml) and eval scores (./scores/eval/akws_br-evalterms.stdlist.xml, ./scores/eval/dtw_br-evalterms.stdlist.xml)&lt;br /&gt;
  ./sample_results/ - Contains the sample results that should be obtained if you run the scripts following the instructions bellow&lt;br /&gt;
&lt;br /&gt;
  STEP BY STEP SAMPLE INSTRUCTIONS&lt;br /&gt;
&lt;br /&gt;
  STEP1 - Prepare the dev scores for training fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml -r ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.rttm -o DEV_SCORES_4FUSION.txt -t queries_in_data.ref -g -z qnorm -m 1 -n 1 ./scores/dev/akws_br-devterms.stdlist.xml ./scores/dev/dtw_br-devterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
  Before running the script, change the MATLAB_BIN variable to be the path to your Matlab binary.&lt;br /&gt;
&lt;br /&gt;
  This script call will generate 2 output files:&lt;br /&gt;
&lt;br /&gt;
  The first one is the DEV_SCORES_4FUSION.txt that contains the scores ready to be used for the following stage. The format is something like:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;score_system1&amp;gt; &amp;lt;score_system2&amp;gt; ... &amp;lt;score_systemN&amp;gt; &amp;lt;groundtruth_label&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  Notice that all systems have produced a score for all candidate detections. &lt;br /&gt;
&lt;br /&gt;
  If the -g option is selected (like in this case) the last column will contain the 0s and 1s for the false and true trials respectively derived from de rttm. If the -g option is not selected, the last column simply contains a column of 1s.&lt;br /&gt;
&lt;br /&gt;
  The second (optional) output file is the queries_in_data.ref. This file simply contains the number of times each query appears in the collection data and it is used later in the fusion stage.&lt;br /&gt;
&lt;br /&gt;
  You can have a look to the general usage of this script typing without arguments:&lt;br /&gt;
  &lt;br /&gt;
    Usage: PrepareForFusion.sh -q &amp;lt;tlistxml&amp;gt; -r &amp;lt;rttm&amp;gt; -o &amp;lt;outputfile&amp;gt; [opts] &amp;lt;stdlistfile1&amp;gt; [stdlistfile2] [stdlistfile3]  ... [stdlistfileN] &lt;br /&gt;
&lt;br /&gt;
     &amp;lt;stdlistfile*&amp;gt; input score stdlist file in the SWS2012 format (*.stdlist.xml)        | - Required argument (at least 1)&lt;br /&gt;
     -q &amp;lt;tlistxml&amp;gt; termlist file in the SWS2012 format (*.tlist.xml)                      | - Required parameter&lt;br /&gt;
     -r &amp;lt;rttm&amp;gt; rttm file in the SWS2012 format (*.rttm)                                   | - Required parameter&lt;br /&gt;
     -o &amp;lt;outputfile&amp;gt; output file name                                                     | - Required parameter&lt;br /&gt;
     -z &amp;lt;value&amp;gt; score z-norm type (none, qnorm, fnorm, qfnorm, fqnorm)                    | - Default: none&lt;br /&gt;
     -g add ground-truth information to the outputfile&lt;br /&gt;
     -t &amp;lt;filename&amp;gt; saves the number of true terms in the reference per query (implies -g)&lt;br /&gt;
     -m &amp;lt;value&amp;gt; apply majority voting fusion with &amp;lt;value&amp;gt; minimum number of votes         | - Default: 1&lt;br /&gt;
     -n &amp;lt;value&amp;gt; method for creating default scores (0: average of the other detections (MV approach); 1; min per query, 2: global min, 3: histogram based)  | - Default: 0&lt;br /&gt;
     -d debug mode, done remove auxiliar files stored in /tmp/tmpdir.$$&lt;br /&gt;
     -h help &lt;br /&gt;
&lt;br /&gt;
     NOTE: Requires Matlab (or octave) and perl. It also depends on the perl scripts align_score_files.pl, heuristicScoring.pl and create_groundtruth_centerdistance.pl &lt;br /&gt;
           that should be located in the same folder of the main script&lt;br /&gt;
&lt;br /&gt;
  STEP2 - Prepare the eval scores for fusion typing the following command:&lt;br /&gt;
&lt;br /&gt;
    ./bin/PrepareForFusion.sh -q ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml -r ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.rttm -o EVAL_SCORES_4FUSION.txt -z qnorm -m 1 -n 1 ./scores/eval/akws_br-evalterms.stdlist.xml  ./scores/eval/dtw_br-evalterms.stdlist.xml&lt;br /&gt;
&lt;br /&gt;
    Notice that in contrast to case of the development scores, we did not use the -g option since we do not need the groudtruth in this case. &lt;br /&gt;
    &lt;br /&gt;
    It is also FUNDAMENTAL to provide the stdlist.xml score files in the same order used in the preparation of the development scores.&lt;br /&gt;
    &lt;br /&gt;
&lt;br /&gt;
  STEP3 - Train fusion on development scores and apply it to eval scores typing the following command:&lt;br /&gt;
&lt;br /&gt;
  ./bin/fusion.sh DEV_SCORES_4FUSION.txt EVAL_SCORES_4FUSION.txt  queries_in_data.ref&lt;br /&gt;
&lt;br /&gt;
  Before running this script you will need to change also the value of the MATLAB_BIN variable and download the Bosaris toolkit and change the variable BOSARIS to the path that contains this toolkit.&lt;br /&gt;
&lt;br /&gt;
  Notice also that the script contains 4 hard-coded variables that are specific to the Mediaeval SWS2013 evaluation. These are the cost parameters P_TARGET, C_MISS, C_FA and the total duration in seconds of the collection data TSPEECH&lt;br /&gt;
&lt;br /&gt;
  This fusion script uses the DEV_SCORES_4FUSION.txt (with groundtruh information) to learn the calibration and fusion parameters that are applied both to the development set and to the evaluation set. &lt;br /&gt;
&lt;br /&gt;
  As mentioned previously the queries_in_data.ref contains the statistics abbout the true number of instances of each query on the data and it is used for hypothesizing missing scores and for computing AWTV.&lt;br /&gt;
&lt;br /&gt;
  This call (that can take a while, several minutes) will produce 1 output file for the dev scores (dev_fusion.scores) and 1 output file for the eval scores (eval_fusion.scores). The format of these output files is as follows:&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
  ...&lt;br /&gt;
  ...&lt;br /&gt;
  &amp;lt;query_id&amp;gt; &amp;lt;file_id&amp;gt; &amp;lt;start_time&amp;gt; &amp;lt;duration&amp;gt; &amp;lt;fusion_score&amp;gt; &amp;lt;decision&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  The &amp;lt;decision&amp;gt; field is 0 or 1 depending respectively if the score is lower or greater than the minimum cost Bayes optimum threshold (see the Interspeech paper for details).&lt;br /&gt;
&lt;br /&gt;
  Additionally, the fusion parameters are stored in the fuse_params.txt.&lt;br /&gt;
&lt;br /&gt;
  The final step consists of converting this result files to the format used in the SWS2013 challenge:&lt;br /&gt;
&lt;br /&gt;
  ./bin/raw2stdslist.sh dev_fusion.scores ./scoring_atwv_sws2013/sws2013_dev/sws2013_dev.tlist.xml &amp;gt; fusion-devterms.stdlist.xml&lt;br /&gt;
  ./bin/raw2stdslist.sh eval_fusion.scores ./scoring_atwv_sws2013/sws2013_eval/sws2013_eval.tlist.xml &amp;gt; fusion-evalterms.stdlist.xml&lt;br /&gt;
  &lt;br /&gt;
  Notice that this script can take a third parameter. This paramater is a threshold value that you can use to apply a different decision threshold to the optimal Bayes one.&lt;br /&gt;
&lt;br /&gt;
  Reference TWV results in Mediaeval SWS2013 task:&lt;br /&gt;
&lt;br /&gt;
			      dev		   eval		&lt;br /&gt;
			mtwv	   atwv		mtwv	atwv&lt;br /&gt;
	  akws_br	0.1571     0.1408 &lt;br /&gt;
	   dtw_br&lt;br /&gt;
	   fusion&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=DTW&amp;diff=7127</id>
		<title>DTW</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=DTW&amp;diff=7127"/>
		<updated>2013-10-02T12:25:16Z</updated>

		<summary type="html">&lt;p&gt;Alberto: Created page with &amp;quot;This page will contain DTW code&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page will contain DTW code&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7126</id>
		<title>STDfusion</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=STDfusion&amp;diff=7126"/>
		<updated>2013-10-02T12:24:52Z</updated>

		<summary type="html">&lt;p&gt;Alberto: Created page with &amp;quot;This page will contain the STD fusion tools used for&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page will contain the STD fusion tools used for&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Tools&amp;diff=7125</id>
		<title>Tools</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Tools&amp;diff=7125"/>
		<updated>2013-10-02T12:22:57Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{TOCright}}&lt;br /&gt;
{| align=&amp;quot;center&amp;quot; style=&amp;quot;border-style: solid; border-width: 1px; border-color: gray80;&amp;quot;&lt;br /&gt;
|[[Image:tools1.jpg|197px]]&lt;br /&gt;
|[[Image:tools2.jpg|197px]]&lt;br /&gt;
|[[Image:tools3.jpg|197px]]&lt;br /&gt;
|}&lt;br /&gt;
Language processing uses tools such as syntactic and semantic analyzers. In order to perform their tasks, some of these  tools use linguistic information (for instance, dictionaries and grammars), making natural language processing by computers closer to the human process.&lt;br /&gt;
&lt;br /&gt;
We are using natural language processing  tools in many of our applications, namely in dialog management, automatic summarization, information retrieval, question answering, discourse analysis, term and emotion extraction.&lt;br /&gt;
Besides applying these tools to text we are applying them to automatic transcriptions of spoken documents, leading to new challenges. &lt;br /&gt;
&lt;br /&gt;
== Morphology ==&lt;br /&gt;
&lt;br /&gt;
* [[MARv]] - a morphossyntactic disambiguation tool;&lt;br /&gt;
* [[monge]] - a word form generator;&lt;br /&gt;
* [[PAsMo]] - a rule-based morphology processor, tag converter, and sentence splitter;&lt;br /&gt;
* [[RuDriCo]] - a rule-based morphology processor;&lt;br /&gt;
* [[SMorph]] - a morphological analyser;&lt;br /&gt;
* [[XA]] - a morphological analyser similar to ispell and jspell;&lt;br /&gt;
* [[YAH]] - yet another hyphenator.&lt;br /&gt;
* [[VerbForms]] - a verbal form generator;&lt;br /&gt;
&lt;br /&gt;
== Syntax ==&lt;br /&gt;
&lt;br /&gt;
* [[Algas]] -- establishes dependency relations between chunks and words&lt;br /&gt;
* [[ParVO]] -- a C++ implementation of Earley's algorithm with attribute unification (as in an attribute grammar).&lt;br /&gt;
* [[SuSAna]] -- susrface syntactic analyser and chunker&lt;br /&gt;
* [[TiraTeimas]] -- verifies if a set of chunks satisfies a set of constraints&lt;br /&gt;
&lt;br /&gt;
== Syntax/Semantics Interface ==&lt;br /&gt;
&lt;br /&gt;
* [[AsDeCopas]] -- applies contextual rules (possibly hierarchically organized) to a graph&lt;br /&gt;
* [[Ogre]] -- transforms a structure where both chunks and words are connected into a dependency structure&lt;br /&gt;
&lt;br /&gt;
== Semantics ==&lt;br /&gt;
&lt;br /&gt;
* [[ATA (system)|ATA]] - automatic term extraction (semantics-like processing)&lt;br /&gt;
&lt;br /&gt;
== Discourse Analysis ==&lt;br /&gt;
&lt;br /&gt;
* [[DID]] - a discourse indentifier.&lt;br /&gt;
&lt;br /&gt;
== Multi-purpose ==&lt;br /&gt;
&lt;br /&gt;
* [[Galinha]] - a portal for building and running applications.&lt;br /&gt;
* [[Language Resources Database|LRDB]] - a language resources database and access framework.&lt;br /&gt;
* [[fstk|FSTk]] - a finite-state transducer library.&lt;br /&gt;
* [[ShReP]] - A Framework for constructing NLP systems.&lt;br /&gt;
&lt;br /&gt;
== Speech Annotation ==&lt;br /&gt;
&lt;br /&gt;
* [[L2F_PhoneAlign]] - A DTW-based phonetic aligner.&lt;br /&gt;
* [[L2F_MuLA]] - A tool to synchronize annotation of speech at various levels of granularity.&lt;br /&gt;
&lt;br /&gt;
== Speech Recognition ==&lt;br /&gt;
&lt;br /&gt;
* [[AUDIMUS]] - Automatic Large Vocabulary Continuous Speech Recognition for the European Portuguese language&lt;br /&gt;
&lt;br /&gt;
== Speech Search ==&lt;br /&gt;
&lt;br /&gt;
* [[DTW]] - DTW system developed for the Mediaeval 2013 Spoken Web Search challenge&lt;br /&gt;
* [[STDfusion]] - Tools for discriminative calibration/fusion of heterogeneous spoken term detection systems&lt;br /&gt;
&lt;br /&gt;
[[category:Tools]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Master_Theses_Themes_at_L2F&amp;diff=7027</id>
		<title>Master Theses Themes at L2F</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Master_Theses_Themes_at_L2F&amp;diff=7027"/>
		<updated>2013-05-10T13:26:47Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* Aferição de técnicas de apresentação&lt;br /&gt;
&lt;br /&gt;
* My-ROBOT: a robotic dialogue based learning companion for a maths game&lt;br /&gt;
&lt;br /&gt;
* Simplificação/tradução automática de textos&lt;br /&gt;
&lt;br /&gt;
* Ferramenta de pesquisa em áudio para advogados&lt;br /&gt;
&lt;br /&gt;
* Microphone network selection and calibration approaches in multi-room environments (Aluno: Miguel Matos, Orientador: Alberto Abad)&lt;br /&gt;
&lt;br /&gt;
* Mobile alert: combining human mobile motion detection and voice analysis&lt;br /&gt;
&lt;br /&gt;
* Os arquivos de áudio de ontem e de hoje&lt;br /&gt;
&lt;br /&gt;
* Query-by-example speaker search in large data speech collections (Aluno: ???, Orientador: Alberto Abad)&lt;br /&gt;
&lt;br /&gt;
* Spoken term detection in speech collections using spoken queries (Aluno: ???, Orientador: Alberto Abad)&lt;br /&gt;
&lt;br /&gt;
* Towards a universal language recognition system (Aluno: ???, Orientador: Alberto Abad)&lt;br /&gt;
&lt;br /&gt;
* Using Shazam-like audio fingerprinting to block advertisements in audio podcasts&lt;br /&gt;
&lt;br /&gt;
Advertisement-blocking applications in internet browsers (e.g. AdBlock in Firefox or Chrome) block advertisements appearing on websites so that they do not display on screen. Audio fingerprinting applications like Shazam recognize a song from small segments by using audio fingerprinting. The objective of this thesis is to design a program which is able to block advertisements found on audio available for download on the web (podcast) on semi-supervised form. The method should exploit fingerprinting techniques like those of Shazam and the fact that Ads do not change from podcast to podcast, unlike the conventional content. The work will include&lt;br /&gt;
&lt;br /&gt;
- Gathering a small set of audio examples from the web containing advertisements (i.e. from BBC podcasts) which is representative of the problem to be solved.&lt;br /&gt;
&lt;br /&gt;
- Using fingerprinting algorithms to identify similar segments in the audio podcasts, store fingerprints and filter out advertisements.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Ensinando um chatbot a responder de modo credível com base em diferentes tipos de feedback&lt;br /&gt;
&lt;br /&gt;
*Just.Chat – respondendo como um humano com base em legendas de filmes&lt;br /&gt;
&lt;br /&gt;
* SERTO – Sobrevivendo a Erros de Reconhecimento, de Teclado e Outros&lt;br /&gt;
&lt;br /&gt;
* POE: PlatafOrma de Escrita&lt;br /&gt;
&lt;br /&gt;
* Dizes-me onde vive a Scarlett Johansson? – desenvolvimento de um sistema de pergunta/resposta para português&lt;br /&gt;
&lt;br /&gt;
* Medicine.Ask: An intelligent search facility for medicine information&lt;br /&gt;
&lt;br /&gt;
* Teach me! Interface verbal para ensino de agentes artificiais&lt;br /&gt;
&lt;br /&gt;
* Aplicação de métodos de aprendizagem ativa em Língua Natural (Aluno: Nuno Aniceto, Orientador: Nuno Mamede)&lt;br /&gt;
&lt;br /&gt;
* Classificador de textos para o ensino de português como segunda língua (Aluno: Pedro Curto, Orientador: Nuno Mamede)&lt;br /&gt;
&lt;br /&gt;
* Corretor ortográfico e identificação de sufixos com Transdutores (Aluno: Marco Ferreira, Orientador: Nuno Mamede)&lt;br /&gt;
&lt;br /&gt;
* Desambiguação semântica de nomes (Aluno: Rita Policarpo, Orientador: Nuno Mamede)&lt;br /&gt;
&lt;br /&gt;
* Desambiguação sintatica/semântica de verbos (Aluno: Goncalo Suissas, Orientador: Nuno Mamede)&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=Master_Theses_Themes_at_L2F&amp;diff=7026</id>
		<title>Master Theses Themes at L2F</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=Master_Theses_Themes_at_L2F&amp;diff=7026"/>
		<updated>2013-05-10T13:26:14Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* Aferição de técnicas de apresentação&lt;br /&gt;
&lt;br /&gt;
* My-ROBOT: a robotic dialogue based learning companion for a maths game&lt;br /&gt;
&lt;br /&gt;
* Simplificação/tradução automática de textos&lt;br /&gt;
&lt;br /&gt;
* Ferramenta de pesquisa em áudio para advogados&lt;br /&gt;
&lt;br /&gt;
* Microphone network selection and calibration approaches in multi-room environments (Aluno: Miguel Matos Orientador: Alberto Abad)&lt;br /&gt;
&lt;br /&gt;
* Mobile alert: combining human mobile motion detection and voice analysis&lt;br /&gt;
&lt;br /&gt;
* Os arquivos de áudio de ontem e de hoje&lt;br /&gt;
&lt;br /&gt;
* Query-by-example speaker search in large data speech collections (Aluno: ??? Orientador: Alberto Abad)&lt;br /&gt;
&lt;br /&gt;
* Spoken term detection in speech collections using spoken queries (Aluno: ??? Orientador: Alberto Abad)&lt;br /&gt;
&lt;br /&gt;
* Towards a universal language recognition system (Aluno: ??? Orientador: Alberto Abad)&lt;br /&gt;
&lt;br /&gt;
* Using Shazam-like audio fingerprinting to block advertisements in audio podcasts&lt;br /&gt;
&lt;br /&gt;
Advertisement-blocking applications in internet browsers (e.g. AdBlock in Firefox or Chrome) block advertisements appearing on websites so that they do not display on screen. Audio fingerprinting applications like Shazam recognize a song from small segments by using audio fingerprinting. The objective of this thesis is to design a program which is able to block advertisements found on audio available for download on the web (podcast) on semi-supervised form. The method should exploit fingerprinting techniques like those of Shazam and the fact that Ads do not change from podcast to podcast, unlike the conventional content. The work will include&lt;br /&gt;
&lt;br /&gt;
- Gathering a small set of audio examples from the web containing advertisements (i.e. from BBC podcasts) which is representative of the problem to be solved.&lt;br /&gt;
&lt;br /&gt;
- Using fingerprinting algorithms to identify similar segments in the audio podcasts, store fingerprints and filter out advertisements.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Ensinando um chatbot a responder de modo credível com base em diferentes tipos de feedback&lt;br /&gt;
&lt;br /&gt;
*Just.Chat – respondendo como um humano com base em legendas de filmes&lt;br /&gt;
&lt;br /&gt;
* SERTO – Sobrevivendo a Erros de Reconhecimento, de Teclado e Outros&lt;br /&gt;
&lt;br /&gt;
* POE: PlatafOrma de Escrita&lt;br /&gt;
&lt;br /&gt;
* Dizes-me onde vive a Scarlett Johansson? – desenvolvimento de um sistema de pergunta/resposta para português&lt;br /&gt;
&lt;br /&gt;
* Medicine.Ask: An intelligent search facility for medicine information&lt;br /&gt;
&lt;br /&gt;
* Teach me! Interface verbal para ensino de agentes artificiais&lt;br /&gt;
&lt;br /&gt;
* Aplicação de métodos de aprendizagem ativa em Língua Natural (Aluno: Nuno Aniceto, Orientador: Nuno Mamede)&lt;br /&gt;
&lt;br /&gt;
* Classificador de textos para o ensino de português como segunda língua (Aluno: Pedro Curto, Orientador: Nuno Mamede)&lt;br /&gt;
&lt;br /&gt;
* Corretor ortográfico e identificação de sufixos com Transdutores (Aluno: Marco Ferreira, Orientador: Nuno Mamede)&lt;br /&gt;
&lt;br /&gt;
* Desambiguação semântica de nomes (Aluno: Rita Policarpo, Orientador: Nuno Mamede)&lt;br /&gt;
&lt;br /&gt;
* Desambiguação sintatica/semântica de verbos (Aluno: Goncalo Suissas, Orientador: Nuno Mamede)&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=People&amp;diff=6945</id>
		<title>People</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=People&amp;diff=6945"/>
		<updated>2013-02-22T18:44:09Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__ &amp;lt;!--[[Image:People1.jpg|thumb|right|350px|Click for larger picture.]]--&amp;gt; &lt;br /&gt;
&amp;lt;!-- == Researchers ==&lt;br /&gt;
&lt;br /&gt;
== Invited Researched ==&lt;br /&gt;
&lt;br /&gt;
== Associated Researchers ==&lt;br /&gt;
&lt;br /&gt;
== Post-Doc Researchers == &amp;lt; ! - - junior researcher + phd - - &amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Junior Researchers == &amp;lt; ! - - phd students - - &amp;gt;&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
== Researchers  ==&lt;br /&gt;
&lt;br /&gt;
{| width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! style=&amp;quot;font-weight: normal; text-align: left; width: 25%;&amp;quot; | [[Ramon Fernandez Astudillo]]&lt;br /&gt;
! style=&amp;quot;font-weight: normal; text-align: left; width: 25%;&amp;quot; | [[Gracinda Carvalho]] &lt;br /&gt;
| style=&amp;quot;font-weight: normal; text-align: left; width: 25%;&amp;quot; | [[Nuno Mamede]]&lt;br /&gt;
| style=&amp;quot;font-weight: normal; text-align: left; width: 25%;&amp;quot; | [[Thomas Pellegrini]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Anabela Barreiro]]&lt;br /&gt;
| [[Joao Paulo Carvalho]] &lt;br /&gt;
| [[David Martins de Matos]]&lt;br /&gt;
| [[Ricardo Daniel Ribeiro|Ricardo Ribeiro]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Fernando Batista]]&lt;br /&gt;
| [[Luísa Coheur]]&lt;br /&gt;
| [[Hugo Meinedo]]&lt;br /&gt;
| [[António Serralheiro|António J. Serralheiro]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Luís Caldas de Oliveira|Luís C. Oliveira]]&lt;br /&gt;
| [[Alberto Abad Gareta]]&lt;br /&gt;
| [[João Paulo Neto]]&lt;br /&gt;
| [[Isabel Trancoso|Isabel M. Trancoso]] (Coordinator)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== PhD Students  ==&lt;br /&gt;
&lt;br /&gt;
{| width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! style=&amp;quot;font-weight: normal; text-align: left; width: 25%;&amp;quot; | [[Gopala Krishna Anumanchipalli]]  &lt;br /&gt;
! style=&amp;quot;font-weight: normal; text-align: left; width: 25%;&amp;quot; | [[José David Lopes]]&lt;br /&gt;
! style=&amp;quot;font-weight: normal; text-align: left; width: 25%;&amp;quot; | [[Helena Moniz]]&lt;br /&gt;
! style=&amp;quot;font-weight: normal; text-align: left; width: 25%;&amp;quot; | [[Hugo Rodrigues]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Rui Correia]]&lt;br /&gt;
| [[Isabel Mascarenhas]]&lt;br /&gt;
| [[Pedro Mota]]&lt;br /&gt;
| [[Paula Cristina Vaz|Paula Cristina Vaz]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Ângela Costa]]&lt;br /&gt;
| [[Luís Marujo]]&lt;br /&gt;
| [[Joana Paulo Pardal]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Luís Garcia]] (external)&lt;br /&gt;
| [[Ana Cristina Mendes]]&lt;br /&gt;
| [[José Portêlo]] &lt;br /&gt;
|-&lt;br /&gt;
| [[Wang Ling]] &lt;br /&gt;
| [[João Miranda]]&lt;br /&gt;
| [[Eugénio Ribeiro]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Research Associates  ==&lt;br /&gt;
&lt;br /&gt;
{| width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Vera Cabarrão]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Cláudio Diniz]] &lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Pedro Fialho]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | &lt;br /&gt;
|-&lt;br /&gt;
| [[Sérgio Curto]]&lt;br /&gt;
| [[Jaime Ferreira]]&lt;br /&gt;
| [[Justyna Kosmala]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Associated Researchers  ==&lt;br /&gt;
&lt;br /&gt;
{| width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Jorge Baptista]] &lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Diamantino Caseiro]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Ciro Martins]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | &lt;br /&gt;
|-&lt;br /&gt;
| [[Miguel Bugalho]] &lt;br /&gt;
| [[João Graça]]&lt;br /&gt;
| [[Maria do Céu Viana]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Masters Students  ==&lt;br /&gt;
&lt;br /&gt;
{| width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Cristiana Amorim]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Pedro Figueirinha]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Sérgio Morais]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Tiago Travanca]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Viviana Cabrita]]&lt;br /&gt;
| [[Alfredo Gomes]]&lt;br /&gt;
| [[Miguel Neto]]&lt;br /&gt;
| [[Pedro Valério]]&lt;br /&gt;
|-&lt;br /&gt;
| [[João Camejo]]&lt;br /&gt;
| [[Tiago Gonçalves]]&lt;br /&gt;
| [[Anna Pompili]] *&lt;br /&gt;
| [[Alexandre Vicente]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Filipe Carapinha]]&lt;br /&gt;
| [[Vahid Keshavarz Hedayati]] *&lt;br /&gt;
| [[Daniel Rosa]]&lt;br /&gt;
| [[Zoran Vitez]]&lt;br /&gt;
|-&lt;br /&gt;
| [[Tagore Dinis]]&lt;br /&gt;
| [[Jorge Jorge]] *&lt;br /&gt;
| [[Ricardo Silva]] *&lt;br /&gt;
|-&lt;br /&gt;
| [[Vanessa Feliciano]]&lt;br /&gt;
| [[Miguel Matos]] *&lt;br /&gt;
| [[João Silvestre Marques]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
[*] Scholarship&lt;br /&gt;
&lt;br /&gt;
== Trainees ==&lt;br /&gt;
&lt;br /&gt;
{| width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Yegor Afanasyev]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Guilherme Ferreira]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | [[Filipe Pires da Silva]]&lt;br /&gt;
! width=&amp;quot;25%&amp;quot; style=&amp;quot;text-align: left; font-weight: normal;&amp;quot; | &lt;br /&gt;
|-&lt;br /&gt;
| &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Administrative Support  ==&lt;br /&gt;
&lt;br /&gt;
{| width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! style=&amp;quot;font-weight: normal; text-align: left; width: 33%;&amp;quot; | [[Teresa Mimoso]] &lt;br /&gt;
! style=&amp;quot;font-weight: normal; text-align: left; width: 33%;&amp;quot; | &lt;br /&gt;
| &lt;br /&gt;
|-&lt;br /&gt;
| &lt;br /&gt;
| &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Former L²F Members  ==&lt;br /&gt;
&lt;br /&gt;
*List of [[Former L²F Members]]&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Logo-dirha.jpg&amp;diff=6711</id>
		<title>File:Logo-dirha.jpg</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Logo-dirha.jpg&amp;diff=6711"/>
		<updated>2012-06-19T11:58:39Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Logo-dirha.png&amp;diff=6710</id>
		<title>File:Logo-dirha.png</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Logo-dirha.png&amp;diff=6710"/>
		<updated>2012-06-19T11:55:46Z</updated>

		<summary type="html">&lt;p&gt;Alberto: uploaded a new version of &amp;amp;quot;File:Logo-dirha.png&amp;amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Logo-dirha.png&amp;diff=6709</id>
		<title>File:Logo-dirha.png</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Logo-dirha.png&amp;diff=6709"/>
		<updated>2012-06-19T11:53:45Z</updated>

		<summary type="html">&lt;p&gt;Alberto: uploaded a new version of &amp;amp;quot;File:Logo-dirha.png&amp;amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
	<entry>
		<id>https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Logo-dirha.png&amp;diff=6708</id>
		<title>File:Logo-dirha.png</title>
		<link rel="alternate" type="text/html" href="https://www.hlt.inesc-id.pt/wiki/index.php?title=File:Logo-dirha.png&amp;diff=6708"/>
		<updated>2012-06-19T11:51:13Z</updated>

		<summary type="html">&lt;p&gt;Alberto: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Alberto</name></author>
	</entry>
</feed>