<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
		>
<channel>
	<title>Comments on: Haghighi and Klein (2009): Simple Coreference Resolution with Rich Syntactic and Semantic Features</title>
	<atom:link href="http://brenocon.com/blog/2009/08/haghighi-and-klein-2009-simple-coreference-resolution-with-rich-syntactic-and-semantic-features/feed/" rel="self" type="application/rss+xml" />
	<link>http://brenocon.com/blog/2009/08/haghighi-and-klein-2009-simple-coreference-resolution-with-rich-syntactic-and-semantic-features/</link>
	<description>cognition, language, social systems; statistics, visualization, computation</description>
	<lastBuildDate>Tue, 25 Nov 2025 13:11:20 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	
	<item>
		<title>By: Breck Baldwin</title>
		<link>http://brenocon.com/blog/2009/08/haghighi-and-klein-2009-simple-coreference-resolution-with-rich-syntactic-and-semantic-features/#comment-10307</link>
		<dc:creator>Breck Baldwin</dc:creator>
		<pubDate>Wed, 26 Aug 2009 22:06:03 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=618#comment-10307</guid>
		<description><![CDATA[Yikes! That system architecture sure looks familiar with the exception of the semantic filtering bootstrapped from Wikipedia--only it was 1995 and I was calling Michael Collins &quot;boy&quot;. It was the UPenn MUC-6 entry that &quot;unofficially&quot; won coref done with Jeff Reynar, Michael Collins, Jason Eisner, Adwait Ratnaparkhi, Joseph Rosenzweig, Anoop Sarkar and Srinivas &lt;a href=&quot;http://www.aclweb.org/anthology-new/M/M95/M95-1015.pdf&quot; rel=&quot;nofollow&quot;&gt;UNIVERSITY OF PENNSYLVANIA :DESCRIPTION OF THE UNIVERSITY OF PENNSYLVANIA SYSTEM USED FOR MUC- 6&lt;/a&gt;
 

The performance improvement is stunning, 87.2% precision, 77.3% recall over our performance of 72% precision, 63% recall. But we had to find our own mentions lemme tell ya, it was up hill both ways to the LINC lab and we thought a megabyte of memory was a pretty huge thing. 

But seriously, when did coreference papers get to report performance without finding their mentions? I know ACE provided that option but it somehow just seems wrong to have gold standard mentions that you only have to do coref over since there is no natural data out there that contains perfect noun phrases that are known to corefer to something. 

Amit Bagga and I did experiments with the MUC-6 data like score the case where all gold standard mentions were made coreferent--you get 100% recall with around 35% precsion as I recall. Those results let us to consider other scoring metrics which then led to B-cubed which additionally involved Alan Bierman. 

A request: Run the same system with a mention detection pass. Those numbers should be exciting.]]></description>
		<content:encoded><![CDATA[<p>Yikes! That system architecture sure looks familiar with the exception of the semantic filtering bootstrapped from Wikipedia&#8211;only it was 1995 and I was calling Michael Collins &#8220;boy&#8221;. It was the UPenn MUC-6 entry that &#8220;unofficially&#8221; won coref done with Jeff Reynar, Michael Collins, Jason Eisner, Adwait Ratnaparkhi, Joseph Rosenzweig, Anoop Sarkar and Srinivas <a href="http://www.aclweb.org/anthology-new/M/M95/M95-1015.pdf" rel="nofollow">UNIVERSITY OF PENNSYLVANIA :DESCRIPTION OF THE UNIVERSITY OF PENNSYLVANIA SYSTEM USED FOR MUC- 6</a></p>
<p>The performance improvement is stunning, 87.2% precision, 77.3% recall over our performance of 72% precision, 63% recall. But we had to find our own mentions lemme tell ya, it was up hill both ways to the LINC lab and we thought a megabyte of memory was a pretty huge thing. </p>
<p>But seriously, when did coreference papers get to report performance without finding their mentions? I know ACE provided that option but it somehow just seems wrong to have gold standard mentions that you only have to do coref over since there is no natural data out there that contains perfect noun phrases that are known to corefer to something. </p>
<p>Amit Bagga and I did experiments with the MUC-6 data like score the case where all gold standard mentions were made coreferent&#8211;you get 100% recall with around 35% precsion as I recall. Those results let us to consider other scoring metrics which then led to B-cubed which additionally involved Alan Bierman. </p>
<p>A request: Run the same system with a mention detection pass. Those numbers should be exciting.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Bob Carpenter</title>
		<link>http://brenocon.com/blog/2009/08/haghighi-and-klein-2009-simple-coreference-resolution-with-rich-syntactic-and-semantic-features/#comment-9323</link>
		<dc:creator>Bob Carpenter</dc:creator>
		<pubDate>Sun, 09 Aug 2009 23:23:35 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=618#comment-9323</guid>
		<description><![CDATA[I also have the peeve that I can&#039;t reproduce people&#039;s papers, because they go into great detail about CRFs (which is standard fare), but then gloss over the details of their features.  I really liked Jenny Finkel&#039;s bakeoff paper on CRFs for Biocreative that described their features and the recent Ratinov and Roth paper on named entity for the same reason.  But even those fall short of a reproducible system description, which I still hold out as one of the hallmarks of science.

Papers that squeeze half a percent improvement on no-longer-held out data seem useless (one more paper evaluating on section 23 of the Treebank or the ModApte split of Reuters and I&#039;ll scream).  

I also like papers that introduce statistical techniques or applications, and those papers are important, too.  For instance, I loved Haghighi and Klein&#039;s previous approach to coref, based on the Dirichlet process, because it seems such a natural fit (other than computationally).

Of course there are non hyponym/hypernym coref cases!  It&#039;d be a stretch to call &quot;Ronald Reagan, the U.S. president, ...&quot; a hypernym case, because &quot;president&quot; isn&#039;t in any sense more generic than &quot;Ronald Reagan&quot; (and it&#039;s temporally dependent; the relation isn&#039;t permanent).  But even eliminating names, you get things like, &quot;the president, a former actor&quot;, where there&#039;s clearly no hypernym/hyponym relation.  

The earliest work I know of for bootstrapping hypernyms/hyponyms was Ann Copestake&#039;s work on LKB from 1990/1991.

We&#039;ve probably both been thinking about natural tagging examples (your footnote [1]), because those are things you can get done with Mechanical Turk!  A deeper problem is that annotation schemes like the Penn Treebank (even just the POS part) are built on theories of underlying linguistic structure which may be nothing like &quot;the truth&quot;, and may not even be useful in real world tasks.  For instance, distributional clustering can be much more effective than POS tags as features in CRFs or parsers, and require no labeled data.  (Though you can also use both in a discriminative model like CRFs.)  But even named entities and coref have a large number of corner cases that need to be documented or just punted (e.g. &quot;erstwhile Phineas Phoggs&quot; from MUC-6 or &quot;the various Bob Dylans&quot; from recent film reviews).]]></description>
		<content:encoded><![CDATA[<p>I also have the peeve that I can&#8217;t reproduce people&#8217;s papers, because they go into great detail about CRFs (which is standard fare), but then gloss over the details of their features.  I really liked Jenny Finkel&#8217;s bakeoff paper on CRFs for Biocreative that described their features and the recent Ratinov and Roth paper on named entity for the same reason.  But even those fall short of a reproducible system description, which I still hold out as one of the hallmarks of science.</p>
<p>Papers that squeeze half a percent improvement on no-longer-held out data seem useless (one more paper evaluating on section 23 of the Treebank or the ModApte split of Reuters and I&#8217;ll scream).  </p>
<p>I also like papers that introduce statistical techniques or applications, and those papers are important, too.  For instance, I loved Haghighi and Klein&#8217;s previous approach to coref, based on the Dirichlet process, because it seems such a natural fit (other than computationally).</p>
<p>Of course there are non hyponym/hypernym coref cases!  It&#8217;d be a stretch to call &#8220;Ronald Reagan, the U.S. president, &#8230;&#8221; a hypernym case, because &#8220;president&#8221; isn&#8217;t in any sense more generic than &#8220;Ronald Reagan&#8221; (and it&#8217;s temporally dependent; the relation isn&#8217;t permanent).  But even eliminating names, you get things like, &#8220;the president, a former actor&#8221;, where there&#8217;s clearly no hypernym/hyponym relation.  </p>
<p>The earliest work I know of for bootstrapping hypernyms/hyponyms was Ann Copestake&#8217;s work on LKB from 1990/1991.</p>
<p>We&#8217;ve probably both been thinking about natural tagging examples (your footnote [1]), because those are things you can get done with Mechanical Turk!  A deeper problem is that annotation schemes like the Penn Treebank (even just the POS part) are built on theories of underlying linguistic structure which may be nothing like &#8220;the truth&#8221;, and may not even be useful in real world tasks.  For instance, distributional clustering can be much more effective than POS tags as features in CRFs or parsers, and require no labeled data.  (Though you can also use both in a discriminative model like CRFs.)  But even named entities and coref have a large number of corner cases that need to be documented or just punted (e.g. &#8220;erstwhile Phineas Phoggs&#8221; from MUC-6 or &#8220;the various Bob Dylans&#8221; from recent film reviews).</p>
]]></content:encoded>
	</item>
</channel>
</rss>
