<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
		>
<channel>
	<title>Comments on: I don&#8217;t get this web parsing shared task</title>
	<atom:link href="http://brenocon.com/blog/2012/03/i-dont-get-this-web-parsing-shared-task/feed/" rel="self" type="application/rss+xml" />
	<link>http://brenocon.com/blog/2012/03/i-dont-get-this-web-parsing-shared-task/</link>
	<description>cognition, language, social systems; statistics, visualization, computation</description>
	<lastBuildDate>Tue, 25 Nov 2025 13:11:20 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	
	<item>
		<title>By: Graphs for SANCL-2012 web parsing results &#124; AI and Social Science &#8211; Brendan O&#039;Connor</title>
		<link>http://brenocon.com/blog/2012/03/i-dont-get-this-web-parsing-shared-task/#comment-222160</link>
		<dc:creator>Graphs for SANCL-2012 web parsing results &#124; AI and Social Science &#8211; Brendan O&#039;Connor</dc:creator>
		<pubDate>Sat, 24 Nov 2012 01:24:30 +0000</pubDate>
		<guid isPermaLink="false">http://brenocon.com/blog/?p=1195#comment-222160</guid>
		<description><![CDATA[[...] of Wall Street Journal annotated data and very little in-domain training data. (Previous discussion here; see Ryan McDonald&#8217;s detailed comment.) Here are some graphs of the results (last page in the [...]]]></description>
		<content:encoded><![CDATA[<p>[...] of Wall Street Journal annotated data and very little in-domain training data. (Previous discussion here; see Ryan McDonald&#8217;s detailed comment.) Here are some graphs of the results (last page in the [...]</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Brendan O'Connor</title>
		<link>http://brenocon.com/blog/2012/03/i-dont-get-this-web-parsing-shared-task/#comment-127941</link>
		<dc:creator>Brendan O'Connor</dc:creator>
		<pubDate>Fri, 09 Mar 2012 04:10:32 +0000</pubDate>
		<guid isPermaLink="false">http://brenocon.com/blog/?p=1195#comment-127941</guid>
		<description><![CDATA[re: getting new annotations the standard way of doing business -- yes.  I remember when Powerset finally hired an annotation expert.  Was way more useful than quite a bit of the other work we were doing.

Thank you for the explanation of your goals, it is helpful.]]></description>
		<content:encoded><![CDATA[<p>re: getting new annotations the standard way of doing business &#8212; yes.  I remember when Powerset finally hired an annotation expert.  Was way more useful than quite a bit of the other work we were doing.</p>
<p>Thank you for the explanation of your goals, it is helpful.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Brendan O'Connor</title>
		<link>http://brenocon.com/blog/2012/03/i-dont-get-this-web-parsing-shared-task/#comment-127938</link>
		<dc:creator>Brendan O'Connor</dc:creator>
		<pubDate>Fri, 09 Mar 2012 04:04:52 +0000</pubDate>
		<guid isPermaLink="false">http://brenocon.com/blog/?p=1195#comment-127938</guid>
		<description><![CDATA[re: semi-supervised learning for twitter POS tagging -- I don&#039;t know the details, but there was one experiment that turned out negative.  It is easy to believe more feature engineering could be a better use of researcher time, though hard to prove such a negative of course.]]></description>
		<content:encoded><![CDATA[<p>re: semi-supervised learning for twitter POS tagging &#8212; I don&#8217;t know the details, but there was one experiment that turned out negative.  It is easy to believe more feature engineering could be a better use of researcher time, though hard to prove such a negative of course.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Ryan McDonald</title>
		<link>http://brenocon.com/blog/2012/03/i-dont-get-this-web-parsing-shared-task/#comment-127933</link>
		<dc:creator>Ryan McDonald</dc:creator>
		<pubDate>Fri, 09 Mar 2012 03:37:35 +0000</pubDate>
		<guid isPermaLink="false">http://brenocon.com/blog/?p=1195#comment-127933</guid>
		<description><![CDATA[A couple points. First, as the website points out:

&quot;It is permissible to use previously constructed lexicons, word clusters or other resources provided that they are made available for other participants.&quot;

So you can use clusters, but in the spirit of open competition, we ask that these resources be made available.

Second, I agree that taking a domain, running system X on the domain, doing an error analysis and then adding features, changing the model or annotating some more data is a very good way to adapt systems. I don&#039;t think anyone is &#039;scared&#039; of this approach. In fact, outside academia, this is the standard way of doing business, not the exception. However, this is not as easy as it sounds. First, you need the resources (human resources that is) to do this for every domain on the web or domain you might be interested in. Second, the annotations you wish to collect must be easily created by you or via a system like mechanical turk. It is one thing to annotate some short twitter posts with 12-15 part of speech tags and a whole other thing to annotate consumer reviews with syntactic structure. I have tried both. They are not comparable. Even the former cannot be done reliably by turkers, which means you will need grad students, staff research scientists or costly third party vendors to do this every time you want to study a new domain.

So the spirit of the competition was to see, from a modeling/algorithm perspective, what are the best methods for training robust syntactic analyzers on the data currently available. By limiting the resources we are trying to make this as much an apples-to-apples comparison as we can. Even this is impossible. Some parsers require lexicons that might have been tuned for specific domains, etc.

Understanding this is still valuable in the &quot;analyze, annotate and iterate&quot; approach. Don&#039;t you want to start off with the best baseline to reduce the amount of human labor required?]]></description>
		<content:encoded><![CDATA[<p>A couple points. First, as the website points out:</p>
<p>&#8220;It is permissible to use previously constructed lexicons, word clusters or other resources provided that they are made available for other participants.&#8221;</p>
<p>So you can use clusters, but in the spirit of open competition, we ask that these resources be made available.</p>
<p>Second, I agree that taking a domain, running system X on the domain, doing an error analysis and then adding features, changing the model or annotating some more data is a very good way to adapt systems. I don&#8217;t think anyone is &#8216;scared&#8217; of this approach. In fact, outside academia, this is the standard way of doing business, not the exception. However, this is not as easy as it sounds. First, you need the resources (human resources that is) to do this for every domain on the web or domain you might be interested in. Second, the annotations you wish to collect must be easily created by you or via a system like mechanical turk. It is one thing to annotate some short twitter posts with 12-15 part of speech tags and a whole other thing to annotate consumer reviews with syntactic structure. I have tried both. They are not comparable. Even the former cannot be done reliably by turkers, which means you will need grad students, staff research scientists or costly third party vendors to do this every time you want to study a new domain.</p>
<p>So the spirit of the competition was to see, from a modeling/algorithm perspective, what are the best methods for training robust syntactic analyzers on the data currently available. By limiting the resources we are trying to make this as much an apples-to-apples comparison as we can. Even this is impossible. Some parsers require lexicons that might have been tuned for specific domains, etc.</p>
<p>Understanding this is still valuable in the &#8220;analyze, annotate and iterate&#8221; approach. Don&#8217;t you want to start off with the best baseline to reduce the amount of human labor required?</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Fernando Pereira</title>
		<link>http://brenocon.com/blog/2012/03/i-dont-get-this-web-parsing-shared-task/#comment-127929</link>
		<dc:creator>Fernando Pereira</dc:creator>
		<pubDate>Fri, 09 Mar 2012 03:21:26 +0000</pubDate>
		<guid isPermaLink="false">http://brenocon.com/blog/?p=1195#comment-127929</guid>
		<description><![CDATA[The purpose of a shared task is to compare models and algorithms, not to compare human annotation and error analysis skill, which depends mainly on a team&#039;s supply of human labor. As for &quot;Our lack of semi-supervised learning was not a weakness,&quot; how do you know? Have you proven that such a better learning algorithm is impossible?]]></description>
		<content:encoded><![CDATA[<p>The purpose of a shared task is to compare models and algorithms, not to compare human annotation and error analysis skill, which depends mainly on a team&#8217;s supply of human labor. As for &#8220;Our lack of semi-supervised learning was not a weakness,&#8221; how do you know? Have you proven that such a better learning algorithm is impossible?</p>
]]></content:encoded>
	</item>
</channel>
</rss>
