<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
		>
<channel>
	<title>Comments on: SF conference for data mining mercenaries</title>
	<atom:link href="https://brenocon.com/blog/2009/01/sf-conference-for-data-mining-mercenaries/feed/" rel="self" type="application/rss+xml" />
	<link>https://brenocon.com/blog/2009/01/sf-conference-for-data-mining-mercenaries/</link>
	<description>cognition, language, social systems; statistics, visualization, computation</description>
	<lastBuildDate>Tue, 25 Nov 2025 13:11:20 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	
	<item>
		<title>By: Bob Carpenter</title>
		<link>https://brenocon.com/blog/2009/01/sf-conference-for-data-mining-mercenaries/#comment-2089</link>
		<dc:creator>Bob Carpenter</dc:creator>
		<pubDate>Fri, 23 Jan 2009 21:47:48 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=415#comment-2089</guid>
		<description><![CDATA[Amen to the data munging being most of the work.  We&#039;re currently working on a customer project and I&#039;ve already written two different parsers for their CSV data.  The customer&#039;s programmers&#039;  efforts to piece together the data from their web logs has been Herculean.  And still we&#039;re not 100% sure of what we have.  There are all kinds of fields with numeric codes that are just plain hard to figure out, even with the business people and some of the coders present -- much of it&#039;s legacy codes that have to be extracted from comments in their .h header files.  

We were getting good results with a feature that turned out to be cheating because while it made sense to use it, the value in the logs didn&#039;t reflect the value in the incoming request, but rather the value in the outgoing response, which indirectly coded the category for classification.  

The kicker is that this data&#039;s only an approximation of the real problem.  But it&#039;s the best we have, and while more data&#039;s better than more learning, some data&#039;s better than nothing.

Other customers have wanted us to find things for them (e.g. forward earnings statements in 10Q footnotes, opinions of cars in blogs, recording artists in news), but there was no existing data, so we had to (help them) create it.  That&#039;s when they run into the problems like whether &quot;Bob Dylans&quot; is a person mention in &lt;a href=&quot;http://www.rollingstone.com/rockdaily/index.php/2007/08/20/the-six-bob-dylans-more-photos-from-todd-haynes-im-not-there-movie/&quot; rel=&quot;nofollow&quot;&gt;The Six Bob Dylans: More Photos From Todd Haynes’ “I’m Not There” Movie&lt;/a&gt;.  It turns out the customers are not semantic grad students or ontology boffins, so they usually just don&#039;t care.

But the real problem with all of this tuning to within a percent of a system&#039;s life is that it&#039;s usually just overfitting when you go out into the wild.  For instance, the customer mentioned above plans to change the overall organization and the instruction text on their site, so that none of our training data will exactly replicate the runtime environment.]]></description>
		<content:encoded><![CDATA[<p>Amen to the data munging being most of the work.  We&#8217;re currently working on a customer project and I&#8217;ve already written two different parsers for their CSV data.  The customer&#8217;s programmers&#8217;  efforts to piece together the data from their web logs has been Herculean.  And still we&#8217;re not 100% sure of what we have.  There are all kinds of fields with numeric codes that are just plain hard to figure out, even with the business people and some of the coders present &#8212; much of it&#8217;s legacy codes that have to be extracted from comments in their .h header files.  </p>
<p>We were getting good results with a feature that turned out to be cheating because while it made sense to use it, the value in the logs didn&#8217;t reflect the value in the incoming request, but rather the value in the outgoing response, which indirectly coded the category for classification.  </p>
<p>The kicker is that this data&#8217;s only an approximation of the real problem.  But it&#8217;s the best we have, and while more data&#8217;s better than more learning, some data&#8217;s better than nothing.</p>
<p>Other customers have wanted us to find things for them (e.g. forward earnings statements in 10Q footnotes, opinions of cars in blogs, recording artists in news), but there was no existing data, so we had to (help them) create it.  That&#8217;s when they run into the problems like whether &#8220;Bob Dylans&#8221; is a person mention in <a href="http://www.rollingstone.com/rockdaily/index.php/2007/08/20/the-six-bob-dylans-more-photos-from-todd-haynes-im-not-there-movie/" rel="nofollow">The Six Bob Dylans: More Photos From Todd Haynes’ “I’m Not There” Movie</a>.  It turns out the customers are not semantic grad students or ontology boffins, so they usually just don&#8217;t care.</p>
<p>But the real problem with all of this tuning to within a percent of a system&#8217;s life is that it&#8217;s usually just overfitting when you go out into the wild.  For instance, the customer mentioned above plans to change the overall organization and the instruction text on their site, so that none of our training data will exactly replicate the runtime environment.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Michael E. Driscoll</title>
		<link>https://brenocon.com/blog/2009/01/sf-conference-for-data-mining-mercenaries/#comment-2088</link>
		<dc:creator>Michael E. Driscoll</dc:creator>
		<pubDate>Fri, 23 Jan 2009 21:26:32 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=415#comment-2088</guid>
		<description><![CDATA[Brendan - 

Nice catch on this conference - Eric Siegel, the organizer, appears to be heavily focused on real-world case studies -- which should be interesting.

As you know, the Bay Area R UseRs group is doing a free, co-located event on Wed evening of the conference -- so if you&#039;re interested in mingling with some PAW folks as well as some R users -- you can sign up at:  http://ia.meetup.com/67/calendar/9573566/ 

Mike]]></description>
		<content:encoded><![CDATA[<p>Brendan &#8211; </p>
<p>Nice catch on this conference &#8211; Eric Siegel, the organizer, appears to be heavily focused on real-world case studies &#8212; which should be interesting.</p>
<p>As you know, the Bay Area R UseRs group is doing a free, co-located event on Wed evening of the conference &#8212; so if you&#8217;re interested in mingling with some PAW folks as well as some R users &#8212; you can sign up at:  <a href="http://ia.meetup.com/67/calendar/9573566/" rel="nofollow">http://ia.meetup.com/67/calendar/9573566/</a> </p>
<p>Mike</p>
]]></content:encoded>
	</item>
</channel>
</rss>

<!-- Dynamic page generated in 0.012 seconds. -->
<!-- Cached page generated by WP-Super-Cache on 2026-04-06 04:50:25 -->
