<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>AI and Social Science - Brendan O&#039;Connor &#187; Best Posts</title>
	<atom:link href="http://brenocon.com/blog/category/best-of/feed/" rel="self" type="application/rss+xml" />
	<link>http://brenocon.com/blog</link>
	<description>cognition, language, social systems; statistics, visualization, computation</description>
	<lastBuildDate>Mon, 08 May 2017 15:29:14 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	
		<item>
		<title>Beautiful Data book chapter</title>
		<link>http://brenocon.com/blog/2009/08/beautiful-data-book-chapter/</link>
		<comments>http://brenocon.com/blog/2009/08/beautiful-data-book-chapter/#comments</comments>
		<pubDate>Wed, 12 Aug 2009 22:14:47 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://anyall.org/blog/?p=634</guid>
		<description><![CDATA[Today I received my copy of Beautiful Data, a just-released anthology of articles about, well, working with data.  Lukas and I contributed a chapter on analyzing social perceptions in web data.  See it here. After a long process of drafting, &#8230; <a href="http://brenocon.com/blog/2009/08/beautiful-data-book-chapter/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
				<content:encoded><![CDATA[<p><img alt="" src="http://assets.doloreslabs.com/blog/beautiful-data.gif" class="alignright" width="85" height="112" /> Today I received my copy of <a href="http://oreilly.com/catalog/9780596157111/">Beautiful Data</a>, a just-released anthology of articles about, well, working with data.  <a href="http://blog.doloreslabs.com/2009/08/beautiful-data/">Lukas</a> and I contributed a chapter on analyzing social perceptions in web data.  <a href="http://anyall.org/bd">See it here.</a> After a long process of drafting, proofreading, re-drafting, and bothering the publishers under rather sudden deadlines, I&#8217;ve resolved to never use graphics again in anything I write :)</p>
<p>Here&#8217;s our final figure, a <a href="http://en.wikipedia.org/wiki/K-means_clustering">k-means</a> clustering of face photos via perceived social attributes (social <a href="http://en.wikipedia.org/wiki/Concept_learning">concepts/types</a>? with <a href="http://en.wikipedia.org/wiki/Prototype_theory">exemplars</a>?):<br />
<a href="http://anyall.org/cluster_table.png"><img src="http://anyall.org/cluster_table.png" alt="" title="cluster_table" width="500" height="593" class="aligncenter size-full wp-image-637" /></a></p>
<p>I just started reading the rest of the book and it&#8217;s very fun.  <a href="http://norvig.com/">Peter Norvig</a>&#8216;s chapter on language models is gripping.  (It does word segmentation, ciphers, and more, in that lovely python-centric tutorial style extending his previous <a href="http://norvig.com/spell-correct.html">spell correction article</a>.)  There are also chapters by many other great researchers and practitioners (some of whom you may have seen around this blog or its neighborhood) like <a href="http://www.stat.columbia.edu/~gelman/">Andrew Gelman</a>, <a href="http://had.co.nz/">Hadley Wickham</a>, <a href="http://mike.teczno.com/">Michal Migurski</a>, <a href="http://jheer.org/">Jeffrey Heer</a>, and still more&#8230;  I&#8217;m impressed just by the talent-gathering-and-organizing operation.  Big kudos to editors <a href="http://kiwitobes.com/">Toby Segaran</a> and <a href="http://www.linkedin.com/in/jhammerb">Jeff Hammerbacher</a>, and O&#8217;Reilly&#8217;s <a href="http://twitter.com/jsteeleeditor">Julie Steele</a>.</p>
<p>I also have an apparently secret code that gets you a discount, so email me if you want it.  I wonder if I&#8217;m not supposed to give out many of them.  Hm.</p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2009/08/beautiful-data-book-chapter/feed/</wfw:commentRss>
		<slash:comments>4</slash:comments>
		</item>
		<item>
		<title>Announcing TweetMotif for summarizing twitter topics</title>
		<link>http://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/</link>
		<comments>http://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/#comments</comments>
		<pubDate>Mon, 18 May 2009 17:40:03 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://anyall.org/blog/?p=515</guid>
		<description><![CDATA[Update (3/14/2010): There is now a TweetMotif paper. Last week, I, with my awesome friends David Ahn and Mike Krieger, finished hacking together an experimental prototype, TweetMotif, for exploratory search on Twitter. If you want to know what people are &#8230; <a href="http://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
				<content:encoded><![CDATA[<p><b>Update</b> (3/14/2010): There is now a <a href="http://anyall.org/oconnor_krieger_ahn.icwsm2010.tweetmotif.pdf">TweetMotif paper</a>.<br />
<hr />
Last week, I, with my awesome friends <a href="http://tweetmotif.com/about">David Ahn and Mike Krieger</a>, finished hacking together an experimental prototype, <a href="http://tweetmotif.com/">TweetMotif</a>, for exploratory search on Twitter.  If you want to know what people are thinking about something, the normal search interface <a href="http://search.twitter.com/">search.twitter.com</a> gives really cool information, but it&#8217;s hard to wade through hundreds or thousands of results.  We take tweets matching a query and group together similar messages, showing significant terms and phrases  that co-occur with the user query.  Try it out at <a href="http://tweetmotif.com">tweetmotif.com</a>.  Here&#8217;s an example for a current hot topic, <a href="http://tweetmotif.com/#%23wolframalpha">#WolframAlpha</a>:</p>
<p><a href="http://anyall.org/blog/wp-content/uploads/2009/05/wa_top.png"><img class="aligncenter size-full wp-image-517" title="wa_top" src="http://anyall.org/blog/wp-content/uploads/2009/05/wa_top.png" alt="" width="500" height="85" /></a><br />
<a href="http://anyall.org/blog/wp-content/uploads/2009/05/wa_bot.png"><img class="aligncenter size-full wp-image-518" title="wa_bot" src="http://anyall.org/blog/wp-content/uploads/2009/05/wa_bot.png" alt="" width="500" height="240" /></a></p>
<p>It&#8217;s currently showing tweets that match both <a href="http://tweetmotif.com/#%23wolframalpha">#WolframAlpha</a> as well as two interesting bigrams: &#8220;queries failed&#8221; and &#8220;google killer&#8221;.  TweetMotif doesn&#8217;t attempt to derive the meaning or sentiment toward the phrases &#8212; NLP is hard, and doing this much is hard enough! &#8212; but it&#8217;s easy for you to look at the tweets themselves and figure out what&#8217;s going on.</p>
<p>Here&#8217;s another fun example right now, a query for <a href="http://tweetmotif.com/#Dollhouse">Dollhouse</a>:</p>
<p><a href="http://anyall.org/blog/wp-content/uploads/2009/05/dollhouse_top.png"><img class="aligncenter size-full wp-image-519" title="dollhouse_top" src="http://anyall.org/blog/wp-content/uploads/2009/05/dollhouse_top.png" alt="" width="500" height="26" /></a><br />
<a href="http://anyall.org/blog/wp-content/uploads/2009/05/dollhouse_bot.png"><img class="aligncenter size-full wp-image-520" title="dollhouse_bot" src="http://anyall.org/blog/wp-content/uploads/2009/05/dollhouse_bot.png" alt="" width="500" height="253" /></a></p>
<p>I love that the #wolframalpha topic has &#8220;infected&#8221; the dollhouse space.  Someone pointed out a connection between them, but really they&#8217;re connected through bot spam.  TweetMotif&#8217;s duplicate detection algorithm found 22 messages here where each is basically a list of all the trending topics.  This seems to be a popular form of twitter spambots.</p>
<p>I learned a ton making this system, and I&#8217;ll try to write more about the technical details in a future post.  It&#8217;s interesting to hear people speculate on how it works; everyone gives a different answer.  I guess this goes to show you that search/NLP is still a pretty unsettled, not-completely-understood area.</p>
<p>There are lots of interesting TweetMotif examples.  More prosaic, less news-y queries like <a href="http://tweetmotif.com/#sandwich">sandwich</a> yield cool things like major ingredients of sandwiches and types of sandwiches.  (These are basically distributional similarity candidates for synonym and meronym acquisition, though a bit too noisy to use in its current form.)  And in a few cases, like for understanding currently unfolding events, TweetMotif might even be useful!  It would be nice to expand the set of usefully served queries.  We&#8217;re occasionally posting interesting queries at <a href="http://twitter.com/tweetmotif">twitter.com/tweetmotif</a>.</p>
<p>And oh yeah.  We have a beautiful iPhone interface!</p>
<p><a href="http://anyall.org/blog/wp-content/uploads/2009/05/astro_mike_iphone_cropped.png"><img class="aligncenter size-medium wp-image-523" title="astro_mike_iphone_cropped" src="http://anyall.org/blog/wp-content/uploads/2009/05/astro_mike_iphone_cropped.png" alt="" width="200" height="300" /></a></p>
<p>Check it out folks.  This is a functional prototype, so you can play with it right now at <a href="http://tweetmotif.com/">tweetmotif.com</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/feed/</wfw:commentRss>
		<slash:comments>6</slash:comments>
		</item>
		<item>
		<title>Comparison of data analysis packages: R, Matlab, SciPy, Excel, SAS, SPSS, Stata</title>
		<link>http://brenocon.com/blog/2009/02/comparison-of-data-analysis-packages-r-matlab-scipy-excel-sas-spss-stata/</link>
		<comments>http://brenocon.com/blog/2009/02/comparison-of-data-analysis-packages-r-matlab-scipy-excel-sas-spss-stata/#comments</comments>
		<pubDate>Mon, 23 Feb 2009 20:18:59 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://anyall.org/blog/?p=421</guid>
		<description><![CDATA[Lukas and I were trying to write a succinct comparison of the most popular packages that are typically used for data analysis.  I think most people choose one based on what people around them use or what they learn in &#8230; <a href="http://brenocon.com/blog/2009/02/comparison-of-data-analysis-packages-r-matlab-scipy-excel-sas-spss-stata/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
				<content:encoded><![CDATA[<p><a href="http://lukasbiewald.com/">Lukas</a> and I were trying to write a succinct comparison of the most popular packages that are typically used for data analysis.  I think most people choose one based on what people around them use or what they learn in school, so I&#8217;ve found it hard to find comparative information.  I&#8217;m posting the table here in hopes of useful comments.</p>
<table border="1" cellspacing="0" cellpadding="3" width="100%">
<tbody>
<tr>
<td><strong>Name</strong></td>
<td><strong>Advantages</strong></td>
<td><strong>Disadvantages</strong></td>
<td><strong>Open source?</strong></td>
<td valign="top"><strong>Typical </strong><strong>users</strong></td>
</tr>
<tr>
<td>R</td>
<td>Library support; visualization</td>
<td>Steep learning curve</td>
<td>Yes</td>
<td valign="top">Finance; Statistics</td>
</tr>
<tr>
<td>Matlab</td>
<td>Elegant matrix support; visualization</td>
<td>Expensive; incomplete statistics support</td>
<td>No</td>
<td valign="top">Engineering</td>
</tr>
<tr>
<td>SciPy/NumPy/Matplotlib</td>
<td>Python (general-purpose programming language)</td>
<td>Immature</td>
<td>Yes</td>
<td valign="top">Engineering</td>
</tr>
<tr>
<td>Excel</td>
<td>Easy; visual; flexible</td>
<td>Large datasets</td>
<td>No</td>
<td valign="top">Business</td>
</tr>
<tr>
<td valign="top">SAS</td>
<td valign="top">Large datasets</td>
<td valign="top">Expensive; outdated programming language</td>
<td valign="top">No</td>
<td valign="top">Business; Government</td>
</tr>
<tr>
<td valign="top">Stata</td>
<td valign="top">Easy statistical analysis</td>
<td valign="top"></td>
<td valign="top">No</td>
<td valign="top">Science</td>
</tr>
<tr>
<td valign="top">SPSS</td>
<td valign="top" colspan="4">Like Stata but more expensive and worse</td>
</tr>
</tbody>
</table>
<p>[<b>7/09 update:</b> tweaks incorporating some of the excellent comments below, esp. for SAS, SPSS, and Stata.]</p>
<p>There&#8217;s a bunch more to be said for every cell.  Among other things: <span id="more-421"></span></p>
<ul>
<li>Two big divisions on the table: The more programming-oriented solutions are R, Matlab, and Python.  More analytic solutions are Excel, SAS, Stata, and SPSS.
<li>Python &#8220;immature&#8221;: matplotlib, numpy, and scipy are all separate libraries that don&#8217;t always get along.  Why does matplotlib come with &#8220;pylab&#8221; which is supposed to be a unified namespace for everything?  Isn&#8217;t scipy supposed to do that?  Why is there duplication between numpy and scipy (e.g. numpy.linalg vs. scipy.linalg)?  And then there&#8217;s package compatibility version hell.  You can use SAGE or Enthought but neither is standard (yet).  In terms of functionality and approach, SciPy is closest to Matlab, but it feels much less mature.</li>
<li>Matlab&#8217;s language is certainly weak.  It sometimes doesn&#8217;t seem to be much more than a scripting language wrapping the matrix libraries.  Python is clearly better on most counts.  R&#8217;s is surprisingly good (Scheme-derived, smart use of named args, etc.) if you can get past the bizarre language constructs and weird functions in the standard library.  Everyone says SAS is very bad.</li>
<li>Matlab is the best for developing new mathematical algorithms.  Very popular in machine learning.</li>
<li>I&#8217;ve never used the Matlab Statistical Toolbox.  I&#8217;m wondering, how good is it compared to R?</li>
<li>Here&#8217;s an <a href="http://www.reddit.com/r/programming/comments/7fg6i/why_are_sasstata_the_default_statistical_tools/">interesting reddit thread</a> on SAS/Stata vs R.</li>
<li>SPSS and Stata in the same category: they seem to have a similar role so we threw them together.  Stata is a lot cheaper than SPSS, people usually seem to like it, and it seems popular for introductory courses.  I personally haven&#8217;t used either&#8230;</li>
<li>SPSS and Stata for &#8220;Science&#8221;: we&#8217;ve seen biologists and social scientists use lots of Stata and SPSS.  My impression is they get used by people who want the easiest way possible to do the sort of standard statistical analyses that are very orthodox in many academic disciplines.  (ANOVA, multiple regressions, t- and chi-squared significance tests, etc.)  Certain types of scientists, like physicists, computer scientists, and statisticians, often do weirder stuff that doesn&#8217;t fit into these traditional methods.</li>
<li>Another important thing about SAS, from my perspective at least, is that it&#8217;s used mostly by an older crowd.  I know dozens of people under 30 doing statistical stuff and only one knows SAS.  At that <a href="http://dataspora.com/blog/predictive-analytics-using-r/">R meetup</a> last week, <a href="http://www.meetup.com/R-Users/members/7654264/">Jim Porzak</a> asked the audience if there were any recent grad students who had learned R in school.  Many hands went up.  Then he asked if SAS was even offered as an option.  All hands went down.  There were boatloads of SAS representatives at that conference and they sure didn&#8217;t seem to be on the leading edge.</li>
<li>But: is there ANY package besides SAS that can do analysis for datasets that don&#8217;t fit into memory?  That is, ones that mostly have to stay on disk?  And exactly how good as SAS&#8217;s capabilities here anyway?</li>
<li>If your dataset can&#8217;t fit on a single hard drive and you need a cluster, none of the above will work. There are a few multi-machine data processing frameworks that are somewhat standard (e.g. Hadoop, MPI) but It&#8217;s an open question what the standard distributed data analysis framework will be.  (Hive? Pig?  Or quite possibly something else.)</li>
<li>(This was an interesting point at the R meetup.  Porzak was talking about how going to MySQL gets around R&#8217;s in-memory limitations.  But Itamar Rosenn and Bo Cowgill (Facebook and Google respectively) were talking about multi-machine datasets that require cluster computation that R doesn&#8217;t come close to touching, at least right now.  It&#8217;s just a whole different ballgame with that large a dataset.)</li>
<li>SAS people complain about poor graphing capabilities.</li>
<li>R vs. Matlab visualization support is controversial.  One view I&#8217;ve heard is, R&#8217;s visualizations are great for exploratory analysis, but you want something else for very high-quality graphs.  Matlab&#8217;s interactive plots are super nice though.  Matplotlib follows the Matlab model, which is fine, but is uglier than either IMO.</li>
<li>Excel has a far, far larger user base than any of these other options.  That&#8217;s important to know.  I think it&#8217;s underrated by computer scientist sort of people.  But it does massively break down at &gt;10k or certainly &gt;100k rows.</li>
<li>Another option: Fortran and C/C++.  They are super fast and memory efficient, but tricky and error-prone to code, have to spend lots of time mucking around with I/O, and have zero visualization and data management support.  Most of the packages listed above run Fortran numeric libraries for the heavy lifting.</li>
<li>Another option: Mathematica.  I get the impression it&#8217;s more for theoretical math, not data analysis.  Can anyone prove me wrong?</li>
<li>Another option: the pre-baked data mining packages.  The open-source ones I know of are Weka and Orange.  I hear there are zillions of commercial ones too.  Jerome Friedman, a big statistical learning guy, has an interesting complaint that they should focus more on traditional things like significance tests and experimental design.  (<a href="http://www-stat.stanford.edu/~jhf/ftp/dm-stat.pdf">Here</a>; the article that inspired <a href="http://anyall.org/blog/2008/12/statistics-vs-machine-learning-fight/">this rant</a>.)</li>
<li>I think knowing where the typical users come from is very informative for what you can expect to see in the software&#8217;s capabilities and user community.  I&#8217;d love more information on this for all these options.</li>
</ul>
<p>What do people think?</p>
<hr />
Aug 2012 update: <a href=http://science.webhostinggeeks.com/poredenje-paketa-analize>Serbo-Croatian translation</a>.<br />
Apr 2015 update: <a href="http://offclouds.com/edu/primerjava-analize-podatkov-paketov-r-matlab-scipy-excel-sas-spss-stata/">Slovenian translation</a>.<br />
May 2017 update: <a href="https://www.homeyou.com/~edu/analise-de-dados">Portugese translation</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2009/02/comparison-of-data-analysis-packages-r-matlab-scipy-excel-sas-spss-stata/feed/</wfw:commentRss>
		<slash:comments>185</slash:comments>
		</item>
		<item>
		<title>Statistics vs. Machine Learning, fight!</title>
		<link>http://brenocon.com/blog/2008/12/statistics-vs-machine-learning-fight/</link>
		<comments>http://brenocon.com/blog/2008/12/statistics-vs-machine-learning-fight/#comments</comments>
		<pubDate>Wed, 03 Dec 2008 08:56:30 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://anyall.org/blog/?p=364</guid>
		<description><![CDATA[10/1/09 update &#8212; well, it&#8217;s been nearly a year, and I should say not everything in this rant is totally true, and I certainly believe much less of it now. Current take: Statistics, not machine learning, is the real deal, &#8230; <a href="http://brenocon.com/blog/2008/12/statistics-vs-machine-learning-fight/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
				<content:encoded><![CDATA[<p><b>10/1/09 update</b> &#8212; well, it&#8217;s been nearly a year, and I should say not everything in this rant is totally true, and I certainly believe much less of it now.  Current take: <i>Statistics</i>, not machine learning, is the real deal, but unfortunately suffers from bad marketing.  On the other hand, to the extent that bad marketing includes misguided undergraduate curriculums, there&#8217;s plenty of room to improve for everyone.</p>
<hr />
So it&#8217;s pretty clear by now that statistics and machine learning aren&#8217;t very different fields. I was recently pointed to <a href="http://www-stat.stanford.edu/~tibs/stat315a/glossary.pdf">a very amusing comparison</a> by the excellent statistician &#8212; and machine learning expert &#8212; <a href="http://www-stat.stanford.edu/~tibs/">Robert Tibshiriani</a>.  Reproduced here:</p>
<table border="1" cellpadding="2" style="margin: 0 auto">
<tr>
<td colspan="2" align="center"><b>Glossary</b></p>
<tr>
<th>Machine learning
<th>Statistics </p>
<tr>
<td>network, graphs
<td>model </p>
<tr>
<td>weights
<td>parameters </p>
<tr>
<td>learning
<td>fitting </p>
<tr>
<td>generalization
<td>test set performance </p>
<tr>
<td>supervised learning
<td>regression/classiﬁcation </p>
<tr>
<td>unsupervised learning
<td>density estimation, clustering </p>
<tr>
<td>large grant = $1,000,000
<td>large grant = $50,000 </p>
<tr>
<td>nice place to have a meeting:<br />
Snowbird, Utah, French Alps</p>
<td>nice place to have a meeting:<br />
Las Vegas in August<br />
</table>
<p>Hah.  Or rather, ouch!  I had two thoughts reading this.  (1) Poor statisticians.  Machine learners invent annoying new terms, sound cooler, and have all the fun.  (2) What&#8217;s wrong with statistics?  They have way less funding and influence than it seems they might deserve.</p>
<p>There are several issues going on here, both substantive and cultural: <span id="more-364"></span></p>
<p>There might be too much re-making-up of terms on the ML side.  But lots of these are useful.  &#8220;Weights&#8221; is a great, intuitive term for the parameters of a linear model.  I use it all the time to explain classifiers and regressions to non-experts.  I was surprised to see &#8220;test set&#8221; on the statistics side; I&#8217;m used to thinking of held-out test set accuracy as an extremely common ML technique, while in statistics model fit is assessed with parametric assumptions for standard errors and such.   I really like cross-validation and bootstrapping as ways of thinking about generalization &#8212; again, something that&#8217;s far easier to grasp than sampling and hypothesis testing approaches to parameter inference &#8212; which keep getting taught to and misunderstood by generations of confused Introduction to Statistics students.  For example, how many times has been explained that: No, a p-value is NOT the probability your model is wrong.  But scientific papers regularly treat significance levels in that manner (look how many stars are on this result!)  On the other hand, cross-validation accuracy *is* something you can interpret as being related to the probability your model is right.</p>
<p>I&#8217;ll also note that there are definitely a number of topics in ML that aren&#8217;t very related to statistics or probability.  Max-margin methods: if all we care about is prediction, why bother using a probability model at all?  Why not just optimize the spatial geometry instead?  SVM&#8217;s don&#8217;t require a lick of probability theory to understand.  (Of course probability-based approaches are huge in ML, but it&#8217;s important to remember they&#8217;re not the only game in town, and there is no necessary reason they must be.)  And then there are non-traditional settings such as <a href="http://hunch.net/?p=277">online learning</a>, <a href="http://en.wikipedia.org/wiki/Reinforcement_learning">reinforcement learning</a>, and <a href="http://en.wikipedia.org/wiki/Supervised_learning#Active_Learning">active learning</a>, where the structure of access to information is in play.  There are certainly plenty of things in statistics that aren&#8217;t considered part of ML &#8212; say, regression diagnostics and significance testing.  Finally, many ML problems involve large, high dimensional data and models, where computational issues are very important.  For example, in statistical machine translation, alignment models are described with probability theory and fit to data, but their structure is complex enough that optimal inference is intractable, and how you do approximate inference (EM, Viterbi, beam search, etc.) is a very major issue.</p>
<p>But the most interesting differences between stats and ML are institutional.</p>
<p>I&#8217;ve been hearing lots of friends compare two dueling courses at Stanford: <a href="http://cs229.stanford.edu/">CS229</a>, the CS department&#8217;s &#8220;machine learning&#8221; course taught by <a href="http://robotics.stanford.edu/~ang/">Andrew Ng</a>; and Stat 315 <a href="http://www-stat.stanford.edu/~tibs/stat315a.html">A</a>/<a href="http://www.stanford.edu/class/stats315b/">B</a>, the Statistics department&#8217;s &#8220;statistical learning&#8221; sequence taught by some combination of <a href="http://www-stat.stanford.edu/~tibs/">Tibshirani</a>, <a href="http://www-stat.stanford.edu/~jhf/">Jerome Friedman</a>, and <a href="http://www-stat.stanford.edu/~hastie/">Trevor Hastie</a>.  These people are all top-of-the-line researchers in the field.  Their courses&#8217; contents are extremely similar; I&#8217;d bet any of them could teach most of the material from the other side.</p>
<p>What differs most is the teaching style.  CS has <a href="http://www.stanford.edu/class/cs229/materials.html">far better lecture notes</a>.  Of course, the stats people wrote a <a href="http://www-stat.stanford.edu/~tibs/ElemStatLearn/">very good book</a>; but better lecture notes win because I can access them later and send them to people for free.  CS students I&#8217;ve talked to think the CS course is better taught; I can&#8217;t find stats students who take the CS course.  (My sample is biased, though I know people in both.)  Finally, the CS course has a big, open-ended project component; the Stats course follows more of a traditional problem set and tests format.</p>
<p>I think this is reflective of the differences in institutional culture between CS and Stats. There&#8217;s an interesting John Langford post on part of the issue, which he calls <a href="http://hunch.net/?p=318">&#8220;The Stats Handicap&#8221;</a>.  He points out that stats Ph.D.&#8217;s have a big disadvantage in the job market because statistics has an old-school journal-oriented publishing culture, so students publish much less and have less experience engaging with a research community.  CS is conference-oriented &#8212; certain conferences have a higher prestige than many journals (e.g. <a href="http://nips.cc/">NIPS</a> in ML, <a href="http://www.sigchi.org/sigchi/chi2008/">CHI</a> in HCI) &#8212; and this results in faster turnaround, dissemination, and collaboration.  (I&#8217;ve heard others make similar comparisons between CS and psychology.)  I&#8217;d expect any discipline with a larger conference emphasis to have better courses since they should reward presentation/teaching skills &#8212; or at least encourage practice &#8212; more than in journal world.</p>
<p>ML sounds like it&#8217;s young, vibrant, interesting to learn, and growing; Stats does not.</p>
<p>Is marketing a problem?  Machine learning terms definitely sound pretty cool.  Maybe the perspective of computational intelligence lends itself to cool names.  Though the Stanford statisticians certainly know how to play this game &#8212; for example, they made up their own names for variants of L1 and L2-regularized regression, leaving annoyed people like me forever googling <a href="http://www-stat.stanford.edu/~tibs/lasso.html">&#8220;lasso&#8221;</a> and <a href="http://en.wikipedia.org/wiki/Ridge_regression">&#8220;ridge&#8221;</a> trying to remember which is which.  (On the other hand, perhaps that&#8217;s child&#8217;s play compared to the true original sin of ML nomenclature: tossing around the highly deceptive term &#8220;neural network&#8221; for a stack of linear functions paired with a <a href="http://en.wikipedia.org/wiki/Backpropagation">wonky, overhyped training algorithm</a>; the combination of which, many years later, still <a href="http://www.stat.columbia.edu/~cook/movabletype/archives/2008/07/predicting-deat.html">causes confusion</a>.  Definitely blame CS for that one.)</p>
<p>Another issue is the definition of statistics itself.  In 1997, Jerome Friedman wrote an extremely interesting analysis of the situation: <a href="http://www-stat.stanford.edu/~jhf/ftp/dm-stat.pdf">&#8220;﻿Data Mining and Statistics: What&#8217;s the Connection?&#8221;</a>.  He points out, quite correctly, the statistical impoverishment of some common approaches to data mining.  You can certainly blame statistics for not marketing its ideas well enough, or blame CS for ignoring statistics.  For example there&#8217;s a good case that lots of genetic algorithms and neural network research was much ado about nothing &#8212; that is, over-complicated cool-sounding hammers looking for nails when all you needed were some time-honored statistical and optimization techniques.  (E.g. why NN when you haven&#8217;t tried a straight-up <a href="http://en.wikipedia.org/wiki/Generalized_linear_model">GLM</a>?  Why GA when you haven&#8217;t tried <a href="http://en.wikipedia.org/wiki/Nelder-Mead_method">Nelder-Mead</a>?)  But this problem has been rectified somewhat &#8212; for example, NLP has seen a big move to simple linear models as the default technique, and NN&#8217;s and GA&#8217;s have fallen from grace in mainstream ML.</p>
<p>Friedman argues part of the problem is in how statisticians approach problems and the world:</p>
<blockquote><p>
One can catalog a long history of Statistics (as a field) ignoring useful methodology developed in other data related fields.  Here are some of them that had seminal beginnings in Statistics but for the most part were subsequently ignored in our field: Pattern Recognition, Neural Networks, Machine Learning, Graphical Models, Chemometrics, Data Visualization.
</p></blockquote>
<p>That is not to say statistics is not important &#8212; it&#8217;s incredibly important.  He quotes Efron as saying &#8220;Statistics has been the most successful information science.&#8221;  However, information science is becoming bigger and broader and more exciting, thanks to computation and ever-increasing amounts of data.  What should statisticians do?  Friedman continues (light editing and emphasis is mine):</p>
<blockquote><p>
One view says that our field should concentrate on that small part of information science that we do best, namely probabilistic inference based on mathematics. If this view is adopted, we should become resigned to the fact that the role of Statistics as a player in the &#8220;information revolution&#8221; will steadily diminish over time.</p>
<p>Another point of view holds that statistics ought to be concerned with <b>data analysis</b>.  The field should be defined in terms of a set of <i>problems</i> &#8212; rather than a set of tools &#8212; that pertain to data.  Should this point of view ever become the dominant one, a big change would be required in our practice and academic programs.</p>
<p>First and foremost, we would have to make peace with computing.  It&#8217;s here to stay; that&#8217;s where the data is.  This has been one of the most glaring omissions in the set of tools that have so far defined Statistics.  Had we incorporated computing methodology from its inception as a fundamental statistical tool (as opposed to simply a convenient way to apply our existing tools) many of the other data related fields would not have needed to exist.  They would have been part of our field.
</p></blockquote>
<p>Friedman wrote this article more than 10 years ago.  All his observations about the importance and increasing prevalence of data and computing power are even more true today than back then.  Has the field of statistics changed?  Not clear.  (I&#8217;d appreciate seeing evidence to the contrary.)</p>
<p>On the other hand a world of data *has* to be increasingly statistical.  The positive spin <a href="http://www-stat.stanford.edu/~ckirby/brad/papers/2005NEWModernScience.pdf">from Efron</a>:</p>
<blockquote><p>
A new generation of scientiﬁc devices, typiﬁed by microarrays, produce data on a gargantuan scale – with millions of data points and thousands of parameters to consider at the same time. These experiments are “deeply statistical”. Common sense, and even good scientiﬁc intuition, won’t do the job by themselves. Careful statistical reasoning is the only way to see through the haze of randomness to the structure underneath. Massive data collection, in astronomy, psychology, biology, medicine, and commerce, is a fact of 21st Century science, and a good reason to buy statistics futures if they are ever offered on the NASDAQ.
</p></blockquote>
<p>I know that I&#8217;m interested in quantitative information science, including statistics and data analysis.  Machine learning has many strengths, but it is definitely an odd way to go about analysis.  But there&#8217;s a good case that statistics, as traditionally defined, is only going to have a smaller role in the future.  &#8220;Data mining&#8221; sounds more relevant, but does it even exist as a coherent subject?  Maybe it&#8217;s time to study a more applied statistical field like <a href="http://en.wikipedia.org/wiki/Econometrics">econometrics</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2008/12/statistics-vs-machine-learning-fight/feed/</wfw:commentRss>
		<slash:comments>132</slash:comments>
		</item>
		<item>
		<title>It is accurate to determine a blog&#8217;s bias by what it links to</title>
		<link>http://brenocon.com/blog/2008/10/it-is-accurate-to-determine-a-blogs-bias-by-what-it-links-to/</link>
		<comments>http://brenocon.com/blog/2008/10/it-is-accurate-to-determine-a-blogs-bias-by-what-it-links-to/#comments</comments>
		<pubDate>Sat, 11 Oct 2008 10:12:00 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://anyall.org/blog/?p=189</guid>
		<description><![CDATA[Here&#8217;s a great project from Andy Baio and Joshua Schachter: they assessed the political biases of different blogs based on which articles they tend link to. Using these political bias scores, they made a cool little Firefox extension that colors &#8230; <a href="http://brenocon.com/blog/2008/10/it-is-accurate-to-determine-a-blogs-bias-by-what-it-links-to/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
				<content:encoded><![CDATA[<p><a href="http://waxy.org/2008/10/memeorandum_colors/">Here&#8217;s a great project from Andy Baio and Joshua Schachter</a>: they assessed the political biases of different blogs based on which articles they tend link to.  Using these political bias scores, they made a cool little Firefox extension that colors the names of different sources on the news aggregator site <a href="http://memeorandum.com/">Memeorandum</a>, like so:</p>
<p><img src="http://waxy.org/random/images/weblog/memeorandum_beforeafter.jpg"></p>
<p>How they computed these biases is pretty neat.  Their data source was the Memeorandum site itself, which shows a particular news story, then a list of different news sites that have written articles about the topic.  Scraping out that data, Joshua constructed the adjacency matrix of sites vs. articles they linked to and ran good ol&#8217; <a href="http://en.wikipedia.org/wiki/Singular_value_decomposition">SVD</a> on it, an algorithm that can be used to summarize the very high-dimensional article linking information in just several numbers (&#8220;components&#8221; or &#8220;dimensions&#8221;) for each news site.  Basically, the algorithm groups together sites that tend to link to the same articles.  It&#8217;s not exactly clustering though; rather, it projects them into a space where sites close to each other had similar linking patterns.  People have used this technique analogously to construct a political spectrum for Congress, by analyzing which legislators tend to vote together.</p>
<p>So here they found that the second dimension of the SVD&#8217;s projected outputs seemed to strongly correlate with their own intuitions of sites&#8217; political biases.  Talk about getting lucky!  This score is used for their coloring visualization, and I personally found the examples pretty accurate.  And they helpfully posted all of their output data <a href="http://waxy.org/2008/10/memeorandum_colors/">with the blog post</a>.</p>
<p>There is a concern though.  The funny thing about SVD (and related algorithms like factor analysis and <a href="http://en.wikipedia.org/wiki/Principal_components_analysis">PCA</a>) is that the numbers that fall out of it don&#8217;t necessarily mean anything.  In fact there have been great controversies when researchers try to interpret its outputs.  For example, if you run PCA on scores from different types of IQ tests, you get a &#8220;<i>g</i> factor&#8221;.  Is it a measure of <a href="http://en.wikipedia.org/wiki/General_intelligence_factor">general human intelligence</a>?  Or is <i>g</i> just a <a href="http://cscs.umich.edu/~crshalizi/weblog/523.html">meaningless statistical artifact</a>?  No one&#8217;s sure.</p>
<p>But for this problem, there is a fair, objective validation &#8212; use 3rd party, human judgments from the web!  I&#8217;ve found before that you can <a href="http://blog.doloreslabs.com/2008/03/crowdsourcing-to-find-media-bias-hillary-vs-obama/">assess media bias on AMT</a> pretty well; but for this, I simply went to a pre-existing site called <a href="http://www.skewz.com">Skewz</a>, which <a href="http://www.skewz.com/source/compare_sources">collects people&#8217;s ratings</a> of the bias of individual articles from news sites.  About 150 sites were rated on Skewz as well as included in the Memeorandum/SVD analysis.</p>
<p>Within that set, it turns out that the SVD&#8217;s second component <b>significantly correlates</b> with Skewz users&#8217; judgments of political bias!  First, here&#8217;s the scatterplot of the &#8220;v2&#8243; SVD dimension against Skewz ratings.  Higher numbers are conservative, lower are liberal:</p>
<p><a href="http://anyall.org/blog/wp-content/uploads/2008/10/pic1.png"><img src="http://anyall.org/blog/wp-content/uploads/2008/10/pic1.png" alt="" title="SVD correlates with political bias judgments" width="500" height="515" class="aligncenter size-full wp-image-193" /></a></p>
<p>So SVD tends to give most sites a neutral score, but when it assigns a strong score, it&#8217;s often right &#8212; or at least, correlates with Skewz users.  Some of the disagreements are interesting &#8212; for example, Skewz thinks The New Republic is liberal, whereas SVD thinks it&#8217;s slightly conservative.  That might mean that TNR links to lots of stories that conservatives tend to like, though its actual content and stances are liberal.  (But don&#8217;t take any particular data point too seriously &#8212; the Skewz data is probably fairly noisy, and the bridging between the datasets introduces more noise too, since Memeorandum and Skewz are based on different sets of articles and such.)</p>
<p>Here&#8217;s a zoom-in on that narrow band in the middle.  There&#8217;s some more successful correlation in there:</p>
<p><a href="http://anyall.org/blog/wp-content/uploads/2008/10/pic2.png"><img src="http://anyall.org/blog/wp-content/uploads/2008/10/pic2.png" alt="" title="SVD correlates with political bias judgments -- zoom-in" width="500" height="515" class="aligncenter size-full wp-image-194" /></a></p>
<p>Here are the actual correlation coefficients with the different SVD outputs.  It turns out the first dimension slightly correlates to political bias as well.  (Joshua explained it as the overall volume of linking.  Do liberals tend to link more?)  But the third through fifth dimensions, which they say were very hard for them to interpret, don&#8217;t correlate at all to these political bias ratings.</p>
<table>
<tr>
<td>SVD component (output dimension) </p>
<th>v1
<th>v2
<th>v3
<th>v4
<th>v5</p>
<tr>
<td>Correlation to Skewz ratings</p>
<td>&nbsp;+.112&nbsp;
<td>&nbsp;+.392&nbsp;
<td>&nbsp;-.011&nbsp;
<td>&nbsp;-.057&nbsp;
<td>&nbsp;-.047&nbsp;<br />
</table>
<p>In conclusion &#8230; this overall result makes me really happy.  A completely unsupervised algorithm, based purely on similarity of linking patterns, gets you a systematic correlation with independent judges&#8217; assessments of bias.  That&#8217;s just sweet.</p>
<p>
Here&#8217;s the entire dataset for the above graphs.  &#8220;score_svd&#8221; is their rescaled version of &#8220;v2&#8243;.  (<a href="http://spreadsheets.google.com/ccc?key=ph4VVGBJ3IpLykg-S2MifWA">Click here</a> to see and download all of it).</p>
<p><iframe width=500 height=200 frameborder=0 src=http://spreadsheets.google.com/pub?key=ph4VVGBJ3IpLykg-S2MifWA></iframe></p>
<p><b>Update:</b> See the comments below.  You can also fit a linear model against all of v1..v5 to predict the Skewz rating as the response.  This fits a little better than using just v2.  Here&#8217;s the scatterplot for the model&#8217;s predictions.<br />
<a href="http://anyall.org/blog/wp-content/uploads/2008/10/pic3.png"><img src="http://anyall.org/blog/wp-content/uploads/2008/10/pic3.png" alt="" title="SVD correlates with political bias judgments" width="290" height="300" class="aligncenter size-medium wp-image-193" /></a></p>
<hr />
<p>Code: I put the Skewz scraper, data, and scripts <a href="http://gist.github.com/16240">up here</a>.</p>
<p>Final note: The correlation coefficients above are via <a href="http://en.wikipedia.org/wiki/Kendall_tau_rank_correlation_coefficient">Kendall-Tau</a>, which is invariant to rescalings of the data.  This data has all sorts of odd spikes and such, and Joshua and Andy themselves rescaled the data for the coloring plugin, so this seemed safest.  And don&#8217;t worry about the small sample size; v1 correlation&#8217;s p-value is .04, v2 correlation p-value is tiny.</p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2008/10/it-is-accurate-to-determine-a-blogs-bias-by-what-it-links-to/feed/</wfw:commentRss>
		<slash:comments>12</slash:comments>
		</item>
		<item>
		<title>Turker classifiers and binary classification threshold calibration</title>
		<link>http://brenocon.com/blog/2008/06/turker-classifiers-and-binary-classification-threshold-calibration/</link>
		<comments>http://brenocon.com/blog/2008/06/turker-classifiers-and-binary-classification-threshold-calibration/#comments</comments>
		<pubDate>Wed, 18 Jun 2008 09:25:00 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://blog.anyall.org/?p=122</guid>
		<description><![CDATA[I wrote a big Dolores Labs blog post a few days ago. Click here to read it. I am most proud of the pictures I made for it:]]></description>
				<content:encoded><![CDATA[<p>I wrote a big Dolores Labs blog post a few days ago.  <a href="http://blog.doloreslabs.com/?p=61">Click here to read it</a>.  I am most proud of the pictures I made for it:</p>
<p><img src="http://blog.doloreslabs.com/wp-content/uploads/2008/06/vertthresh.png"></p>
<p><img src="http://blog.doloreslabs.com/wp-content/uploads/2008/06/confusionbars.png"></p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2008/06/turker-classifiers-and-binary-classification-threshold-calibration/feed/</wfw:commentRss>
		<slash:comments>1</slash:comments>
		</item>
		<item>
		<title>Are women discriminated against in graduate admissions? Simpson&#8217;s paradox via R in three easy steps!</title>
		<link>http://brenocon.com/blog/2008/04/are-women-discriminated-against-in-graduate-admissions-simpsons-paradox-via-r-in-three-easy-steps/</link>
		<comments>http://brenocon.com/blog/2008/04/are-women-discriminated-against-in-graduate-admissions-simpsons-paradox-via-r-in-three-easy-steps/#comments</comments>
		<pubDate>Sun, 13 Apr 2008 09:04:00 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://blog.anyall.org/?p=116</guid>
		<description><![CDATA[R has a fun built-in package, datasets: a whole bunch of easy-to-use, interesting tables of data. I found the famous UC Berkeley admissions data set, from a 1970&#8242;s study of whether sex discrimination existed in graduate admissions. It&#8217;s famous for &#8230; <a href="http://brenocon.com/blog/2008/04/are-women-discriminated-against-in-graduate-admissions-simpsons-paradox-via-r-in-three-easy-steps/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
				<content:encoded><![CDATA[<p><a href=http://www.statmethods.net/>R</a> has a fun built-in package, <a href=http://finzi.psych.upenn.edu/R/library/datasets/html/00Index.html>datasets</a>: a whole bunch of easy-to-use, interesting tables of data.  I found the famous UC Berkeley admissions data set, from a 1970&#8242;s study of whether sex discrimination existed in graduate admissions.  It&#8217;s famous for illustrating a particular statistical paradox.  Thanks to R&#8217;s awesome mosaic plots interface, we can see this really easily.</p>
<p>UCBAdmissions is a three-dimensional table (like a matrix): Admit Status x Gender x Dept, with counts for each category as the matrix&#8217;s values.  R&#8217;s default printing shows the basics just fine.  Here&#8217;s the data for just the first of six departments:</p>
<blockquote><pre>&gt; <i>UCBAdmissions</i>
, , Dept = A

          Gender
Admit      Male Female
  Admitted  512     89
  Rejected  313     19

...
</pre>
</blockquote>
<p>Overall, <b>women have a lower admittance rate than men</b>:</p>
<blockquote><pre>&gt; <i>apply(UCBAdmissions,c(1,2),sum)</i>

          Gender
Admit         M    F
  Admitted 1198  557
  Rejected 1493 1278
</pre>
</blockquote>
<p>This is the phenomenon that prompted a lawsuit against Berkeley which prompted the study that collected this data.</p>
<p>R&#8217;s plot function is overloaded to do a mosaic plot for this sort of categorical data.  Very cool.  With just</p>
<blockquote><pre>&gt; plot(UCBAdmissions)
</pre>
</blockquote>
<p>or, playing around after reading <a href=http://www.statmethods.net/advgraphs/mosaic.html>Quick-R&#8217;s page on this</a>:</p>
<blockquote><pre>&gt; install.packages(”vcd”)
> library(vcd)
> mosaic(UCBAdmissions, condvars=c('Dept'))
</pre>
</blockquote>
<p>We have a plot showing admittance and gender breakdowns per department:</p>
<p><center><img src="http://farm4.static.flickr.com/3155/2409960034_1fb97ffbd1_o.png"></center></p>
<p><b>In each department, women have similar admittance rates as men.</b>  This seems to be at odds with the fact that women have a lower admittance rate overall.  This discrepancy is an example of <a href=http://en.wikipedia.org/wiki/Simpson%27s_paradox>Simpson&#8217;s paradox</a>.</p>
<p>This mosaic also shows the explanation: <b>Selective departments have more female applicants.</b>  It&#8217;s easy to see since the departments are ordered by selectiveness.  Departments A and B let in many applicants, but they&#8217;re mostly male.  The reverse is true for the rest.  This means that the overall female population takes big admittance hits in departments C through F, while lots of males get in via departments A and B.</p>
<p>I think these mosaic plots are impressive for visualizing categorical proportions for high dimensional data sets.  Well, by “high” I think I mean, more than 2.  I can&#8217;t think of a better way to see several cross relationships in categorical data at once.  And the only tuning I needed to do was play around a bit with the order of those three dimensions.</p>
<p>Sources:</p>
<ul>
<li>R&#8217;s <a href=http://finzi.psych.upenn.edu/R/library/datasets/html/UCBAdmissions.html>UCBAdmissions</a> help page.  It comes with the standard download of R.</p>
<li>R&#8217;s <a href=http://finzi.psych.upenn.edu/R/library/vcd/html/mosaic.html>vcd::mosaic</a> function.  I recommend the <a href=http://cran.r-project.org/web/packages/vcd/vignettes/strucplot.pdf>pdf vigenette</a> about it, which has many more pictures of cool mosaic plots.
<li><strike>I would post the original 1975 Science paper, but it&#8217;s not freely available.  I hate academic publishers.</strike>  Here&#8217;s the paper, at least for now:
<ul>
<li> Bickel, P. J., Hammel, E. A., and O&#8217;Connell, J. W. (1975) Sex bias in graduate admissions: Data from Berkeley. Science, 187, 398–403.  <a href="http://anyall.org/science_1975_sex_bias_graduate_admissions_data_berkeley.pdf" rel="nofollow">[PDF]</a></ul>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2008/04/are-women-discriminated-against-in-graduate-admissions-simpsons-paradox-via-r-in-three-easy-steps/feed/</wfw:commentRss>
		<slash:comments>8</slash:comments>
		</item>
		<item>
		<title>color name study i did</title>
		<link>http://brenocon.com/blog/2008/03/color-name-study-i-did/</link>
		<comments>http://brenocon.com/blog/2008/03/color-name-study-i-did/#comments</comments>
		<pubDate>Tue, 18 Mar 2008 16:54:00 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://blog.anyall.org/?p=108</guid>
		<description><![CDATA[Link: Where does “Blue” end and “Red” begin? I&#8217;m writing some posts on blog.doloreslabs.com and this is the best one so far. Methodology-wise, along the lines of my earlier Amazon Mechanical Turk moral decisions survey&#8230;]]></description>
				<content:encoded><![CDATA[<p>Link: <a href="http://blog.doloreslabs.com/?p=11">Where does “Blue” end and “Red” begin?</a></p>
<p>I&#8217;m writing some posts on blog.doloreslabs.com and this is the best one so far.  Methodology-wise, along the lines of my earlier Amazon Mechanical Turk <a href="http://socialscienceplusplus.blogspot.com/2008/01/moral-psychology-on-amazon-mechanical.html">moral decisions survey</a>&#8230;</p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2008/03/color-name-study-i-did/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Food Fight</title>
		<link>http://brenocon.com/blog/2008/01/food-fight/</link>
		<comments>http://brenocon.com/blog/2008/01/food-fight/#comments</comments>
		<pubDate>Thu, 31 Jan 2008 07:21:00 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://blog.anyall.org/?p=104</guid>
		<description><![CDATA[Absolutely amazing &#8212; a short film chronicling conflicts from World War II &#8212; as food. I think this has to have the highest amount of Wikipedia-linkable references per second of any film I&#8217;ve seen. Yes, it&#8217;s U.S.-centric, but so is &#8230; <a href="http://brenocon.com/blog/2008/01/food-fight/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
				<content:encoded><![CDATA[<p>Absolutely amazing &#8212; a <a href="http://homepage.mac.com/stefannadelman/foodfight/index.htm">short film</a> chronicling conflicts from World War II &#8212; as food.</p>
<p><embed src="http://www.atomfilms.com:80/a/autoplayer/shareEmbed.swf?keyword=food_fight" width="426" height="350"></embed></p>
<p>I think this has to have the highest amount of Wikipedia-linkable references per second of any film I&#8217;ve seen.  Yes, it&#8217;s U.S.-centric, but so is Wikipedia, which makes cataloguing it easier.  At the very least:
<ul>
<li><a href="http://en.wikipedia.org/wiki/World_War_II">World War II</a>
<ul>
<li><a href="http://en.wikipedia.org/wiki/Racial_policy_of_Nazi_Germany">Persecution of Jews</a> and <a href="http://en.wikipedia.org/wiki/The_Holocaust">The Holocaust</a></li>
<li><a href="http://en.wikipedia.org/wiki/Battle_of_France">Invasion of France</a></li>
<li><a href="http://en.wikipedia.org/wiki/The_Blitz">The Blitz</a></li>
<li><a href="http://en.wikipedia.org/wiki/Attack_on_Pearl_Harbor">Pearl Harbor</a>, <a href="http://en.wikipedia.org/wiki/Pacific_Ocean_theater_of_World_War_II">Pacific Theater</a></li>
<li><a href="http://en.wikipedia.org/wiki/Normandy_Campaign">D-Day</a></li>
<li><a href="http://en.wikipedia.org/wiki/Liberation_of_Paris">Liberation of France</a><br />Invasion of Germany &#8211; <a href="http://en.wikipedia.org/wiki/Western_Front_%28World_War_II%29">Western</a> and <a href="http://en.wikipedia.org/wiki/Eastern_Front_%28World_War_II%29">Eastern</a> fronts</li>
<li><a href="http://en.wikipedia.org/wiki/Atomic_bombings_of_Hiroshima_and_Nagasaki">Atomic bombing of Hiroshima</a></li>
</ul>
</li>
<li><a href="http://en.wikipedia.org/wiki/1948_Arab-Israeli_War">1948 Arab-Israeli War</a></li>
<li><a href="http://en.wikipedia.org/wiki/Korean_War">Korean War</a></li>
<li><a href="http://en.wikipedia.org/wiki/Cuban_Missile_Crisis">Cuban Missile Crisis</a></li>
<li><a href="http://en.wikipedia.org/wiki/Vietnam_War">Vietnam War</a>
<ul>
<li><a href="http://en.wikipedia.org/wiki/First_Indochina_War">French Indochina War</a></li>
<li><a href="http://en.wikipedia.org/wiki/Role_of_United_States_in_the_Vietnam_War">U.S. involvement</a></li>
</ul>
</li>
<li><a href="http://en.wikipedia.org/wiki/Nuclear_arms_race">US/USSR nuclear arms race</a></li>
<li><a href="http://en.wikipedia.org/wiki/First_Gulf_War">First Gulf War<br /></a>
<ul><a href="http://en.wikipedia.org/wiki/First_Gulf_War"></a>
<li><a href="http://en.wikipedia.org/wiki/Iraqi_invasion_of_Kuwait">Invasion of Kuwait</a></li>
</ul>
<li><a href="http://en.wikipedia.org/wiki/Israeli-Palestinian_conflict">Israeli-Palestinian conflict</a>
<ul>
<li>Judging from its timing in the film, maybe specifically the <a href="http://en.wikipedia.org/wiki/First_Intifada">First Intifada</a>?</li>
</ul>
</li>
<li><a href="http://en.wikipedia.org/wiki/September_11%2C_2001_attacks">9/11 attacks</a></li>
<li><a href="http://en.wikipedia.org/wiki/War_in_Afghanistan_(2001%E2%80%93present)">War in Afghanistan</a>
<ul>
<li>Taliban falls, but <a href="http://en.wikipedia.org/wiki/Islamic_Emirate_of_Waziristan">some escape</a>, with <a href="http://en.wikipedia.org/wiki/Location_of_Osama_bin_Laden">Osama bin Laden</a></li>
</ul>
</li>
<li><a href="http://en.wikipedia.org/wiki/Iraq_War">2003 Iraq war</a> and <a href="http://en.wikipedia.org/wiki/Post-invasion_Iraq,_2003%E2%80%932006">aftermath</a></li>
</ul>
<p>Many lessons learned about food violence, 20th century war, and Wikipedia <a href="http://en.wikipedia.org/wiki/Meronymy">meronymic</a> article relationships&#8230;</p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2008/01/food-fight/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Moral psychology on Amazon Mechanical Turk</title>
		<link>http://brenocon.com/blog/2008/01/moral-psychology-on-amazon-mechanical-turk/</link>
		<comments>http://brenocon.com/blog/2008/01/moral-psychology-on-amazon-mechanical-turk/#comments</comments>
		<pubDate>Sun, 20 Jan 2008 01:44:00 +0000</pubDate>
		<dc:creator>brendano</dc:creator>
				<category><![CDATA[Best Posts]]></category>

		<guid isPermaLink="false">http://blog.anyall.org/?p=102</guid>
		<description><![CDATA[There&#8217;s a lot of exciting work in moral psychology right now. I&#8217;ve been telling various poor fools who listen to me to read something from Jonathan Haidt or Joshua Greene, but of course there&#8217;s a sea of too many articles &#8230; <a href="http://brenocon.com/blog/2008/01/moral-psychology-on-amazon-mechanical-turk/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
				<content:encoded><![CDATA[<p>There&#8217;s a lot of exciting work in moral psychology right now.  I&#8217;ve been telling various poor fools who listen to me to read something from <a href="http://people.virginia.edu/~jdh6n/">Jonathan Haidt</a> or <a href="http://www.wjh.harvard.edu/~jgreene/">Joshua Greene</a>, but of course there&#8217;s a sea of too many articles and books of varying quality and intended audience.  But just last week Steven Pinker wrote a great NYT magazine article, <a href="http://www.nytimes.com/2008/01/13/magazine/13Psychology-t.html">&#8220;The Moral Instinct,&#8221;</a> which summarizes current research and tries to spell out a few implications.  I recommend it highly, if just for presenting so many awesome examples.  (Yes, this blog has <a href=/2007/11/how-did-freud-become-respected-humanist.html>poked fun</a> at Pinker before.  But in any case, he is a brilliant expository writer.  <a href=http://pinker.wjh.harvard.edu/books/tli/>The Language Instinct</a> is still one of my favorite popular science books.)</p>
<p>For a while now I&#8217;ve been thinking that recruiting subjects online could lend itself to collecting some really interesting behavioral science data.  A few months ago I tried doing this with <a href=http://www.mturk.com/>Amazon Mechanical Turk</a>, a horribly misnamed web service that actually lets you create web-based tasks and pay online workers do them.  Its canonical commercial applications include tedious tasks like <a href=http://www.techcrunch.com/2007/10/21/powerset-testing-results-at-mechanical-turk/>search quality</a> evaluation or image labeling, where you really need human data to perform well.  You put up, say, several thousand images you want classified as &#8220;porn&#8221; or &#8220;not-porn&#8221;, say you&#8217;ll pay workers $0.01 to label ten images, then sit back and watch the data roll in.</p>
<p>So AMT advertises itself as a data annotation or machine learning substitute system, but I think its main innovation is finding out that there are lots and lots of people with free time willing to do online work for very, very low amounts of money.  You can run any task you want, including surveys, and people happily respond for mere pennies.  (Far below minimum wage, I might add &#8212; their motivation seems to be more like casual gaming or so.)  To that end, I tried out running one of the standard moral psych survey questions to see what would happen &#8212; the so-called <a href=http://en.wikipedia.org/wiki/Trolley_problem>&#8220;trolley problem&#8221;</a>:</p>
<blockquote><p>A runaway trolley is hurtling down a track towards five people who have been tied down in its path.  If nothing happens, they will be killed.  Fortunately, you have a switch which would divert the trolley to a different track.  Unfortunately, the other track has one person tied down to it.  Should you flip the switch?</p></blockquote>
<p>It&#8217;s supposed to be a classic dilemma of <a href=http://plato.stanford.edu/entries/consequentialism/>consequentialist</a> vs. <a href=http://plato.stanford.edu/entries/ethics-deontological/>deontological</a> moral reasoning.  Is it acceptable to sacrifice for the greater good?  Is it permissible to take an action that will cause a preventable death?  And so on.  I think it&#8217;s neat just because when I pose it to people, different folks really do disagree, give different answers, and are willing to argue about it.  There are some interesting recent fMRI findings (due to Greene I think?) that people who refuse to flip the switch seem to be engaged in a more emotional response, whereas those who do seem to be using deliberative reasoning systems.  (Some, like Greene and Pinker, seem to go further and argue this is a substantive normative reason to favor flipping the switch; whether you feel like getting sucked into that debate, though, there&#8217;s clearly something interesting happening here.)</p>
<p>So I ran this on AMT; the particpants (they call themselves <a href=http://turkers.proboards80.com/>&#8220;turkers&#8221;</a>) had to answer yes or no.  Turns out 77% say they&#8217;d flip the tracks.</p>
<p>I also ran two variant scenarios of the same logical dilemma, to sacrifice one person to save five:<br />
<blockquote>A trolley is hurtling down a track towards five people.  You are on a bridge under which it will pass, and you can stop it by dropping a heavy weight in front of it.  As it happens, there is a very fat man next to you &#8211; your only way to stop the trolley is to push him over the bridge and onto the track, killing him to save five.  Should you proceed?</p></blockquote>
<p>and<br />
<blockquote>A brilliant transplant surgeon has five patients, each in need of a different organ, each of whom will die without that organ. Unfortunately, there are no organs available to perform any of these five transplant operations. A healthy young traveler, just passing through the city the doctor works in, comes in for a routine checkup. In the course of doing the checkup, the doctor discovers that his organs are compatible with all five of his dying patients. Suppose further that if the young man were to disappear, no-one would suspect the doctor.  Should the doctor sacrifice the man to save his other patients?</p></blockquote>
<p>These two, of course, feel a lot harder to say &#8220;Yes&#8221; to, but if you were willing to say &#8220;Yes&#8221; to the original question, it is hard to justify why.  The participants&#8217; repsonses followed what you would expect: fewer said &#8220;Yes&#8221; to these scenarios.  Here are the Yes/No responses to each of the questions (100 responses for each):</p>
<table border=1>
<tr>
<th>Question
<th>Yes
<th>No<br />
<tr>
<td>surgeon
<td>2
<td>98<br />
<tr>
<td>fat man
<td>30
<td>70<br />
<tr>
<td>switch, save 5
<td>77
<td>23<br />
<tr>
<td>switch, save 10
<td>82
<td>18<br />
<tr>
<td>switch, save 15
<td>83
<td>17<br />
<tr>
<td>switch, save 20
<td>83
<td>17<br /></table>
<p>Only two people thought it was acceptable to sacrifice for organs, and only half as many would push the fat man as would flip the switch.  I also ran variants of the switch version with more and more people on the tracks; the Yes response creeps upwards but never reaches 100%.  The differences among the first three questions are statistically significant (unpaired t-tests, all p<.001 (this seems like the wrong test, can anyone correct me?)).</p>
<p>What&#8217;s amazing is how fast responses happen.  I started getting responses just minutes after posting the question.  I actually posted each of the six questions as a separate, standalone task; but many of the turkers who did one found the rest in the task pool and did them too.  (So what was supposed to be a between-subjects design fell into something else, oops!)  The whole thing cost $6 and was done in a matter of hours.  It&#8217;s very encouraging &#8212; AMT allows you to very quickly iterate and try out different designs and such.  It&#8217;s a bit of a pain to use, though; Amazon has certainly done a poor job in exploiting its full potential.  (They have a form builder which was good enough to quickly write up these tasks, but to do anything moderately sophisticated, even just getting your data back out, you have to write programs against their <a href=http://aws.amazon.com/mturk>somewhat mediocre API</a>; you have to know how to use an XML parser, etc.  <a href=http://doloreslabs.com/>Hm.</a>)</p>
<p>I also tried an explicitly within-subject version, where each participant answered the three basic versions.  I was interested in consistency &#8212; presumably very few people would sacrifice for organs but refuse to divert the trolley.  For 141 participants, here are the frequencies of the different answer triples:</p>
<table border=1>
<tr>
<th>% with this response triple
<th>flip switch?
<th>push fat man?
<th>sacrifice traveler for organs?<br />
<tr>
<td>42.6
<td>Y
<td>N
<td>N<br />
<tr>
<td>29.8
<td>Y
<td>Y
<td>N<br />
<tr>
<td>20.6
<td>N
<td>N
<td>N<br />
<tr>
<td>5.0
<td>Y
<td>Y
<td>Y<br />
<tr>
<td>0.7
<td>Y
<td>N
<td>Y<br />
<tr>
<td>0.7
<td>N
<td>Y
<td>Y<br />
<tr>
<td>0.7
<td>N
<td>Y
<td>N<br /></table>
<p>I personally find the most common responses coherent with my own gut reactions &#8212; from left to right, I feel less and less good about sacrificing in each case.  Perhaps all people feel the same gut reactions, and use different ad hoc reasons to draw the line in different places?</p>
<p>I&#8217;m sorry that this post started with neat moral psychology then degenerated into methodology, but hey it&#8217;s fun.  I&#8217;ve seen only two instances of any sort of research paper being written using AMT, both by computer scientists; here&#8217;s a nice blog post on an <a href=http://behind-the-enemy-lines.blogspot.com/2007/08/experiences-using-amazon-mechanical.html>information retrieval experiment</a> (it&#8217;s a great blog, btw); and someone mentioned to me <a href=http://www2007.org/htmlpapers/paper461/>this one on data processing accuracy</a> also.  Anyone know of any?  It&#8217;s clearly an interesting approach.</p>
]]></content:encoded>
			<wfw:commentRss>http://brenocon.com/blog/2008/01/moral-psychology-on-amazon-mechanical-turk/feed/</wfw:commentRss>
		<slash:comments>21</slash:comments>
		</item>
	</channel>
</rss>
