<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
		>
<channel>
	<title>Comments on: Announcing TweetMotif for summarizing twitter topics</title>
	<atom:link href="https://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/feed/" rel="self" type="application/rss+xml" />
	<link>https://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/</link>
	<description>cognition, language, social systems; statistics, visualization, computation</description>
	<lastBuildDate>Tue, 25 Nov 2025 13:11:20 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	
	<item>
		<title>By: brendano</title>
		<link>https://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/#comment-166497</link>
		<dc:creator>brendano</dc:creator>
		<pubDate>Thu, 19 Jul 2012 14:20:03 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=515#comment-166497</guid>
		<description><![CDATA[It was too annoying to maintain :)]]></description>
		<content:encoded><![CDATA[<p>It was too annoying to maintain :)</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Joseph Turian</title>
		<link>https://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/#comment-166443</link>
		<dc:creator>Joseph Turian</dc:creator>
		<pubDate>Thu, 19 Jul 2012 08:14:45 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=515#comment-166443</guid>
		<description><![CDATA[Why did you take http://www.tweetmotif.com/ down?

I wanted to try it.]]></description>
		<content:encoded><![CDATA[<p>Why did you take <a href="http://www.tweetmotif.com/" rel="nofollow">http://www.tweetmotif.com/</a> down?</p>
<p>I wanted to try it.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: suscunnisuase</title>
		<link>https://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/#comment-27745</link>
		<dc:creator>suscunnisuase</dc:creator>
		<pubDate>Sat, 05 Jun 2010 01:36:36 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=515#comment-27745</guid>
		<description><![CDATA[18 Jun 2007 Just about everyone who watches sports agrees that ESPN&#039;s Erin Andrews is smoking hot. If you don&#039;t know Erin, she is the sideline reporter &lt;a href=&quot;http://erinandrewspeephole.net/&quot; rel=&quot;nofollow&quot;&gt;Erin Andrews Peep Video Vid Caps&lt;/a&gt; P. Continue Reading &quot; Erin Andrews&#039; stalker gets 27-month prison term Former Florida Gators dazzler and current ESPN sideline reporter Erin Andrews has signed on to be a. All the latest on Erin Andrews, including pics and videos, is at Chickipedia. Discover vital facts about Erin Andrews : she was born on as Erin Andrews. Erin Andrews - Wikipedia, the free encyclopedia. Erin Andrews (born May 4, 1978) is an Erin Andrews Dazzlers. Jill Arrington Pictures. Pagination. 1. 2. 3. 4. 5. What many people may not know is that Erin was a &quot;Dazzler&quot; at her time in Florida? Dancing with the Stars 2010 Cast: Erin Andrews won&#039;t be a victim. Watch Erin Andrews Peephole Video, Erin Andrews is getting the Paris Hilton effect. Erin Andrews&#039; &quot;Dancing with the Stars&quot; efforts continued Monday night. JoeSportsFan.com MediaSpace Page For Erin Andrews.
  see more:
&lt;a href=&quot;http://briarcliffvillage.org/bb/index.php?action=profile;u=12271&quot; rel=&quot;nofollow&quot;&gt;Erin Andrews Peephole Images&lt;/a&gt;
 Erin Andrews Peephole VideoErin Andrews Peep Camera]]></description>
		<content:encoded><![CDATA[<p>18 Jun 2007 Just about everyone who watches sports agrees that ESPN&#8217;s Erin Andrews is smoking hot. If you don&#8217;t know Erin, she is the sideline reporter <a href="http://erinandrewspeephole.net/" rel="nofollow">Erin Andrews Peep Video Vid Caps</a> P. Continue Reading &#8221; Erin Andrews&#8217; stalker gets 27-month prison term Former Florida Gators dazzler and current ESPN sideline reporter Erin Andrews has signed on to be a. All the latest on Erin Andrews, including pics and videos, is at Chickipedia. Discover vital facts about Erin Andrews : she was born on as Erin Andrews. Erin Andrews &#8211; Wikipedia, the free encyclopedia. Erin Andrews (born May 4, 1978) is an Erin Andrews Dazzlers. Jill Arrington Pictures. Pagination. 1. 2. 3. 4. 5. What many people may not know is that Erin was a &#8220;Dazzler&#8221; at her time in Florida? Dancing with the Stars 2010 Cast: Erin Andrews won&#8217;t be a victim. Watch Erin Andrews Peephole Video, Erin Andrews is getting the Paris Hilton effect. Erin Andrews&#8217; &#8220;Dancing with the Stars&#8221; efforts continued Monday night. JoeSportsFan.com MediaSpace Page For Erin Andrews.<br />
  see more:<br />
<a href="http://briarcliffvillage.org/bb/index.php?action=profile;u=12271" rel="nofollow">Erin Andrews Peephole Images</a><br />
 Erin Andrews Peephole VideoErin Andrews Peep Camera</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: brendano</title>
		<link>https://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/#comment-6367</link>
		<dc:creator>brendano</dc:creator>
		<pubDate>Fri, 22 May 2009 18:15:31 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=515#comment-6367</guid>
		<description><![CDATA[Oh yeah and on hashtags.  I personally think their semantics are kind of the same thing as a set-of-words view of tweets.  (If you had good NLP, you shouldn&#039;t need tags for textual data, right? -- especially when tags contribute to the 140 limit just as much as normal text!)  Everyone else seems to think they&#039;re important though and that they should be linkified in the UI for a new search and such.

Hashtags do have one very important aspect: people use them when they intentionally want their message to be grouped with other messages using that same hashtag.  That alone makes them worthy of more in-depth special treatment.]]></description>
		<content:encoded><![CDATA[<p>Oh yeah and on hashtags.  I personally think their semantics are kind of the same thing as a set-of-words view of tweets.  (If you had good NLP, you shouldn&#8217;t need tags for textual data, right? &#8212; especially when tags contribute to the 140 limit just as much as normal text!)  Everyone else seems to think they&#8217;re important though and that they should be linkified in the UI for a new search and such.</p>
<p>Hashtags do have one very important aspect: people use them when they intentionally want their message to be grouped with other messages using that same hashtag.  That alone makes them worthy of more in-depth special treatment.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: brendano</title>
		<link>https://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/#comment-6365</link>
		<dc:creator>brendano</dc:creator>
		<pubDate>Fri, 22 May 2009 18:03:39 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=515#comment-6365</guid>
		<description><![CDATA[Tthanks for the comments.  Significant phrases are extracted against a general twitter corpus by comparing likelihoods under the two corpus&#039; language models.

There is something of a runtime partitioning scheme as you mentioned.  Tweetmotif has a very strong coherence constraint: a theme label MUST appear in all of the messages within that theme.  (I think for analytic purposes, it&#039;s nice to minimize hidden variables and be transparent.)  We start by just taking all the significant phrases as clusters, then treat it as a deduplication problem (kinda like agglomerative clustering) to reduce redundancy.  Cluster merges are only allowed to keep messages in the *intersection* -- not the union as in agglomerative clustering -- so they&#039;re only done when only a small number of messages would be lost in a merge (e.g. the cluster-cluster jaccard similarity is high).  Then there&#039;s a decision for what label to use for the new cluster; sometimes we end up with a skip n-gram as the label, which is really neat.

This is a pretty conservative as far as redundancy-reducing goes and there&#039;s still a lot left, as you point out.  But it&#039;s *much* better than just throwing up the top-k phrases :)

The correct way to do syntactic filtering for significant phrases is an interesting problem.  Restricting to NPs and adjectives is definitely a possibility.  I thought certain PP&#039;s and verbs can be pretty interesting, but they&#039;re sometimes mundane.  Sometimes things that look like punctuation are informative ... for example when a &quot;+&quot; comes up, it&#039;s usually because it&#039;s being used in a &quot;X + Y = Z&quot; construction, e.g. &quot;sunshine + park = happiness&quot;.  The unicode music symbol ♫ is another common but interesting token.

We thought the first version of this should err on the side of keeping as much as possible.  (wouldn&#039;t it be nice to have a magical completely unsupervised analysis system with no biases implied by the implementor? :) )  Well, there is already one important filtering step.  We take the raw n-grams then throw out ones that look like they cross syntactic boundaries -- e.g. ones that contain a right-binding function word like &quot;the&quot; at the rightmost position.  Results are much worse without this step...

The system&#039;s overall result quality is much better on non-news topics like &quot;sandwich&quot;.  I&#039;m not sure why -- it might be because, on news topics, a small set of things are said in many different ways, and tweetmotif fails to abstract over them and ends up with too much redundancy.

Twitter conventions are really interesting.  Hashtags are treated specially: they are only allowed to be unigrams and never allowed to join anything else as an n-gram, since usually they have unordered tag-like set semantics.  (There are some instances where that doesn&#039;t hold .. &quot;i went to the store and bought a #macintosh and bla...&quot; but they seem rare.)  Also @-replies are treated specially too.  Everything still boils boils down to textual n-grams which is better than nothing, but not quite right, of course.  There&#039;s very obvious structured information here.]]></description>
		<content:encoded><![CDATA[<p>Tthanks for the comments.  Significant phrases are extracted against a general twitter corpus by comparing likelihoods under the two corpus&#8217; language models.</p>
<p>There is something of a runtime partitioning scheme as you mentioned.  Tweetmotif has a very strong coherence constraint: a theme label MUST appear in all of the messages within that theme.  (I think for analytic purposes, it&#8217;s nice to minimize hidden variables and be transparent.)  We start by just taking all the significant phrases as clusters, then treat it as a deduplication problem (kinda like agglomerative clustering) to reduce redundancy.  Cluster merges are only allowed to keep messages in the *intersection* &#8212; not the union as in agglomerative clustering &#8212; so they&#8217;re only done when only a small number of messages would be lost in a merge (e.g. the cluster-cluster jaccard similarity is high).  Then there&#8217;s a decision for what label to use for the new cluster; sometimes we end up with a skip n-gram as the label, which is really neat.</p>
<p>This is a pretty conservative as far as redundancy-reducing goes and there&#8217;s still a lot left, as you point out.  But it&#8217;s *much* better than just throwing up the top-k phrases :)</p>
<p>The correct way to do syntactic filtering for significant phrases is an interesting problem.  Restricting to NPs and adjectives is definitely a possibility.  I thought certain PP&#8217;s and verbs can be pretty interesting, but they&#8217;re sometimes mundane.  Sometimes things that look like punctuation are informative &#8230; for example when a &#8220;+&#8221; comes up, it&#8217;s usually because it&#8217;s being used in a &#8220;X + Y = Z&#8221; construction, e.g. &#8220;sunshine + park = happiness&#8221;.  The unicode music symbol ♫ is another common but interesting token.</p>
<p>We thought the first version of this should err on the side of keeping as much as possible.  (wouldn&#8217;t it be nice to have a magical completely unsupervised analysis system with no biases implied by the implementor? :) )  Well, there is already one important filtering step.  We take the raw n-grams then throw out ones that look like they cross syntactic boundaries &#8212; e.g. ones that contain a right-binding function word like &#8220;the&#8221; at the rightmost position.  Results are much worse without this step&#8230;</p>
<p>The system&#8217;s overall result quality is much better on non-news topics like &#8220;sandwich&#8221;.  I&#8217;m not sure why &#8212; it might be because, on news topics, a small set of things are said in many different ways, and tweetmotif fails to abstract over them and ends up with too much redundancy.</p>
<p>Twitter conventions are really interesting.  Hashtags are treated specially: they are only allowed to be unigrams and never allowed to join anything else as an n-gram, since usually they have unordered tag-like set semantics.  (There are some instances where that doesn&#8217;t hold .. &#8220;i went to the store and bought a #macintosh and bla&#8230;&#8221; but they seem rare.)  Also @-replies are treated specially too.  Everything still boils boils down to textual n-grams which is better than nothing, but not quite right, of course.  There&#8217;s very obvious structured information here.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Bob Carpenter</title>
		<link>https://brenocon.com/blog/2009/05/announcing-tweetmotif-for-summarizing-twitter-topics-with-a-dash-of-nlp/#comment-6339</link>
		<dc:creator>Bob Carpenter</dc:creator>
		<pubDate>Thu, 21 May 2009 19:07:05 +0000</pubDate>
		<guid isPermaLink="false">http://anyall.org/blog/?p=515#comment-6339</guid>
		<description><![CDATA[There are lots of ways to build this app.  In fact, we were just talking to a customer about building something query refinement by phrase suggestion, which is very similar technically.  My fave implementation of this idea is scirus.com (by FAST and Elsevier) which gives you suggestions in &quot;refine your search&quot;.

First, do significant phrase extraction.  You can either do this per query result set (very expensive, but much better results if you have enough results) or statically (perhaps with updates).  Best is to do it versus some background corpus (e.g. Wikipedia), but plain old collocation extraction is OK.  Then index the phrases along with the tweets.

At run time, when you search, you pull back the sig phrases in the tweets as well as the results, so you get a set of tweets and a tweet-phrase relation.  You could either do some kind of fancy dynamic-programming optimization to get an optimal partition by phrase according to some metric (diversity/coherence -- see below), or you could just report the top few results and accept some redundancy (realizing that a phrase appearing in all or most hits isn&#039;t useful in organizing results).   

The real trick for this application is optimizing diversity (low similarity among the clusters) and high coherence (high similarity of tweets within a cluster).  Right now, tweemotif&#039;s not doing a very good job of this (clusty.com also has issues -- it&#039;s a tough problem).  

Another thing you might consider for tweetmotif is restricting results to phrases like noun phrases; getting &quot;by the black&quot; is wrong phrasally for a search on the band &quot;Black Keys&quot;, and plain old &quot;-&quot; is just wrong, which I got as a term for search &#039;Raveonettes&#039;.  And &quot;flu&quot; pulled back &quot;h1n1 flu&quot;, &quot;swine flu&quot;, and &quot;case of h1n1&quot;, which just isn&#039;t diverse enough.  I had better luck on the drilldowns for topics like &quot;banh mi&quot;.  You should probably also pay attention to Twitter conventions like &quot;#&quot; for tags.]]></description>
		<content:encoded><![CDATA[<p>There are lots of ways to build this app.  In fact, we were just talking to a customer about building something query refinement by phrase suggestion, which is very similar technically.  My fave implementation of this idea is scirus.com (by FAST and Elsevier) which gives you suggestions in &#8220;refine your search&#8221;.</p>
<p>First, do significant phrase extraction.  You can either do this per query result set (very expensive, but much better results if you have enough results) or statically (perhaps with updates).  Best is to do it versus some background corpus (e.g. Wikipedia), but plain old collocation extraction is OK.  Then index the phrases along with the tweets.</p>
<p>At run time, when you search, you pull back the sig phrases in the tweets as well as the results, so you get a set of tweets and a tweet-phrase relation.  You could either do some kind of fancy dynamic-programming optimization to get an optimal partition by phrase according to some metric (diversity/coherence &#8212; see below), or you could just report the top few results and accept some redundancy (realizing that a phrase appearing in all or most hits isn&#8217;t useful in organizing results).   </p>
<p>The real trick for this application is optimizing diversity (low similarity among the clusters) and high coherence (high similarity of tweets within a cluster).  Right now, tweemotif&#8217;s not doing a very good job of this (clusty.com also has issues &#8212; it&#8217;s a tough problem).  </p>
<p>Another thing you might consider for tweetmotif is restricting results to phrases like noun phrases; getting &#8220;by the black&#8221; is wrong phrasally for a search on the band &#8220;Black Keys&#8221;, and plain old &#8220;-&#8221; is just wrong, which I got as a term for search &#8216;Raveonettes&#8217;.  And &#8220;flu&#8221; pulled back &#8220;h1n1 flu&#8221;, &#8220;swine flu&#8221;, and &#8220;case of h1n1&#8243;, which just isn&#8217;t diverse enough.  I had better luck on the drilldowns for topics like &#8220;banh mi&#8221;.  You should probably also pay attention to Twitter conventions like &#8220;#&#8221; for tags.</p>
]]></content:encoded>
	</item>
</channel>
</rss>

<!-- Dynamic page generated in 0.035 seconds. -->
<!-- Cached page generated by WP-Super-Cache on 2026-04-13 01:30:23 -->
