Speaker: Steven M. “I’m just sayin'” Cohen
[More links to cool stuff that I did not included can be found at the presentation wiki linked above.]
Google Reader is now more popular than Bloglines, which Cohen thinks has to do with the amount of money that Google can sink into it. Both have tools that tell you how many people are subscribed/reading it, which can be helpful in convincing administrators to support the use of RSS feeds from various sources. Offline feed readers don’t make much sense, since so often the things you are reading will direct you to other sources online.
If you’re not using Google Reader, do it now.
No, really. Steven says to do it.
Google + Feedburner = advertisements on your feeds, which means that they are now revenue generating, like the ads on your website. RSS is no longer sucking away your revenue source, so get over it and add feeds for your content! Plus, anyone using Page2RSS can scrape your content and turn it into a feed, so really, you should give them something that benefits you, too.
LibWorm is a site that indexes library-related blogs and news sources, and it provides RSS feeds, so use it for keeping current if you’re not already doing so.
Follow what is been twittered on your topic of choice using TweetScan. Follow all of your friends’ online activities at FriendFeed (notification once a day, which seems possibly even reasonably infrequent enough that I might actually use it).
Go check out his top
ten eleven twelve favorite tools. They’re all really cool and worth playing with.
I have been using Feed on Feeds as my RSS agregator for the past month, but I have decided to go back to using Bloglines. I liked the clean lines of Feed on Feeds, as well as the ability to host my feeds on my own website. However, it uses Magpie RSS to parse the … Continue reading “rss agregator”
I have been using Feed on Feeds as my RSS agregator for the past month, but I have decided to go back to using Bloglines. I liked the clean lines of Feed on Feeds, as well as the ability to host my feeds on my own website. However, it uses Magpie RSS to parse the feeds, and it can be quite persnickety if the feed does not completely validate. This limited me in the feeds I could track, as well as causing headaches every time I tried to update the feeds. Also, I couldn’t get the silent update feature to work. I tweaked my crontab file until I was blue in the face, but nothing worked. Overall, Bloglines requires less maintenance or headaches on my part. Feed on Feeds has great potential, but for now, I will give it some time to mature.
Will RSS feeds overload the ‘net?
Wired News has a short article about RSS feed readers and the potential they have for increasing web traffic. I knew about this article because it was listed in the RSS feed that I get from Wired. Go figure. Anyway, the author and others are concerned that because aggregators are becoming more and more popular among those who like to read regularly published electronic content, eventually a large chunk of web traffic will consists of desktop aggregators regularly downloading that data throughout the day.
The trouble is, aggregators are greedy. They constantly check websites that use RSS, always searching for new content. Whereas a human reader may scan headlines on The New York Times website once a day, aggregators check the site hourly or even more frequently.
If all RSS fans used a central server to gather their feeds (such as Bloglines or Shrook), then there wouldn’t be as much traffic, because these services check feeds once per hour at most, regardless of the number of subscribers. So, if you have 100 people subscribed to your feed, rather than getting 100 hits every hour (or some other frequency), you would only get one. The article notes two difficulties with this scenario. First, a lot of RSS fans prefer their desktop aggregators to a web-based aggregator such as Bloglines. Second, the Shrook aggregator is not free, and probably that will be the model that its competitors will take.
I don’t completely agree with the premise that having a central server distributing content to feed subscribers will reduce the flow of traffic on the ‘net anymore than it currently is. Whether my aggregator checks my feeds once an hour or whether Bloglines does it for me, I still use up bandwidth when I log in and read the content on the Bloglines site. For some feeds, if I want to read the whole entry or article, I still have to click to the site. Frankly, I think the problem has more to do with aggregators that “are not complying with specifications that reduce how often large files are requested.”
Readers are supposed to check if the RSS file has been updated since the last visit. If there has been no update, the website returns a very small “no” message to the reader.
But Murphy says the programs often don’t remember when they last checked, or use the local computer’s clock instead of the website’s clock, causing the reader to download entries over and over.
Perhaps the best thing for us to do is to educate ourselves about which RSS aggregator we use and how it may affect the bandwidth of the feeds we download through it.