Donating our OLPC XO

Way back at linux.conf.au 2008 there was a large OLPC XO giveaway, but with the rider do something wonderful with this, or give it to someone who will. Neither Andrew nor I received one directly, but Matthew Garrett gave his to Andrew essentially on the grounds that he wasn’t going to do anything wonderful with it. (If I have the chronology right, Matthew had a stack of laptops in his possession at the time and did things to them regularly, generally making them sleep on demand.)

In any event, neither Andrew nor I did anything wonderful with the XO: Andrew intended to look at some point at Python or Python application startup times (the Bazaar team have a bunch of tricks in that regard), but two years is a lot of intending.

Still, better late than never. In the spirit of the original giveaway, we’ve handed it over to be taken to New Zealand by someone going to linux.conf.au 2010. It will be donated to the Wellington OLPC testers group, who meet weekly to work on various projects and who are somewhat short on machines.

If you are similarly (morally) bound by the linux.conf.au 2008 giveaway conditions, aren’t doing anything wonderful with your XO, and are going to linux.conf.au 2010 or can get your XO there, you could do likewise. You could drop off to Tabitha Roder at the education miniconf, the OLPC stand at Open Day or otherwise get in touch with her. (You probably want to let her know yours is coming anyway, so she has a sense of whether to expect one or two, or a truckload.)

Other possibilities include getting involved in the Sydney group or checking if they’d have a use for laptop donations. (They meet more regularly than that wiki page implies; they are now meeting at SLUG.) I don’t know what the status of the OLPC library is. The webpage being down is probably not a great sign, but perhaps collaborators would help John out there. You’d at least be doing something meta-wonderful.

Clean up IMAP folders

Per Matt Palmer’s blog entry OfflineIMAP and Deleting Folders users of any mail sorting recipe that creates new mail folders a lot tend to find that over time they accumulate a lot of mail folders for, eg, email lists they are no longer subscribed to. And most IMAP clients will waste time checking those folders for new mail all the time.

Matt wrote:

Now, of course, someone’s going to point me to a small script that finds all of your local empty folders and deletes them locally then issues an IMAP “delete folder” command on the server. But I had fun working all this out, so it’s not a complete waste.

I haven’t quite done this, I’ve just written a script that detects and deletes empty remote folders. (For me, offlineimap does not have the behaviour of creating new remote folders, so I haven’t bothered cleaning up local folders.)

It’s good: it’s speeding up my mail syncs a whole lot, deleting these old folders I haven’t received mail in for about five years. I’ve got full details and the script available for download (as you’d expect, it’s short): Python script to delete empty IMAP folders.

OLPC

This morning at Bruce Schneier’s keynote it was announced that they wanted to give a One Laptop Per Child XO laptop to the people at the conference who were going to do something incredibly cool with it. Except… they didn’t have a way of determining who those people were. So, they were given away to conference attendees whose names were chosen at random. The condition is that they recipients should either do something wonderful or pass it on to someone who will.

Did we get one? No. But Matthew Garrett gave us his. And by ‘us’ I mean ‘Andrew’. But still.

Ideas for wonderful things accepted.

My talk at OSDC: the Planet Feed Reader

I gave a thirty minute presentation at the Open Source Developers’ Conference yesterday about the Planet software and the associated communities and conventions, focusing more on the latter since one of my reviewers suggested that the social aspects are more interesting than the code. My slides [PDF format, 2.1MB] are now available for the amusement of the wider public.

Much of the discussion of history was a recap of my Planet Free Software essay and the discussion of Planet conventions was a loose recap of accumulated wisdom, including:

  1. using bloggers’ real names, or at least the ones which they attach to email (usually real names) in addition to common IRC/IM handles is useful for putting faces to blog entries to contributions;
  2. once the convention of using real faces and real names is established, people get upset when the conventions are broken (quoth Luis Villa: I’m not sure who/what this ubuntu-demon is, but ‘random animated head without a name meandering about doing a lot of engineering work to fix a problem that should not exist’ was not what I was looking for when I was looking for information on planet ubuntu); and
  3. life blogging is of interest to an extent, many developers would actually like to feel that they’re friends with each other, but the John Fleck case on Planet GNOME shows that there are limits.

Much of the rest was due to Luis Villa’s essay on blogging in the corporate open source context, but as I wasn’t allowed to set assigned reading to the audience I was able to pad out by half an hour by including that content.

Mostly it was a fun experiment in doing slides in a format other than six bullet points per slide, six slides per section, six sections per talk format; incorporating badly rescaled images in various places; and using Beamer so I was surprised to end up hosting a Planet BoF (Birds of a Feather) session, discussing it from the point of view of someone running a Planet (the editor). Some of the topics that came up were:

  • trying to start communities via Planet sites, rather than enhancing them, by, say, starting a environmental politics Planet;
  • the possibility of introducing some of the newer blog ideas to the Free Software world (like carnivals);
  • allowing a community to edit a Planet, and editorial policies in general;
  • potential problems with aggregating libellous or illegal content (another reason some editors apparently insist on real names);
  • alternative aggregators;
  • banning RSS in favour of Atom;
  • whether it is possible or wise to filter people’s feeds without their consent;
  • moving to the Venus branch of Planet; and
  • making Venus trunk.

I may propose a blogging BoF at linux.conf.au and, if I do so, I’ll even plan some discussion points, which will make it less random.

Logging into the OSDC wireless network

I have a wireless login script for attendees of OSDC who use Ubuntu, Debian, or anything else that can run scripts on connecting to a network and has essentially the same iwconfig output:

 eth1      IEEE 802.11g  ESSID:"Monash-Conference-Net" Mode:Managed  Frequency:2.437 GHz  Access Point: 00:13:7F:9D:36:C0 

To save some tiny amount of time when connecting to the wireless, stick my osdc-login script in your /etc/network/if-up.d directory or equivalent and give it similar permissions to what’s already in there. You can get the latest version of the script at https://gitlab.com/puzzlement/osdc-2006-monash-wireless-login/raw/master/osdc-login, or through Bazaar git, with the repository at https://gitlab.com/puzzlement/osdc-2006-monash-wireless-login/tree/master. It’s very small, but feel free to send me improvements (although if using Bazaar git, please don’t check in a version containing the real username and password).

You need to replace INSERTCONFERENCELOGINHERE with the appropriate username and INSERTCONFERENCEPWHERE with the password. By running the script you will be agreeing to Monash’s terms of service, which are here.

Syndication, aggregation, and HTTP caching headers

Syndication, aggregation, and HTTP caching headers

I’ve seen various people in various places lately who were very unhappy about someone requesting their RSS feed every 30 seconds, or minute, or half hour, or whatever, and re-downloading it every time at a cost of megabytes in bandwidth. I’ve also seen people growing unhappy with the Googlebot for re-downloading their entire site every day.

So, a quick heads-up: there is a way for a client to say “hey, I have an old copy of your page, do you have anything newer, or can I use this one?” and for the server to say “hey, I haven’t changed since the last time you viewed me! use the copy you downloaded then!” Total bandwidth cost: about 300 bytes per request. That’s still a bit nasty for an ‘every 30 seconds’ request, but it means you won’t get cranky at the 10 minute people anymore. Introducing Caching in HTTP (1.1)!

The good news! Google’s client already does the client half of this. Many of the major RSS aggregaters do the client half of this (but alas, not all, there’s a version of Feed on Feeds that re-downloads my complete feed every half hour or so). And major servers already implement this… for static pages (files on disk).

The bad news! Since dynamic pages are generated on the fly, there’s no way for the server software to tell if they’ve changed. Only the generating scripts (the PHP or Perl or ASP or whatever) have the right knowledge. Dynamic pages need to implement the appropriate headers themselves. And because this is HTTP-level (the level of client and server talking their handshake protocol to each other prior to page transmission) not HTML level (the marked-up content of the page itself), I can’t show you any magical HTML tags to put in your template. The magic has to be added to the scripts by programmers.

End users of blogging tools, here’s the lesson to take away: find out if your blogging software does this. If you have logs that show the return value (200 and 404 are the big ones), check for occurrences of 304 (this code means “not modified”) in your logs. If it’s there, your script is setting the right headers and negotiating with clients correctly. Whenever you see a 304, that was a page transmission saved. If you see 200, 200, 200, 200 … for requests from the same client on a page you know you weren’t changing (counting all template changes), then you don’t have this. Nag your software developers to add it. (If you see it only for particular clients, then unfortunately it’s probably the client’s fault. The Googlebot is a good test, since it has the client side right.) An appropriate bug title would be I don’t think your software sets the HTTP cache validator headers, and explain that the Googlebot keeps hitting unchanged pages and is getting 200 in response each time.

RSS aggregater implementers and double for robot implementers: if you’ve never heard of the If-None-Match and If-Modified-Since headers, then you’re probably slogging any page you repeatedly request. Your users on slow or expensive connections hate you, or would if they knew the nature of your evil. Publishers of popular feeds hate you. Have a read of the appropriate bits of the spec and start actually storing pages you download and not re-downloading them! Triple for images!

Weblog and CMS software implementers: if you’ve never heard of the Last-Modified and/or ETag headers, learn about them, and add the ability to generate them to your software.

Creative Commons License
Syndication, aggregation, and HTTP caching headers by Mary Gardiner is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.