Archive for the ‘Uncategorized’ Category

Stuart Metcalfe

Paypal support coming to Ubuntu Software Centre

It’s been a long time in development, but we’re finally drawing close to releasing Paypal support in Ubuntu Pay, the payment service behind Ubuntu Software Centre. Here are a couple of screenshots to whet your appetite…

We’re aiming to launch this new feature before Christmas.

Stuart Metcalfe

Ubuntu SSO improved experience logging in to sites with OpenID

Anyone who uses the OpenID service on Ubuntu SSO (or Launchpad) to log in to sites on the Internet will know that we currently only send your username unless the site is registered with us. This makes logging in with Ubuntu SSO using OpenID less-than-great if the site needs your name or email address, for instance, to create an account for you. Next week we’re rolling out an update which removes this restriction and gives you control over which personal data you want to share.

I’ve personally wanted this feature for a while so it’s great to see it finally land.

Stuart Metcalfe

Ubuntu developer portal preview

In Ubuntu 10.10 we introduced the ability to purchase software through the Ubuntu Software Centre.  Because of the existing manual process to get new ISVs (independent software vendors) signed up, however, the number of applications available for purchase has grown slower than we’d like.  The solution we’ve selected to this problem is to have a standard developer program for ISVs to sign up to and to provide an online self-service portal where they can manage their applications.  As we’re about to launch the first preview of this portal, we wanted to show the experience a first-time user will have from agreeing to the developer terms to actually having their new software appear in the Software Centre.

For this initial release, access is limited to ISVs we have an existing relationship with.  Once we’ve ironed out the early wrinkles, we’ll be opening the doors to everyone.  If you would like to hear when the service launches publicly, keep an eye on the Canonical blog and

Anthony Lenton

Debug Django with the Werkzeug debugger

The great Django command-extensions app gives you an easy way to integrate the Werkzeug debugger into your Django site.

For if you’ve never used this debugger before, imagine the regular Django technical details debugging page, with all same super useful information — environment variables, sql queries, and a full traceback with even the local variables and a bit of code context for each frame.
Now imagine you could have a pdb prompt running within each frame of the traceback, actually running interactively via ajax within the frames that you see.  That’s the kind of raw power Werkzeug gives you.

A big disclaimer though: this is not for running on production sites.  You’ll be giving anybody that happens to stumble upon a failure page an interactive (unrestricted) python terminal running with the web server’s permissions on your box.

So, this is just what out the box django-command-extension’s runserver_plus gives you.  With a bit of tweaking, here are a couple of other neat things you can do:

Make your server multithreaded

There are times when you need a multithreaded server in your development environment.  I’ve seen arguments about this being a smell for bad design, so here ar two reasonable scenarios I’ve come across:

  • Developing a web API, you might have a view that provides a test consumer for the API, not intended to be available on production.  The test consumer view will make a (blocking) call out to the API that’s provided by the same server, hanging everything if you don’t have multithreadedness.
  • Or, in larger systems where you have more than one application server, you might have a view or two that do a bit of communication between servers (to see if everything’s ok, to provide a neat server status screen), that you’ll need to test calling out to the same server during development.

For whatever reason it may be, if you need to enable threading, you can just set threaded=True when you call Werkzeug’s run_simple function:

Debug your whole wsgi app instead of just Django

If you’re running Django as part of a larger wsgi stack (a non-Django web api, or some authentication wsgi middleware), you probably would love to be able to run the whole stack in your development server, and even have those same great debugging features available for everything, not only Django.
You can do this by modifying the wsgi app that Werkzeug debugs:

It’s something with raw power that keeps getting better the more you look…

Ricardo Kirkner

configglue: a configuration library on steroids

What is configglue?

configglue is a library that glues together Python’s optparse.OptionParser and
ConfigParser.ConfigParser, so that you don’t have to repeat yourself when you
want to export the same options to a configuration file and a commandline

The main features of configglue are:

  • ini-style configuration files
  • schema-based configuration
  • commandline integration
  • configuration validation

Why would I want to use configglue?

Some of the benefits of using configglue are that it allows you to:

  • separate configuration declaration (which options are available) from
    definition (what value does each option take)
  • validate configuration files (there are no required options missing, prevent
    typos in option names, assert each option value is of the correct type)
  • use standards-compatible configuration files (standard ini-files)
  • use standard types out of the box (integer, string, bool, tuple, list, dict)
  • create your own custom types beyond what’s provided in the library
  • easily support commandline integration
  • override options locally by using several configuration files (useful for
    separating configuration files for different environments)

configglue and django-configglue are already available in Ubuntu 10.10 (Maverick), so they can be installed via apt-get. configglue should already be installed if you have the desktop edition, as it’s being used by Ubuntu One’s client.

Who else is using configglue?

  • Ubuntu Pay
  • Ubuntu Software Center
  • Ubuntu Single Sign On
  • Ubuntu One

Got curious?

You can find a quickstart guide for configglue on and you can get its code at

As an additional bonus, there is another project called django-configglue
which allows you to use all the benefits of configglue on your Django
projects. You can find a quickstart guide for django-configglue on and you can get its code at

Michael Foord

Continuous Integration with Django and Hudson

After the release of Ubuntu Maverick Meerkat we have a brief time of housekeeping in the ISD team, where we work on clearing our technical debt and implementing improvements to our development processes. One of these improvements has been getting continuous integration setup for some of our projects. Continuous integration means not just having an automated test suite, but having an automated system for running the tests (continuously)

We have settled on Hudson as our CI server. Despite being a Java based tool it is a popular choice in the Python world, mainly because Hudson is both easy to install / configure and provides an attractive and powerful web based interface out-of-the-box.

You can use Hudson to report test run pass or fail, and view the console output from a test run, with virtually no work at all. Simply provide a shell command for running your tests and off you go. Hudson works best when your test run generates an XML description of the test run in a poorly specified (but widely implemented) format called JUnit-XML. This is the same JUnit that was the original inspiration for the Python unittest testing library.

With JUnit XML describing your test run you can use the Hudson UI to view individual test failures and generate pretty graphs for how the time taken by tests changes over successive builds.

Our projects are based on Django, which in turn uses the standard unittest test runner for running tests. Python has a wealth of different choices for coaxing JUnit-XML out of a unittest test run. As we’re deploying on Ubuntu Lucid servers we needed a solution easy to integrate with the Lucid distribution of Django (approximately version 1.1.2).

After trying several alternatives we settled on the pyjunitxml package, which just happens to be the creation of Robert Collins, a fellow Canonical developer.

For a suite (unittest terminology for a collection) of tests, getting a junit-xml report from pyjunitxml is gloriously easy. Here’s the code:

    import junitxml
    with open('xmlresults.xml', 'w') as report:
        result = junitxml.JUnitXmlResult(report)

If you’re familiar with unittest code you may be surprised that there is no test runner involved. The unittest TextTestRunner class is useful for generating console output of a test run, but as this isn’t needed for a run under continuous integration we only need a test suite and the test result object from junitxml.

To integrate this into a standard django test run we had to copy the run_tests function from the django.test.simple module and add a parameter to use this code when running under Hudson.

Unsurprisingly our projects are managed on Launchpad and use Bazaar for version control. Although not enabled by default Hudson ships with a plugin for Bazaar integration. Here’s a guide for setting up Hudson for a Launchpad / Bazaar based project. It assumes you have a script “” for running your test suite with the junitxml code active:

First install hudson. For debian and Ubuntu this link gives the details:

(You will also need a version of Java and the jre installed which I don’t think that guide covers.)

Once installed Hudson will start on boot. Immediately after install you may need to start it manually:

sudo /etc/init.d/hudson restart

You can view the locally running Hudson from this URL:


You need to enable the bzr plugin in Hudson. You can do this through the Hudson plugin manager:


Find it in the ‘available’ tab of the plugin manager. You will need to restart Hudson after installing the plugin.

Next you need to give the Hudson user access to your repository. You can this step if your repository is public and can be fetched by an anonymous user.

  • Switch to the Hudson user: sudo su – hudson
  • Tell bzr your launchpad user: bzr lp-login your-username
  • Generate ssh keys without a passphrase: ssh-keygen -t rsa
  • Add the public key to your user on launchpad.

Next create a new Hudson job, with whatever name you want.

  • Select Build a free-style software project
  • Source code management: Bazaar
  • Repository URL: lp:~your-username/project-name/branch
  • Add build step -> Execute Shell:
  • Post-build action: Check Publish JUnit test result report
  • Test report XMLs: xmlresults.xml

Finally: Build Now

You can watch the build in progress via the console output.

There is no console output during the test run itself, the results go to the junit xml file used by Hudson to report the results. The console output is still useful as any build errors or unhandled exceptions will show up there.

At the moment Hudson is happily running scheduled builds on my local machine. The next step is to run it on our private cloud and decide whether or not to use a commit hook to trigger the builds. I have a personal preference for scheduled builds (as close to back-to-back as possible). Multiple runs per revision gives you the opportunity to shake out fragilities in your tests and collect better data (more data == better data) for performance tests.

Ricardo Kirkner

Ubuntu Pay is open for translations

We are pleased to announce that Ubuntu Pay, the new payment service that allows you to buy commercial software (by means of the Ubuntu Software Centre) or subscriptions to services like Ubuntu One (in the near future), is ready to start accepting translations from the community.

Currently, we support a restricted set of languages, namely

– Chinese
– English
– German
– Polish
– Spanish

However, we would like to be able to provide a native language interface for everyone! So, if you’d like to see this service translated to your language, you can help us by

a) fixing translation mistakes on currently supported languages
b) adding translations for your language, even if it’s not supported right now

We’ll try to adopt new languages as they grow a community behind them that can help keep the translations up to date.

To start contributing your translations, just go to the translations overview page.

Łukasz Czyżykowski

tox – New Tool For Testing Python Libraries

What is tox?

tox is a tool for testing python libraries (or more generally Python projects) against multiple Python implementations. Before tox, this would have been quite involved and any project wanting to check had to create its own solution. With tox it’s just a matter of creating a simple .ini file and running it as often as you need.

It uses virtualenv to properly manage and separate environments.

It’s compatible with all Python versions starting from 2.4, currently up to n 3.2. Support for Jython and PyPy is also included.

How to use it?

First thing is to create tox.ini file in your project’s directory, next to the Format is quite simple. Here’s example:

envlist = py26,py27

deps = nose
commands = nosetests

Where envlist is a list of all Python implementations you want the tests (py26, py27 are built-in, one can also create custom environments) to run in. deps is a list of all dependencies your package requires, other than packages specified in file. Finally commands is a list of all commands which tox will run in that test environment. Here it’s only running nose, but you can also have commands for building documentation or anything else that you want to ensure that it works.

With this file in package directory running tox is just a matter of:

$ tox

It will create each environment, install required packages (using pip), run all the commands and finally report success or failure. It saves all output to log files and caches all downloaded files, so no need worry about that.

Real life example

To test tox on a real project I’ve chosen configglue, our configuration library. It is always a good idea to make sure it runs properly on various Python versions. So, there will be no surprises in the future.

First step was to download the source code ($ bzr branch lp:configglue). Next in the project’s directory. I’ve created very simple tox.ini file displayed below:

envlist = py25,py26,jython

deps = nose
commands = nosetests []


As you can see, one addition is testing against Jython.

Another small caveat with configglue is that one file in the source code requires small addition to be compatible with Python 2.5 (from __future__ import with_statement). That doesn’t ensure that all tests will pass, it just prevents syntax error from being raised.

As you can see you can customize each test environment separately by using testenv:pyver section name.

Running tox with configglue reveals that it works perfectly on Python 2.5 and 2.6 but one test fails on Jython (offending line _open = __builtins__['open']).

Final Notes

Additionally tox has support for working with Hudson and should be fairly straightforward to integrate with other CI solutions.

Because this project is really young (released on June 2010) there’s not so many real life examples. One of them is from mock library; it’s tox.ini is fairly complex, as it includes building documentation using Sphinx.

You can find an even more complicated example in the kombu source code.

tox.ini in the tox source code has an example of integration with Hudson.

Of course, tox site contains a list of examples.

Szilveszter Farkas

Outlining ISD’s deployment process

In this blog entry I would like to describe our deployment strategies we use at the different stages of our development process. The stages are the following:

  1. local development
  2. QA
  3. staging (+ QA)
  4. production (+ QA)

For local development everyone is welcome to use his preferred way, but most of us bet on Virtualenv. Especially given the fact that we maintain a large number of different projects, it makes our lives a lot easier to have a separate environment for each of those. We also have to make sure that we align with the production environment, which is in some cases still Python 2.5-based, but we’re currently in transition to 2.6.

If a feature or bugfix is ready to be QA’d, we deploy the application to an Amazon EC2 instance. Our team mate, Łukasz Czyżykowski, wrote a collection of extensions to Fabric, that provides a few useful functions (e.g. using private PPAs very easily). With a few dozens of lines of simple Python code, we can deploy the whole application to a running EC2 instance. We also use EC2 to QA all the features and bugfixes targeted at a release together before deploying to staging, so that if there is an issue, we can re-deploy very quickly (during the next two stages, QA is mainly about testing regressions).

The staging and production environments are identical from the deployment process perspective. We simply create a binary Debian package from our application: Launchpad’s PPA feature makes the build process a breeze. The main reason we decided to go with Debian packages is that we can also specify system level dependencies, not only Python packages (and of course there’s some dogfooding involved since the company supports Ubuntu). This also requires that all of the team members have packaging skills, so we had several training sessions, and a two-day online sprint where we packaged lazr.restful and all of its dependencies which were not available in Ubuntu 8.04 (around 30 packages, half of them backports, half of them new packages – thanks to our hard-working team members, these are available for Ubuntu 10.04 as well).

For configuration we don’t use Django’s built-in settings mechanism, but a custom solution that will be open sourced in the near future (one more reason to keep an eye on our blog). It consists of two components: schemaconfig is responsible for parsing the config files (which are INI-style, but have some extra features, like layering, typing, and support for data structures like lists and dictionaries – basically we looked around for solutions, and stole a little bit from everywhere to put together one that fits us most), and there’s django-settings which is a glue between schemaconfig and Django’s settings (so in the end we still use django.conf.settings). One of the biggest problems we had with our previous setup that it was very prone to human error, and that caused us unexpected deployment issues between staging and production. This is solved by the layering and the non-Python style of the config files, so they are easily manageable by both us and IS (our operations team).

Watch this blog for more about schemaconfig and other exciting projects and articles!