01 Oct 2014

feedPlanet Ubuntu

Mark Shuttleworth: Exchange controls in SA provide no economic guarantees of stability, but drive up the cost of cross-border relationships for everyone

The South African Supreme Court of Appeal today found in my favour in a case about exchange controls. I will put the returned funds of R250m plus interest into a trust, to underwrite constitutional court cases on behalf of those who's circumstances deny them the ability to be heard where the counterparty is the State. Here is a statement in full:

Exchange controls may appear to be targeted at a very small number of South Africans but their consequences are significant for all of us: especially those who are building relationships across Southern Africa such as migrant workers and small businesses seeking to participate in the growth of our continent. It is more expensive to work across South African borders than almost anywhere else on Earth, purely because the framework of exchange controls creates a cartel of banks authorized to act as the agents of the Reserve Bank in currency matters.

We all pay a very high price for that cartel, and derive no real benefit in currency stability or security for that cost.

Banks profit from exchange controls, but our economy is stifled, and the most vulnerable suffer most of all. Everything you buy is more expensive, South Africans are less globally competitive, and cross-border labourers, already vulnerable, pay the highest price of all - a shame we should work to address. The IMF found that "A study in South Africa found that the comparative cost of an international transfer of 250 rand was the lowest when it went through a friend or a taxi driver and the highest when it went through a bank." The World Bank found that "remittance fees punish poor Africans". South Africa scores worst of all, and according to the Payments Association of South Africa and the Reserve Bank, this is "..mostly related to the regulations that South African financial institutions needed to comply with, such as the Financial Intelligence Centre Act (Fica) and exchange-control regulations."

Today's ruling by the Supreme Court of Appeal found administrative and procedural fault with the Reserve Bank's actions in regards to me, and returned the fees levied, for which I am grateful. This case, however, was not filed solely in pursuit of relief for me personally. We are now considering the continuation of the case in the Constitutional Court, to challenge exchange control on constitutional grounds and ensure that the benefits of today's ruling accrue to all South Africans.

This is a time in our history when it will be increasingly important to defend constitutional rights. Historically, these are largely questions related to the balance of power between the state and the individual. For all the eloquence of our Constitution, it will be of little benefit to us all if it cannot be made binding on our government. It is expensive to litigate at the constitutional level, which means that such cases are imbalanced - the State has the resources to make its argument, but the individual often does not.

For that reason, I will commit the funds returned to me to today by the SCA to a trust run by veteran and retired constitutional scholars, judges and lawyers, that will selectively fund cases on behalf of those unable to do so themselves, where the counterparty is the state. The mandate of this trust will extend beyond South African borders, to address constitutional rights for African citizens at large, on the grounds that our future in South Africa is in every way part of that great continent.

This case is largely thanks to the team of constitutional lawyers who framed their arguments long before meeting me; I have been happy to play the role of model plaintiff and to underwrite the work, but it is their determination to correct this glaring flaw in South African government policy which inspired me to support them.

For that reason I will ask them to lead the establishment of this new trust and would like to thank them for their commitment to the principles on which our democracy is founded.

This case also has a very strong personal element for me, because it is exchange controls which make it impossible for me to pursue the work I am most interested in from within South Africa and which thus forced me to emigrate years ago. I pursue this case in the hope that the next generation of South Africans who want to build small but global operations will be able to do so without leaving the country. In our modern, connected world, and our modern connected country, that is the right outcome for all South Africans.


01 Oct 2014 1:48pm GMT

Alan Pope: XDA Developer Conference 2014

The XDA Developer community had its second conference last weekend, this time in Manchester, UK. We were asked to sponsor the event and were happy to do so. I went along with Daniel Holbach from the Community Team and Ondrej Kubik from the Phone Delivery Team at Canonical.

This was my first non-Ubuntu conference for a while, so it was interesting for me to meet people from so many different projects. As well as us representing Ubuntu Phone, there were guys from the Jolla project showing off SailfishOS and their handset and ports. Asa Dotzler was also there to represent Mozilla & FirefoxOS.

Daniel did a small Ubuntu app development workshop which enabled us to learn a lot from our materials and process around App Dev Schools which we'll feed back to later sessions. Ondrej gave a talk to a packed room about hardware bring-up and porting Ubuntu to other devices. It was well receieved and explained the platform nicely. I talked about the history of Ubuntu phone and what the future might hold.

There were other sponsor booths including big names like nVidia showing off the Sheild tablet and Sony demonstrating their rather bizarre Smart EyeGlass technology. Oppo and OnePlus had plenty of devices to lust after too including giant phones with beautiful displays. I enjoyed a bunch of the talks including MediaTek making a big announcement, and demonstrating their new LinkIT One platform.

The ~200 attendees were mostly pretty geeky guys whose ages ranged from 15 to 50. There were Android developers, ROM maintainers, hardware hackers and tech enthusiasts who all seemed very friendly and open to discuss all kinds of tech subjects at every opportunity.

One thing I'd not seen at other conferences which was big at XDA:DevCon was the hardware give-aways. The organisers had obtained a lot of tech from the sponsors to give away. This ranged from phone covers through bluetooth speakers, mobile printers, hardware hacking kits through to phones, smart watches & tablets, including an Oppo Find 7, pebble watch and nVidia Sheild & controller. These were often handed out as a 'reward' for attendees asking good questions, or as (free) raffle prizes. It certainly kept everyone on their toes and happy! I was delighted to see an Ubuntu community member get the Oppo Find 7 :) I was rewarded with an Anker MP141 Portable Bluetooth Speaker during one talk for some reason :)

On the whole I found the conference to be an incredibly friendly, well organised event. There was plenty of food and drink at break times and coffee and snacks in between with relaxing beers in the evening. A great conference which I'd certainly go to again.


01 Oct 2014 10:09am GMT

Svetlana Belkin: Thoughts on Having a Meta Open Science Community

Over the last week, I started to think about how to improve the collaboration between the Open Science groups and researchers and also between the groups themselves. One of the ideas that I thought about using simple tools that are around in other Open * places (mainly Open Source/Linux distros). These tools are your forums (Discourse and other ones), Planet feeds, and wikis. Using these creates a meta community where members of the community can start there and get themselves involved in one or more groups. Open Science seems to lack this meta community.

Even though I think that meta community is not present, I do think that there is one group that can maintain this meta community and that group is the Open Knowledge Foundation Network (OKFN). They have a working group for Open Science. Therefore, I think, if they take the time and the resources, then it could happen or else some other group can be created for this.

What this meta community tool-wise needs:

Planet Feeds

Since I'm an official Ubuntu Member, I'm allowed to add my blog's feed to Planet Ubuntu. Planet Ubuntu allows anyone to read blog posts from many Ubuntu Members because it's one giant feed reader. This is well needed for Open Science, as Reddit doesn't work for academia. I asked on the Open Science OKFN mailing list and five people e-mailed me saying that they are interested in seeing one. My next goal is to ask the folks of Open Science OKFN for help on building a Planet for Open Science.


I can only think of one forum, which is the Mozilla Science Lab one, that I wrote about last a few hours ago. Having some general forum allows users to talk about various projects to job posting for their groups. I don't know if Discourse would be the right platform for the forums. To me, it's dynamicness is a bit too much at times.


I have no idea if a wiki would work for this meta Open Science community but at least having a guide that introduces newcomers to the groups is worthwhile to have. There is a plan for a guide.

I hope these ideas can be used by some group within the Open Science community and allow it the grow.

01 Oct 2014 1:30am GMT

30 Sep 2014

feedPlanet Ubuntu

Svetlana Belkin: Mozilla Science Lab Forums Now Open

I am pleased to announce that the Mozilla Science Lab now has a forum that anyone can use. Anyone can introduce themselves in this topic or the category.

30 Sep 2014 10:08pm GMT

Ubuntu Server blog: Server team meeting minutes: 2014-09-30



People present (lines said)


30 Sep 2014 7:24pm GMT

Adam Stokes: sosreport (SoS) version 3.2 released

The sos team is pleased to announce the release of sos-3.2. This release includes a large number of enhancements and fixes, including:


30 Sep 2014 5:55pm GMT

Ubuntu Kernel Team: Kernel Team Meeting Minutes – September 30, 2014

Meeting Minutes

IRC Log of the meeting.

Meeting minutes.


20140930 Meeting Agenda

Release Metrics and Incoming Bugs

Release metrics and incoming bug data can be reviewed at the following link:

Status: Utopic Development Kernel

The Utopic kernel remainds rebased on the v3.16.3 upstream stable
kernel. The latest uploaded to the archive is 3.16.0-19.26. Please
test and let us know your results.
Also, Utopic Kernel Freeze is next week on Thurs Oct 9. Any patches
submitted after kernel freeze are subject to our Ubuntu kernel SRU
Important upcoming dates:
Thurs Oct 9 - Utopic Kernel Freeze (~1 week away)
Thurs Oct 16 - Utopic Final Freeze (~2 weeks away)
Thurs Oct 23 - Utopic 14.10 Release (~3 weeks away)

Status: CVE's

The current CVE status can be reviewed at the following link:


Status: Stable, Security, and Bugfix Kernel Updates - Trusty/Precise/Lucid

Status for the main kernels, until today (Sept. 30):

Open Discussion or Questions? Raise your hand to be recognized

No open discussion.

30 Sep 2014 5:15pm GMT

Mark Shuttleworth: Fixing the internet for confidentiality and security

"The Internet sees censorship as damage and routes around it" was a very motivating tagline during my early forays into the internet. Having grown up in Apartheid-era South Africa, where government control suppressed the free flow of ideas and information, I was inspired by the idea of connecting with people all over the world to explore the cutting edge of science and technology. Today, people connect with peers and fellow explorers all over the world not just for science but also for arts, culture, friendship, relationships and more. The Internet is the glue that is turning us into a super-organism, for better or worse. And yes, there are dark sides to that easy exchange - internet comments alone will make you cry. But we should remember that the brain is smart even if individual brain cells are dumb, and negative, nasty elements on the Internet are just part of a healthy whole. There's no Department of Morals I would trust to weed 'em out or protect me or mine from them.

Today, the pendulum is swinging back to government control of speech, most notably on the net. First, it became clear that total surveillance is the norm even amongst Western democratic governments (the "total information act" reborn). Now we hear the UK government wants to be able to ban organisations without any evidence of involvement in illegal activities because they might "poison young minds". Well, nonsense. Frustrated young minds will go off to Syria precisely BECAUSE they feel their avenues for discourse and debate are being shut down by an unfair and unrepresentative government - you couldn't ask for a more compelling motivation for the next generation of home-grown anti-Western jihadists than to clamp down on discussion without recourse to due process. And yet, at the same time this is happening in the UK, protesters in Hong Kong are moving to peer-to-peer mechanisms to organise their protests precisely because of central control of the flow of information.

One of the reasons I picked the certificate and security business back in the 1990′s was because I wanted to be part of letting people communicate privately and securely, for business and pleasure. I'm saddened now at the extent to which the promise of that security has been undermined by state pressure and bad actors in the business of trust.

So I think it's time that those of us who invest time, effort and money in the underpinnings of technology focus attention on the defensibility of the core freedoms at the heart of the internet.

There are many efforts to fix this under way. The IETF is slowly become more conscious of the ways in which ideals can be undermined and the central role it can play in setting standards which are robust in the face of such inevitable pressure. But we can do more, and I'm writing now to invite applications for Fellowships at the Shuttleworth Foundation by leaders that are focused on these problems. TSF already has Fellows working on privacy in personal communications; we are interested in generalising that to the foundations of all communications. We already have a range of applications in this regard, I would welcome more. And I'd like to call attention to the Edgenet effort (distributing network capabilities, based on zero-mq) which is holding a sprint in Brussels October 30-31.

20 years ago, "Clipper" (a proposed mandatory US government back door, supported by the NSA) died on the vine thanks to a concerted effort by industry to show the risks inherent to such schemes. For two decades we've had the tide on the side of those who believe it's more important for individuals and companies to be able to protect information than it is for security agencies to be able to monitor it. I'm glad that today, you are more likely to get into trouble if you don't encrypt sensitive information in transit on your laptop than if you do. I believe that's the right side to fight for and the right side for all of our security in the long term, too. But with mandatory back doors back on the table we can take nothing for granted - regulatory regimes can and do change, as often for the worse as for the better. If you care about these issues, please take action of one form or another.

Law enforcement is important. There are huge dividends to a society in which people to make long term plans, which depends on their confidence in security and safety as much as their confidence in economic fairness and opportunity. But the agencies in whom we place this authority are human and tend over time, like any institution, to be more forceful in defending their own existence and privileges than they are in providing for the needs of others. There has never been an institution in history which has managed to avoid this cycle. For that reason, it's important to ensure that law enforcement is done by due process; there are no short cuts which will not be abused sooner rather than later. Checks and balances are more important than knee-jerk responses to the last attack. Every society, even today's modern Western society, is prone to abusive governance. We should fear our own darknesses more than we fear others.

A fair society is one where laws are clear and crimes are punished in a way that is deemed fair. It is not one where thinking about crime is criminal, or one where talking about things that are unpalatable is criminal, or one where everybody is notionally protected from the arbitrary and the capricious. Over the past 20 years life has become safer, not more risky, for people living in an Internet-connected West. That's no thanks to the listeners; it's thanks to living in a period when the youth (the source of most trouble in the world) feel they have access to opportunity and ideas on a world-wide basis. We are pretty much certain to have hard challenges ahead in that regard. So for all the scaremongering about Chinese cyber-espionage and Russian cyber-warfare and criminal activity in darknets, we are better off keeping the Internet as a free-flowing and confidential medium than we are entrusting an agency with the job of monitoring us for inappropriate and dangerous ideas. And that's something we'll have to work for.

30 Sep 2014 2:24pm GMT

Dustin Kirkland: Apply updates to multiple systems simultaneously using Byobu and Shift-F9

A StackExchange question, back in February of this year inspired a new feature in Byobu, that I had been thinking about for quite some time:

Wouldn't it be nice to have a hot key in Byobu that would send a command to multiple splits (or windows?

This feature was added and is available in Byobu 5.73 and newer (in Ubuntu 14.04 and newer, and available in the Byobu PPA for older Ubuntu releases).

I actually use this feature all the time, to update packages across multiple computers. Of course, Landscape is a fantastic way to do this as well. But if you don't have access to Landscape, you can always do this very simply with Byobu!

Create some splits, using Ctrl-F2 and Shift-F2, and in each split, ssh into a target Ubuntu (or Debian) machine.

Now, use Shift-F9 to open up the purple prompt at the bottom of your screen. Here, you enter the command you want to run on each split. First, you might want to run:

sudo true

This will prompt you for your password, if you don't already have root or sudo access. You might need to use Shift-Up, Shift-Down, Shift-Left, Shift-Right to move around your splits, and enter passwords.

Now, update your package lists:

sudo apt-get update

And now, apply your updates:

sudo apt-get dist-upgrade

Here's a video to demonstrate!

In a related note, another user-requested feature has been added, to simultaneously synchronize this behavior among all splits. You'll need the latest version of Byobu, 5.87, which will be in Ubuntu 14.10 (Utopic). Here, you'll press Alt-F9 and just start typing! Another demonstration video here...


30 Sep 2014 1:44pm GMT

Raphaël Hertzog: My Debian LTS report for September

Thanks to the sponsorship of multiple companies, I have been paid to work 11 hours on Debian LTS this month.

CVE triagingI started by doing lots of triage in the security tracker (if you want to help, instructions are here) because I noticed that the dla-needed.txt list (which contains the list of packages that must be taken care of via an LTS security update) was missing quite a few packages that had open vulnerabilities in oldstable.

In the end, I pushed 23 commits to the security tracker. I won't list the details each time but for once, it's interesting to let you know the kind of things that this work entailed:

And when I say "I reviewed" it's a simplification for this kind of process:

CVE triaging is often almost half the work in the general process: once you know that you are affected and that you have a patch, the process to release an update is relatively straightforward (sometimes there's still work to do to backport the patch).

Once I was over that first pass of triaging, I had already spent more than the 11 hours paid but I still took care of preparing the security update for python-django. Thorsten Alteholz had started the work but got stuck in the process of backporting the patches. Since I'm co-maintainer of the package, I took over and finished the work to release it as DLA-65-1.

No comment | Liked this article? Click here. | My blog is Flattr-enabled.

30 Sep 2014 1:24pm GMT

The Fridge: Ubuntu Weekly Newsletter Issue 385

Welcome to the Ubuntu Weekly Newsletter. This is issue #385 for the week September 22 - 28, 2014, and the full version is available here.

In this issue we cover:

The issue of The Ubuntu Weekly Newsletter is brought to you by:

If you have a story idea for the Weekly Newsletter, join the Ubuntu News Team mailing list and submit it. Ideas can also be added to the wiki!

Except where otherwise noted, content in this issue is licensed under a Creative Commons Attribution 3.0 License BY SA Creative Commons License

30 Sep 2014 3:16am GMT

29 Sep 2014

feedPlanet Ubuntu

Stuart Langridge: The next big thing is privacy

The way you beat an incumbent is by coming up with a thing that people want, that you do, and that your competitors can't do.

Not won't. Can't.

How did Apple beat Microsoft? Not by making a better desktop OS. They did it by shifting the goalposts. By creating a whole new field of competition where Microsoft's massive entrenched advantage didn't exist: mobile. How did Microsoft beat Digital and the mainframe pushers? By inventing the idea that every desktop should have a real computer on it, not a terminal.

How do you beat Google and Facebook? By inventing a thing that they can't compete against. By making privacy your core goal. Because companies who have built their whole business model on monetising your personal information cannot compete against that. They'd have to give up on everything that they are, which they can't do. Facebook altering itself to ensure privacy for its users… wouldn't exist. Can't exist. That's how you win.

If you ask actual people whether they want privacy, they say, yes. Always. But if you then ask, are they, are we, prepared to give that privacy up to get things? They say yes again. They, we, want privacy, but not as much as we want stuff. Not as much as we want to talk to one another. Giving up our personal data to enable that, that's a reasonable cost to pay, because we don't value our personal data. Some of that's because there's no alternative, and some of that's because nobody's properly articulated the alternative.

Privacy will define the next major change in computing.

We saw the change to mobile. The change to social. These things fundamentally redefined the way technology looked to the mainstream. The next thing will be privacy. The issue here is that nobody has worked out a way of articulating the importance of privacy which convinces actual ordinary people. There are products and firms trying to do that right now. Look at Blackphone. Look at the recent fertile ground for instant messaging with privacy included from Telegram and Threema and Whisper System's Text Secure. They're all currently basically for geeks. They're doing the right thing, but they haven't worked out how to convince real people that they are the right thing.

The company who work out how to convince people that privacy is important will define the next five years of technology.

Privacy, historically the concern of super-geeks, is beginning to poke its head above the parapet. Tim Berners-Lee calls for a "digital Magna Carta". The EFF tries to fix it and gets their app banned because it's threatening Google's business model to have people defend their own data. The desire for privacy is becoming mainstream enough that the Daily Mash are prepared to make jokes about it. Apple declare to the world that they can't unlock your iPhone, and Google are at pains to insist that they're the same. We're seeing the birth of a movement; the early days before the concern of the geeks becomes the concern of the populace.

So what about the ind.ie project?

The ind.ie project will tell you that this is what they're for, and so you need to get on board with them right now. That's what they'll tell you.

The ind.ie project is to open source as Brewdog are to CAMRA. Those of you who are not English may not follow this analogy.

CAMRA is the Campaign for Real Ale: a British society created in the 1970s and still existing today who fight to preserve traditionally made beer in the UK, which they name "real ale" and have a detailed description of what "real ale" is. Brewdog are a brewer of real ale who were founded in 2007. You'd think that Brewdog were exactly what CAMRA want, but it is not so. Brewdog, and a bunch of similar modern breweries, have discovered the same hatred that new approaches in other fields also discovered. In particular, Brewdog have done a superb job at bringing a formerly exclusive insular community into the mainstream. But that insular community feel resentful because people are making the right decisions, but not because they've embraced the insular community. That is: people drink Brewdog beer because they like it, and Brewdog themselves have put that beer into the market in such a way that it's now trendy to drink real ale again. But those drinking it are not doing it because they've bought into CAMRA's reasoning. They like real ale, but they don't like it for the same reasons that CAMRA do. As Daniel Davies said, every subculture has this complicated relationship with its "trendy" element. From the point of view of CAMRA nerds, who believe that beer isn't real unless it has moss floating in it, there is a risk that many new joiners are fair-weather friends just jumping on a trendy bandwagon and the Brewdog popularity may be a flash in the pan. The important point here is that the new people are honestly committed to the underlying goals of the old guard (real ale is good!) but not the old guard's way of articulating that message. And while that should get applause, what it gets is resentment.

Ind.ie is the same. They have, rather excellently, found a way of describing the underlying message of open source software without bringing along the existing open source community. That is, they've articulated the value of being open, and of your data being yours without it being sold to others or kept as commercial advantage, but have not done so by pushing the existing open source message, which is full of people who start petty fights over precisely which OS you use and what distribution A did to distribution B back in the mists of prehistory. This is a deft and smart move; people in general tend to agree with the open source movement's goals, but are hugely turned off by interacting with that existing open source movement, and ind.ie have found a way to have that cake and eat it.

Complaints from open source people about ind.ie are at least partially justified, though. It is not reasonable to sneer at existing open source projects for knowing nothing about users and at the same time take advantage of their work. It is not at all clear how ind.ie will handle a bunch of essential features - reading an SD card, reformatting a drive, categorising applications, storing images, sandboxing apps from one another, connecting to a computer, talking to the cloud - without using existing open source software. The ind.ie project seem confident that they can overlay a user experience on this essential substrate and make that user experience relevant to real people rather than techies; but it is at best disingenuous and at worst frankly offensive to simultaneously mock open source projects for knowing nothing about users and then also depend on their work to make your own project successful. Worse, it ignores the time and effort that companies such as Canonical have put in to user testing with actual people. It's blackboard economics of the worst sort, and it will have serious repercussions down the line when the ind.ie project approaches one of its underlying open source projects and says "we need this change made because users care" and the project says "but you called us morons who don't care about users" and so ignores the request. Canonical have suffered this problem with upstream projects, and they were nowhere near as smugly, sneeringly dismissive as ind.ie have been of the open source substrate on which they vitally depend.

However, they, ind.ie, are doing the right thing. The company who work out how to convince people that privacy is important will define the next five years of technology. This is not an idle prediction. The next big wave in technology will be privacy.

There are plenty of companies right now who would say that they're already all over that. As mentioned above, there's Blackphone and Threema and Telegram and ello and diaspora. All of them are contributors and that's it. They're not the herald who usher in the next big wave. They're ICQ, or Friends Reunited: when someone writes the History Of Tech In The Late 2010s, Blackphone and ello and Diaspora will be footnotes, with the remark that they were early adopters of privacy-based technology. There were mp3 players before the iPod. There were social networks before Facebook. All the existing players who are pushing privacy as their raison d'etre and writing manifestos are creating an environment which is ripe for someone to do it right, but they aren't themselves the agent of change; they're the Diamond Rio who come before the iPod, the ICQ who come before WhatsApp. Privacy hasn't yet found its Facebook. When it does, that Facebook of privacy will change the world so that we hardly understand that there was a time when we didn't care about it. They'll take over and destroy all the old business models and make a new tech universe which is better for us and better for them too.

I hope it comes soon.

29 Sep 2014 11:41pm GMT

Ben Howard: Cloud Images and Bash Vulnerabilities

Cloud Images and Bash Vulnerabilities

The Ubuntu Cloud Image team has been monitoring the bash vulnerabilities. Due to the scope, impact and high profile nature of these vulnerabilties, we have published new images. New cloud images to address the lastest bash USN-2364-1 [1, 8, 9] are being released with a build serials of 20140927. These images include code to address all prior CVEs, including CVE-2014-6271 [6] and CVE-2014-7169 [7], and supersede images published in the past week which addressed those CVEs.

Please note: Securing Ubuntu Cloud Images requires users to regularly apply updates[5]; using the latest Cloud Images are insufficient.

Addressing the full scope of the Bash vulnerability has been an iterative process. The security team has worked with the upstream bash community to address multiple aspects of the bash issue. As these fixes have become available, the Cloud Image team has published daily[2]. New released images[3] have been made available at the request of the Ubuntu Security team.

Canonical has been in contact with our public Cloud Partners to make these new builds available as soon as possible.

Cloud image update timeline

Daily image builds are automatically triggered when new package versions become available in the public archives. New releases for Cloud Images are triggered automatically when a new kernel becomes available. The Cloud Image team will manually trigger new released images when either requested by the Ubuntu Security team or when a significant defect requires.

Please note: Securing Ubuntu cloud images requires that security updates be applied regularly [5], using the latest available cloud image is not sufficient in itself. Cloud Images are built only after updated packages are made available in the public archives. Since it takes time to build the images, test/QA and finally promote the images, there is time (sometimes considerable) between public availablity of the package and updated Cloud Images. Users should consider this timing in their update strategy.

[1] http://www.ubuntu.com/usn/usn-2364-1/
[2] http://cloud-images.ubuntu.com/daily/server/
[3] http://cloud-images.ubuntu.com/releases/
[4] https://help.ubuntu.com/community/Repositories/Ubuntu/
[5] https://wiki.ubuntu.com/Security/Upgrades/
[6] http://people.canonical.com/~ubuntu-security/cve/2014/CVE-2014-6271.html
[7] http://people.canonical.com/~ubuntu-security/cve/2014/CVE-2014-7169.html
[8] http://people.canonical.com/~ubuntu-security/cve/2014/CVE-2014-7187.html
[9] http://people.canonical.com/~ubuntu-security/cve/2014/CVE-2014-7186.html

29 Sep 2014 5:45pm GMT

Ubuntu Podcast from the UK LoCo: S07E26 – The One Where Underdog Gets Away

We're back with Season Seven, Episode Twenty-Six of the Ubuntu Podcast! Just Alan Pope and Laura Cowen with a set of interviews from Mark Johnson this week.

Download Ogg Download MP3 Play in Popup

In this week's show:

python 2.x:
python -m SimpleHTTPServer [port]

python 3.x
python -m http.server [port]

We'll be back next week, so please send your comments and suggestions to: podcast@ubuntu-uk.org
Join us on IRC in #uupc on Freenode
Leave a voicemail via phone: +44 (0) 203 298 1600, sip: podcast@sip.ubuntu-uk.org and skype: ubuntuukpodcast
Follow us on Twitter
Find our Facebook Fan Page
Follow us on Google+

29 Sep 2014 3:25pm GMT

Charles Butler: Juju Powered Radio! A protoduction experiment.

September 26'th I undertook a rather daunting task of trialing something I strongly believe in that really took me out of my comfort zone and put me front and center of an audience's attention, for not only my talents, but also the technical implementation of their experience.

The back story

I've been amateur DJ'ing on Secondlife for about the last 7 months, and recently left the metaverse to pursue a podcast format of my show(s). What I found was I really missed the live interaction with people during the recording of the set. It was great to get feedback, audience participation, and I could really gauge the flow of energy that I'm broadcasting. To some this may sound strange, but when your primary interaction is over text, and you see a feed erupt with actions as you put on more high energy music, it just 'clicks' and makes sense.

The second aspect to this was I wanted to showcase how you can get moving with Juju in less than a week to bring a production ready app online and ready for scale (depending on the complexity of the app of course). It's been a short while since I've pushed a charm from scratch into the charm store - and this will definately get me re-acquiainted with the process our new users go through on their Juju journey.

So, I've got a habit of mixing my passions in life. If you know me very well you know that I am deeply passionate about what I'm working on, my hobbies, and the people that I surround myself with that i consider my support network. How can I leverage this to showcase and run a 'Juju lab' study?

The Shoutcast charm is born

I spent a sleepless night hacking away at a charm for a SHOUTCast DNAS server. They offer several PAAS, scaling solutions that might work for people that are making money off of their hobby - but I myself prefer to remain an enthusiast and not turn a profit from my hobby. Juju is a perfect fit for deploying pretty much anything, and making sure that all the components work together in a distributed service environment. It's getting better every day - proof of this is the Juju GUI just announced machine view - where you can easily do co-location of services on the same server, and get a deep dive look at how your deployment is comprised of machines vs services.

Observations & Lessons

Testing what you expect, never yields the unexpected

Some definate changes to just the shoutcast charm itself are in order.

Machine Metrics tell most of the story

I deployed this setup on Digital Ocean to run my 'lab test' - as the machines are cheap, performant, and you get 1TB of transfer unmetered before you have to jump up a pricing teir. This is a great mixture for testing the setup. But how well did the VPS perform?

I consumed 2 of the 'tiny' VPS servers for this. And the metrics of the transcoders were light enough that it barely touched the CPU. As a matter of fact I saw more activity out of supporting infra services such as LogStash, than I did out of the SHOUTCast charm. Excellent work on the implementation Shoutcast devs. This was a pleasant surprise!

Pre-scaling was the winner

Having a relay setup out of the gate really helped to mitigate issues as I saw people get temporary hiccups in their network. I saw several go from the primary stream to the relay and finish out the duration of the broadcast connected there.

The fact that the clients supported this, tells me that any time I do this live, I need to have at bare minimum 2 hosts online transmitting the broadcast.

Had this been a single host - every blip in the network would yield dead airspace before they realized something had gone wrong.

Juju Scaled Shoutcast Service

Supportive people are amazing, and make what you do, worthwhile

Those that tuned in genuinely enjoyed that I had the foresight to pre-record segments of the show to interact with them. This was more so I could investigate the server(s), watch htop metrics, refresh shoutcast, etc. However the fan interaction was genuinely empowering. I found myself wanting to turn around and see what was said next during the live-mixing segments.

The Future for Radio Bundle Development

Putting the auto in automation

I've found a GREAT service that I want to consume and deploy to handle the station automation side of this deployment. SourceFabric produces Airtime which makes setting up Radio Automation very simple, and supports such advanced configurations as mixing in Live DJ's into your lineup on a schedule. How awesome is this? It's open-source to boot!

I'm also well on my way to having revision 1 of this bundle completed, since I started the blog post on Friday. Hacked on the bundle through the weekend, and landed here on Monday.

I'll be talking more about this after it's officially unveiled in Brussels.

Where to find the 'goods'

The Shoutcast Juju Charm can be found on Launchpad: lp:~lazypower/charms/trusty/shoutcast/trunk or github

The up-coming Airtime Radio Automation Charm can be found on github

Actual metrics and charts to be uploaded at a later date, once I've sussed out how I want to parse these and present them.

29 Sep 2014 1:36pm GMT

28 Sep 2014

feedPlanet Ubuntu

Aurélien Gâteau: Experimenting with keyboard shortcuts

A few weeks ago, I decided to make an experiment and completely rework the global shortcuts of my KDE desktop. I wanted them to make a bit more sense instead of being the agglomerated result of inspirations from other systems, and was ready to pay the cost of brain retraining.

My current shortcut setup relies on a few "design" decisions:

I am still playing with it, but it is stabilizing these days, so I thought I'd write a summary of what I came up with:

Window management

Virtual desktop

Application launch


How does it feel?

I was a bit worried about the muscle-memory retraining, but it went quite well. Of course I am a bit lost nowadays whenever I use another computer, but that was to be expected.

One nice side-effect I did not foresee is that this change turned the Win modifier into a sort of quasimode: all global workspace operations are done by holding the Win key. I said "sort of" because some operations requires you to release the Win key before they are done, for example when switching from one window to another, no shortcuts work as long as the window switcher is visible, so one needs to release the Win key after switching and press it again to do something else. I notice this most often when maximizing left or right.

Another good point of this approach is that, almost no shortcuts use function keys. This is a good thing because: a) it can be quite a stretch for small hands to hold both the Win or Alt modifier together with a function key and b) many laptops these days come with the function keys mapped to multimedia controls and need another modifier to be held to become real function keys, some other laptops do not even come with any function keys at all! (heresy I know, but such is the world we live in...)

What about you, do you have unusual shortcut setups?

Flattr this

28 Sep 2014 6:09pm GMT