Among the first modern computers outside military applications such as ‘Colossus’ at Bletchley Park was EDSAC, built at Cambridge University. This became the basis for LEO 1 (or Lyons Electronic Office 1), the first commercial computer, developed for the J Lyons and Co. bakery and restaurant chain. Meanwhile at Manchester University, the Manchester Baby had evolved into the Manchester Mark 1 which became the basis for the Ferranti Mark 1. At this time computer software was provided and maintained free with the hardware as it would only run on the hardware for which it was written — a situation common until recently with embedded computer systems such as in-car electronics and mobile ’phones.
The first software capable of running on different hardware was developed at IBM’s research centre at Hursley in Hampshire. However, most people were not aware of this until the 1970s when the development of Unix for mainframes and CP/M for micro-computers allowed people to write software independently of the hardware in much the same way as the Android operating system has allowed people to write software for a wide variety of mobile ’phones. But, as programmers did not want people to buy their software and hand it on to anyone else, they hit on the idea of ‘licensing’ the software to individual users who were barred from modifying it in any way. Among those who profited from this new approach were Bill Gates and Paul Allen who had written some software for one of the newfangled micro-computers and went on to build a multinational company, Microsoft, on the back of this new approach.
At the same time computer companies were beginning to sell computers to organisations which did not have skilled technicians on whom they could rely to report and, in some cases, solve problems for the mutual benefit of all. Companies without skilled technicians expected the computer company to solve all their problems. Not wanting to be called on to sort out problems caused by an unskilled person, computer companies began charging for supporting the software and imposing restrictions on what people could do with it.
IBM had begun to sell some of its software separately from its hardware in 1969 and in 1982, at the end of the long running anti-trust litigation brought against it by the US Department of Justice, it began charging for all its software separately from its hardware though some people thought it ‘underpriced’ it to keep the competition at bay (Glass, 2005).
Meanwhile, as the US economy flagged, US legislators hit on the idea of extending copyright to provide a greater incentive for innovation. In 1980 it became possible to copyright software in the US while the courts became more favourable to patenting software, a position formalised in law in 1990.
During the 1980s media publishers began to argue that strengthening copyright laws would benefit the US economy and by 1990 they had persuaded the US government to include in trade deals the requirement that other countries would respect US copyright laws. In 1992 breaking copyright law became a criminal, not just a civil, offence in the US and in 1994 the US was able to make respecting its copyright laws part of GATT (General Agreement on Trade and Tariffs), which was succeeded the following year by the World Trade Organization.
In 1998 the media publishers persuaded Bill Clinton to sign into law the Digital Millennium Copyright Act which imposes penalties for circumventing digital copyright systems and have continued to pursue copyright ‘pirates’ ever since.
These changes had been justified as part of neo-liberal economic theory about creating a free market but they actually had the effect of severely curtailing the freedom of the individual which is supposed to be at the heart of such theories (Coleman, 2013).
Though the original developers of Unix were based at AT&T’s Bell Labs, much of the development of the software had taken place in universities. As bodies in receipt of public funds they were required by US legislation to make their research publicly available. Initially, they did this by printing the code for the software in research papers but eventually the length of the code became too great to include in papers and they began publishing it separately as computer files. To comply with US legislation, they created permissive licences, that is, ones allowing users to modify and publish their changes to the code, of which the MIT (Massachusetts Institute of Technology) and BSD (Berkeley Systems Department) licences are probably the most well-known.
Because AT&T was virtually a monopoly communications company in the US, it had been barred from selling computing equipment but, in 1984, it was broken up and released from this restriction. Once the ban was lifted, AT&T began charging people for Unix. This created problems for the universities whose contributions to Unix were now being sold, in apparent contravention of their legal obligations, and annoyed many of the academics who had contributed to it. Over the next decade, there were numerous attempts to resolve this impasse including developing different versions of Unix based on the universities’ work, of which BSD Unix, which is used by Apple, is probably the most well-known. The impasse was finally resolved in 1993 when the US computer company, Novell, bought the rights to Unix from AT&T and declared that it would not charge anyone for using Unix.1
Before this, programmers had been exchanging free (and not so free) software through locally based ‘Bulletin Boards’ but, concerned at the increasing restrictions being placed on programmers by the commercialisation of software and fearful that his work might be hijacked by his employer, the Massachusetts Institute of Technology, Richard Stallman, who worked at the Artificial Intelligence Laboratory, resigned his post in 1984 and recruited a number of like-minded programmers to create the Free Software Foundation in 1985 and articulate in the ‘GNU Manifesto’ the four freedoms:2
In 1989 he formalised these principles in the GPL (GNU General Public License) which, in its several versions, is now used to protect free and open source software. It differs markedly from traditional forms of copyright in asserting both the right of the author to be identified as the author of the software and the rights of the user to use the software in the ways set out in the GNU Manifesto.
Stallman and his colleagues tried to develop a free version of Unix — a project as yet unfinished — in the course of which they managed to write a lot of free software, sometimes collectively known as the GNU utilities, which runs on most Unix-like systems. Their stance was supported by Tim Berners-Lee (1999), who developed the World Wide Web, which has meant that most of the software used to run the Internet is free and open source software.
1 In 2011 Novell was taken over by Attachmate, who announced that they would continue Novell’s approach to the rights to Unix. When Attachmate was taken over by MicroFocus in 2014, there was no change to this policy.
2 The term is used in imitation of the four freedoms articulated by US President Franklin D. Roosevelt on January 6, 1941.
The term ‘open source’ was not coined until 1998 (Biancuzzi, 2008) and it is a little misleading since, apart from the proprietary software developed since the 1970s and ‘shareware,’ that is, software distributed free but normally with a request to make a donation to the author, most software is ‘open source’ in the sense that anyone can read the code. The term was coined in 1998 to cover a new way of creating and supporting computer software, initiated in 1988 by Intel. They hired Michael Tiemann to write the software for a new engineering chip and then gave it away with the chip, telling customers who were themselves not sufficiently skilled to use it to contact Michael for help. In association with John Gilmore and David Henkel-Wallace, he built up Cygnus Support to support these customers and develop a variety of free and open source software including a number of the GNU utilities.3
The key feature of this model is that creating software and supporting it became separate activities with charges only being made for support. It was taken up by other companies, most notably the Swedish database company MySQL4 which gave away its software but offered support contracts for people needing help to get the most out of it.
3 In 1999 Cygnus Solutions, as it was called by then, merged with RedHat.
4 In 2008 Sun Microsystems, a long-standing manufacturer of high performance computers, acquired MySQL for $1 billion dollars. Since Oracle acquired Sun Microsystems in 2010, MySQL is now supported by Oracle. Meanwhile, one of original developers of MySQL created MariaDB as an alternative to MySQL as a result of fears about the direction in which Sun Microsystems was taking MySQL.
With increasing adoption of Unix on larger computers in the 1980s, Andrew Tannenbaum had created a smaller alternative, Minix, to use in teaching students the principles of Unix. In 1991 a computing student at Helsinki University, Linus Torvalds, became so impatient waiting to get onto the university computer that he used Minix as the starting point for building his own operating system which, with the help of the systems administrator at the university, who called it ‘Linux,’ he put on the Internet. Within 36 hours he had people contacting him and within 18 months they had developed a kernel which worked well enough with the software which the universities and Richard Stallman and his colleagues had developed to make it a viable system for use on PCs using Intel processors. A contributory factor was the GPL (General Public License) created by the Free Software Foundation which facilitated collaboration among developers.
Among those who spotted its potential were Patrick Volkerding who released Slackware on 17 July 1993 and Debbie and Ian Murdoch, whose Debian distribution (17 September 1993) underpins most desktop Linux systems. Volunteers continue to maintain and develop Debian which remains the most popular basis for Linux distributions, including Ubuntu, the most popular. Another was the German company S.u.S.E. (now SUSE) which distributed Softlanding Linux System and Slackware before creating a German version of Slackware in 1994 and their own distribution in 1996.5 Along with RedHat, a US company set up in 1993 to distribute Linux and now part of IBM, they are among the companies supporting Linux for business customers.
Apart from the fact that Linux would run on PCs as opposed to the mainframes needed for most varieties of Unix, it also had the advantage of more similarities than differences among the different Linux distributions, something which could not be said for Unix at the time.
In 1996 Digital Equipment Corporation approached Linus Torvalds to adapt Linux for their new Alpha processors;6 he wrote this up for his Masters degree, in the process making it much easier to use the Linux kernel with any computer chip. This flexibility meant that Linux began to appear on a wide range of devices from supercomputers (Linux runs all the top 500 supercomputers) to webcams.
5 In 2003 they were absorbed by Novell; when Novell was taken over by Attachmate in 2011, Novell and SUSE became separate divisions, an arrangement which continued after MicroFocus acquired Attachmate in 2014. In 2018 MicroFocus announced that they would sell SUSE as a going concern to Swedish investment company, EQT Partners, who in turn announced that SUSE would continue to operate as an independent company.
6 Digital Equipment Corporation was acquired in 1998 by Compaq and they in turn were acquired in 2002 by Hewlett-Packard who continued to produce the Alpha processor until 2004.
During the 1990s another group of programmers developed some free server software to facilitate communication between computer programs on different computers. IBM was so impressed by Apache that it wanted to use it; however, in the light of the anti-trust settlement IBM had made in 1982, it wanted to know the provenance of every line of code they used. At the time, the Apache developers had collaborated without keeping a check on who did what; so they had to go through the code and document who had done what and they also had to create a legal entity with whom IBM could make an agreement. This led to the creation of the Apache Software Foundation in 1999 and to IBM’s programmers beginning to contribute code to the Apache server.
It was also the start of a major shift towards the open source model for software development and support which Intel had pioneered. Major computer companies now contribute both programming and financial resources to charitable foundations who manage software development rather than competing among themselves to develop software. They then make their money offering specialised support to users as Michael Tiemann had first done.
Among these charitable foundations are the Free Software Foundation, the World Wide Web Consortium (W3C), Software in the Public Interest (originally set up to host the Debian distribution but now acting as the umbrella organisation for a number of projects) and ICANN, which manages the allocation of Internet domain names. The most well-known charities involved in developing free and open source software are probably:
They were joined in 2010 by the Document Foundation which was set up to continue the development, under the name LibreOffice,7 of OpenOffice, the open source office suite developed under the auspices of Sun Microsystems, and in 2013 by the MariaDB Foundation, set up to develop MariaDB as an alternative to MySQL after it had been acquired by Sun Microsystems.
All also provide software which runs on proprietary systems such as Microsoft Windows while Microsoft has contributed to the Linux Foundation to enable Windows computers to work more efficiently with Linux computers.
7 The name OpenOffice is owned by Oracle as a result of its acquisition of Sun Microsystems. Oracle declined an invitation to join the Document Foundation and in 2011 offered the OpenOffice suite to the Apache Foundation.
In 2007, when Apple released the iPhone, 95% of the devices connected to the Internet were running Microsoft Windows; today less than 15% are as Apple’s iOS and Android, which has Linux at its heart, dominate the tablet and smartphone markets and most people no longer use a PC to connect to the Internet.
In 2013 Eben Upton, formerly an admissions tutor at Cambridge University who had become frustrated by the low level of computing knowledge of applicants to the university’s computing courses, launched the Raspberry Pi, a credit card sized computer on a chip running Linux and intended for schools. It became an overnight sensation and has gone on to be developed in a variety of ways and used in ways, like creating a supercomputer by running 64 of them in parallel, which the inventors never envisaged.
It is important to stress that no one who contributes to free and open source software loses copyright; indeed, because everyone who contributes retains copyright, free and open source programs often have very long lists of all those who have contributed unlike proprietary programs where copyright is owned by the company. So an immediate advantage for programmers is that, if they are going for a job, there is a publicly available record of what they have done which any potential employer can examine.
A second advantage is that all software is peer reviewed, either directly by other programmers working on the program or indirectly by users using and commenting on it. Though this system is not foolproof, potential weaknesses are more likely to be revealed in free and open source software than they are in proprietary software (Mockus et al., 2005).
A third advantage is the wide range of people involved in software development — from full-time programmers through employees whose companies allow them to contribute code as part of their work and people who program in their spare time to people who have limited programming skills but are prepared to contribute as users though bug-hunting and developing documentation. The majority of contributors to a project may only contribute one small change but the sum total of those small changes makes the key contributors’ job much easier. No individual company could ever gain access to such a wide spectrum of skills or feedback during the development of software.
A fourth advantage is that you can adapt the software to meet your needs without having to ask anyone’s permission and in most cases without telling anyone. You only need to understand what the software licence says if you decide to share your changes with anyone else. The most restrictive licence, known as GPL3, requires you to share any changes with everyone and prohibits you from linking the software with proprietary software; others have lower requirements about publishing the changes and do not place any limits on links with proprietary software.
Linus Torvalds is somewhat bemused at how far his student project has gone and in particular the way in which it has opened up the possibility for third world countries like Nepal, Bhutan and Vietnam to develop home-grown computer systems. Three technical features of Linux have been particularly helpful, its limited demands on hardware, the adoption of the
utf-8 encoding of Unicode and the way in which the instructions that appear on the screen are stored. But it has also had a major impact on the ways people work together.
The Linux kernel consists of a small core and modules which can be available all the time, on-call or not at all. So programmers only use the modules relevant to what they want to do; that is why Linux can run on a supercomputer or a digital television. If your computer has limited hardware, you can create your own version to suit the limited hardware, something highly relevant in developing countries.
Unicode is a sequence of all the characters in all the known languages, living and dead, in the world. Because there are so many characters, a font only covers a limited number of the possible characters but it is theoretically possible, using the right combination of fonts, to write documents in any known language using any Linux program. Since the
utf-8 encoding has also been adopted as the only language encoding to be used on websites under HTML5 and ICANN has approved the use of non-Roman scripts for website names, something only Unicode can handle comprehensively, Linux is the perfect operating system for website development.
More importantly, because the instructions that appear on screen are stored in simple lists within the program, it is relatively easy, if tedious in programs that have lots of messages for the user, for someone to create a translation of any Linux program as has happened in Nepal, Bhutan and Vietnam. No one needs to ask permission and it depends not on a company decision but on the willingness of volunteers or, in some cases, governments to support people to do the job.
Fifty years ago Richard Titmuss (1970) found that, in relation to blood donation, people who give blood free are more likely than paid donors to be committed to the quality of what they are giving. The FOSS community has gone one step further in bringing together paid and voluntary contributors in the joint pursuit of quality. Volunteers, whose only public reward is to see their name in the list of contributors, overwhelmingly outnumber those who are supported by their employers and the leaders of free and open source software projects have had to hone their relationship skills to sustain that commitment.
There have been some problems with cultures where giving and, in particular, giving feedback is not the norm, where people have been reluctant to engage fully in the FOSS ethos. There have also been problems with people who have adopted exploitative or domineering approaches to software development and there has sometimes been frustration with those who only take and never give back. But, relative to the success of FOSS worldwide, these have been irritations rather than major hurdles to its development. FOSS has developed, and continues to develop, out of a spirit of giving, whether by companies, governments or individuals. It is a model for co-operation between government, commercial and voluntary organisations that has stood the test of time, delivering quality outcomes to far more people in the world than possibly any other community or voluntary enterprise ever.
Berners-Lee, T. (1999). Weaving the Web: the past, present and future of the World Wide Web by its inventor. London: Orion Business Books.
Biancuzzi, F. (2008, 7 April). Open source decade: 10 years after the Free Software Summit. http://arstechnica.com/information-technology/2008/04/free-software-summit-10th/.
Coleman, E. G. (2013). Coding freedom: the ethics and aesthetics of hacking. Oxford: Princeton University Press.
Glass, R. L. (2005). Standing in front of the open source steamroller. In J. Feller, B. Fitzgerald, S. A. Hissam, and K. R. Lakhani (Eds.), Perspectives on free and open source software, Chapter 4, pp. 81–92. Cambridge MA/London: MIT Press.
Mockus, A., R. T. Fielding, and J. D. Herbsleb (2005). Two case studies of open source software development: Apache and Mozilla. In J. Feller, B. Fitzgerald, S. A. Hissam, and K. R. Lakhani (Eds.), Perspectives on free and open source software, Chapter 10, pp. 163–209. Cambridge MA/London: MIT Press.
Titmuss, R. M. (1970). The gift relationship: from human blood to social policy., London: Allen & Unwin.