Linux is a free UNIX clone that was developed by volunteers around the world. Although Linux is almost a decade old, it went largely unnoticed by the general public until a couple of years ago. Since then it has become very popular with individual users, universities and large corporations. For example, IBM has made it a major part of their business strategy for server hardware. Many software companies already offer Linux versions of their software, including products such as Oracle, Labview and MSC Nastran. Other companies have developed embedded applications using Linux. Currently Linux can be found in devices as small as a wristwatch  and as large as a Beowulf class supercomputer . The popularity of Linux is based on three factors:
This workshop will present the Linux operating system in general, and its current status in computing. This will be followed by a discussion of the basic features and operation of a Linux system. User applications will also be discussed, and demonstrated. Finally the future of the operating system will be discussed. An outline of the presentation is given below.
Linux is an open source operating system. It is open because users and developers can use the source code any way they want. This allows anyone to customize it, improve it and add desired features. As a result Linux is dynamic, evolving to respond to the desires and needs of the users. In contrast, closed operating systems are developed by a single corporation using static snapshots of market models and profit driven constraints.
Linux is free. This allows companies to use it without adding cost to products. It also allows people to trade it freely. And, with the profit motive gone, developers have a heightened sense of community interest. The Linux community has developed a tremendous spirit because of these core development concepts.
Linux has existed since the early 1990s , but it grew out of previous developments in computing . It was originally developed to be a Unix clone that would run on low cost computer hardware. Unix was developed in the 1970s. Through the 1970s and early 1980s it was used on large computers in companies and universities. During this time many refinements and enhancements were made. By the mid 1980s Unix was being used on many lower priced computers. By the end of the 1980s most universities were making use of Unix computers in computer science and engineering programs. This created a wealth of graduates who understood what they could expect from a mature operating system. But, it also created a demand to be able to do high level work at home on low priced machines.
Early in the 1990s Linux started as a project to create a Unix clone that would run on a personal computer. This project gained momentum quickly and by the mid 1990s it was ready for users. The first groups to adopt it were hobbyists, academics and internet services. At this time the general public was generally unaware of Linux but by the end of the 1990s it was beginning to enter the public sphere. By 2000 it had entered the popular press, and it was cited as a major threat to at least one existing operating system vendor. Now, it is available off-the-shelf in software and book stores.
Modern computers have ample power for most computer applications. This is more true for Linux. At present there are versions of linux that will run on any platform from an IBM mainframe computer to a Palm Pilot. The smallest Linux installations can fit on a single floppy disk, and run on a diskless computer with a few MB of memory. On the other end of the spectrum, Linux will run on most high end computer systems. An average user would expect reasonable performance on a computer with an old Pentium 100 processor, 64MB of memory, and 2 GB of disk space. On newer computers the performance of the operating system is extremely fast. The list below gives some idea of the capabilities, but complete lists of supported hardware are available .
By itself an operating system is somewhat useless, software applications are added to give desired functionality. Some of the common applications that a computer might be used for are listed below. Linux will support all of these applications, and more, with the right software .
A partial list of advantages and disadvantages is given below. The cost, stability and open nature of the system have been winning over a large number of corporate adopters. But, adoption has been slowed by people who don’t understand the nature of free software or have a perception that it is difficult to use. In some cases there are also some software packages that are not available for Linux, and won’t run under simulators  - the most notable of these applications are first person shooting games.
There are multiple distributions of Linux. While these all contain the Linux Kernel, they often include different pieces of software, and installation processes vary somewhat. The basic licensing agreement that Linux is distributed under requires that even if it is sold for a fee, it must be made available at no cost and it may also be copied freely by the user. As a result you can often download these distributions over the network at no cost . The total download size can be up to 600MB. An alternative is to buy a distribution (the typical cost is $30) which includes a floppy disk, a CD-ROM and a brief manual. These can be found at any store that sells software. Sometimes the distribution will have a higher cost for ‘deluxe’ versions - this more costly package often includes telephone support.
In total there are hundreds of Linux distributions. Many of these are specialized for features such as embedded systems, foreign languages, internet servers and security. The list below is for user-friendly installation and usage. The most successful of these distributions is Redhat. Some distributions, such as Mandrake, are based on the Redhat distribution, but with enhancements.
Each distribution of Linux will have a slightly different installation procedure, but they all follow the basic steps below. The total time to install Linux will between one to two hours. Users with a high level of knowledge can opt to do advanced setup, and new users will have the option of letting the system suggest installation options.
This section is a brief overview of the Linux operating system. The intention is to overview the basic components in the operating system. An administrator can manage the operating system using the graphical user interface (GUI), or using typed commands. New users often prefer to use the system using the GUI. Advanced users often prefer to use commands to administer the system, they are often faster and more reliable.
The GUI in Linux is actually two programs working together. The basic program is called X windows, and it provides basic connection to the screen, mouse, keyboard and sound card. The look-and-feel of the GUI is provided by the window manager. One simple window manager is called ‘fvwm’ and it can behave like Windows 95/98. Newer window managers include Gnome and KDE. While these both provide similar capabilities and features, most users develop personal preferences for a single window manager.
The directory and file structure of Linux is hierarchical, much like other popular operating systems. The main directory for the system is call root and is indicated with a single slash ‘/’. There are a number of subdirectories listed below that are used for storing system files, user files, temporary files and configuration files. A sample of the standard directories are shown below, and can be viewed with a file manager, or with keyboard commands. If other disks are used, such as a CDROM, or floppy disk, they are mounted under the root directory. (i.e., there are no ‘C’, ‘A’ or other drives, they are all under ‘/’)
Devices and settings can be configured under X-windows using graphical tools. Settings can also be configured with text files, but this is not necessary. Examples of settings that the user or root might want to change are:
Linux follows very strict conventions for file and directory permissions. These require that each file and directory be given specific permissions for public reading, writing and execution. Each user is given their own account with a password, so that access to the system is controlled. Only the root user can access all files and directories on the system. Other users are limited to files they own, or files that have been marked public. Typically the root user is only used for administration, and normal users use non-root accounts. This generally keeps the system safe from careless damage, and security breaches. Each user has their own home directory, normally in the ‘/home’ directory. The permissions for files and directories are set so that the user has complete control over that directory.
Security is not a significant problem for a computer that is not connected to a network, and passwords will protect it from ‘honest thieves’. When connected to a network there is potential for security problems. These problems become more serious when the computer is connected to the network 24 hours a day. General rules to keep a computer safe (this applies to non-Linux computers also) are:
Most users focus less on the Operating System, and more on the programs that it will run. The task list below includes many of the applications that would be desired by the average user. Most of the listed applications are free, with the exception of the games. Many of these packages are a standard part of Linux distributions.
Linux use is growing quickly, and this is expected to increase. Within administrative circles Linux is now a proven commodity. IBM  has made Linux a core component of their business strategy. Home depot is adopting it for point-of-sale terminals. Mexico city is replacing their current operating system with Linux to redirect money to fighting poverty. In academic circles Linux has had a long history, and in many cases it has been adopted department wide in many computer science and some electrical engineering programs.
Hugh Jack is an Assistant Professor in the Padnos School of Engineering at Grand Valley State University. He has been teaching there since 1996 in the areas of manufacturing and controls. His research areas include programmable logic controllers, process planning, robotics and rapids prototyping. He previously taught at Ryerson Polytechnic university for 3 years. He holds a Bachelors in Electrical Engineering, and Masters and Doctorate in Mechanical Engineering from the University of Western Ontario.