A database is nothing more than a set of related information. A telephone book, for example, is a database of the names, phone numbers, and addresses of all people living in a particular region. While a telephone book is certainly a ubiquitous and frequently used database, it suffers from the following:
Finding a person’s telephone number can be time-consuming, especially if the telephone book contains a large number of entries.
A telephone book is indexed only by last/first names, so finding the names of the people living at a particular address, while possible in theory, is not a practical use for this database.
From the moment the telephone book is printed, the information becomes less and less accurate as people move into or out of a region, change their telephone numbers, or move to another location within the same region.
The same drawbacks attributed to telephone books can also apply to any manual data storage system, such as patient records stored in a filing cabinet. Because of the cumbersome nature of paper databases, some of the first computer applications developed were database systems, which are computerized data storage and retrieval mechanisms. Because a database system stores data electronically rather than on paper, a database system is able to retrieve data more quickly, index data in multiple ways, and deliver up-to-the-minute information to its user community.
Early database systems managed data stored on magnetic tapes. Because there were generally far more tapes than tape readers, technicians were tasked with loading and unloading tapes as specific data was requested. Because the computers of that era had very little memory, multiple requests for the same data generally required the data to be read from the tape multiple times. While these database systems were a significant improvement over paper databases, they are a far cry from what is possible with today’s technology. (Modern database systems can manage terabytes of data spread across many fast-access disk drives, holding tens of gigabytes of that data in high-speed memory, but I’m getting a bit ahead of myself.)
This section contains some background information about pre-relational database systems. For those readers eager to dive into SQL, feel free to skip ahead a couple of pages to the next section.
Over the first several decades of computerized database systems, data was stored and represented to users in various ways. In a hierarchical database system, for example, data is represented as one or more tree structures. Figure 1-1 shows how data relating to George Blake’s and Sue Smith’s bank accounts might be represented via tree structures.
George and Sue each have their own tree containing their accounts and the transactions on those accounts. The hierarchical database system provides tools for locating a particular customer’s tree and then traversing the tree to find the desired accounts and/or transactions. Each node in the tree may have either zero or one parent and zero, one, or many children. This configuration is known as a single-parent hierarchy.
Another common approach, called the network database system, exposes sets of records and sets of links that define relationships between different records. Figure 1-2 shows how George’s and Sue’s same accounts might look in such a system.
In order to find the transactions posted to Sue’s money market account, you would need to perform the following steps:
Find the customer record for Sue Smith.
Follow the link from Sue Smith’s customer record to her list of accounts.
Traverse the chain of accounts until you find the money market account.
Follow the link from the money market record to its list of transactions.
One interesting feature of network database systems is demonstrated by the set
product records on the far right of Figure 1-2. Notice that each
product record (Checking, Savings, etc.) points to
a list of
account records that are of that
Account records, therefore, can
be accessed from multiple places (both
customer records and
records), allowing a network database to act as a multiparent
Both hierarchical and network database systems are alive and well today, although generally in the mainframe world. Additionally, hierarchical database systems have enjoyed a rebirth in the directory services realm, such as Microsoft’s Active Directory and the Red Hat Directory Server, as well as with Extensible Markup Language (XML). Beginning in the 1970s, however, a new way of representing data began to take root, one that was more rigorous yet easy to understand and implement.
In 1970, Dr. E. F. Codd of IBM’s research laboratory published a paper titled “A Relational Model of Data for Large Shared Data Banks” that proposed that data be represented as sets of tables. Rather than using pointers to navigate between related entities, redundant data is used to link records in different tables. Figure 1-3 shows how George’s and Sue’s account information would appear in this context.
There are four tables in Figure 1-3
representing the four entities discussed so far:
transaction. Looking across the top of the
customer table in Figure 1-3, you can see three
(which contains the customer’s ID number),
fname (which contains the customer’s first name), and
lname (which contains the customer’s last name).
Looking down the side of the
you can see two rows, one containing George Blake’s data
and the other containing Sue Smith’s data. The number of columns that a table
may contain differs from server to server, but it is generally large enough not
to be an issue (Microsoft SQL Server, for example, allows up to 1,024 columns
per table). The number of rows that a table may contain is more a matter of
physical limits (i.e., how much disk drive space is available) and
maintainability (i.e., how large a table can get before it becomes difficult to
work with) than of database server limitations.
Each table in a relational database includes information that uniquely
identifies a row in that table (known as the primary key),
along with additional information needed to describe the entity completely.
Looking again at the
customer table, the
cust_id column holds a different number
for each customer; George Blake, for example, can be uniquely identified by
customer ID #1. No other customer will ever be assigned that identifier, and no
other information is needed to locate George Blake’s data in the
Every database server provides a mechanism for generating unique sets of numbers to use as primary key values, so you won’t need to worry about keeping track of what numbers have been assigned.
While I might have chosen to use the combination of the
columns as the primary key (a primary key consisting of two or more columns is
known as a compound key), there could easily be two or more
people with the same first and last names that have accounts at the bank.
Therefore, I chose to include the
column in the
customer table specifically for
use as a primary key column.
In this example, choosing
lname as the primary
key would be referred to as a
natural key, whereas the choice of
cust_id would be referred to as a
surrogate key. The decision whether to employ
natural or surrogate keys is a topic of widespread debate, but in this
particular case the choice is clear, since a person’s last name may change
(such as when a person adopts a spouse’s last name), and primary key columns
should never be allowed to change once a value has been assigned.
Some of the tables also include information used to navigate to another table;
this is where the “redundant data” mentioned earlier comes in. For example, the
account table includes a column called
cust_id, which contains the unique
identifier of the customer who opened the account, along with a column called
product_cd, which contains the unique
identifier of the product to which the account will conform. These columns are
known as foreign keys, and they serve the same purpose as
the lines that connect the entities in the hierarchical and network versions of
the account information. If you are looking at a particular account record and
want to know more information about the customer who opened the account, you
would take the value of the
and use it to find the appropriate row in the
customer table (this process is known, in relational database
lingo, as a join; joins are introduced in Chapter 3 and probed deeply in Chapters 5 and
It might seem wasteful to store the same data many times, but the relational
model is quite clear on what redundant data may be stored. For example, it is
proper for the
account table to include a
column for the unique identifier of the customer who opened the account, but it
is not proper to include the customer’s first and last names in the
account table as well. If a customer were to
change her name, for example, you want to make sure that there is only one place
in the database that holds the customer’s name; otherwise, the data might be
changed in one place but not another, causing the data in the database to be
unreliable. The proper place for this data is the
customer table, and only the
cust_id values should be included in other tables. It is also not
proper for a single column to contain multiple pieces of information, such as a
name column that contains both a person’s
first and last names, or an
that contains street, city, state, and zip code information. The process of
refining a database design to ensure that each independent piece of information
is in only one place (except for foreign keys) is known as
Getting back to the four tables in Figure 1-3, you may wonder how you would
use these tables to find George Blake’s transactions against his checking
account. First, you would find George Blake’s unique identifier in the
customer table. Then, you would find the row in
account table whose
cust_id column contains George’s unique identifier
product_cd column matches the row
product table whose
name column equals “Checking.” Finally, you would
locate the rows in the
transaction table whose
account_id column matches the unique identifier from the
account table. This might sound complicated, but
you can do it in a single command, using the SQL language, as you will see
I introduced some new terminology in the previous sections, so maybe it’s time for some formal definitions. Table 1-1 shows the terms we use for the remainder of the book along with their definitions.
Table 1-1. Terms and definitions
Something of interest to the database user community. Examples include customers, parts, geographic locations, etc.
An individual piece of data stored in a table.
A set of columns that together completely describe an entity or some action on an entity. Also called a record.
A set of rows, held either in memory (nonpersistent) or on permanent storage (persistent).
Another name for a nonpersistent table, generally the result of an SQL query.
One or more columns that can be used as a unique identifier for each row in a table.
Along with Codd’s definition of the relational model, he proposed a language called DSL/Alpha for manipulating the data in relational tables. Shortly after Codd’s paper was released, IBM commissioned a group to build a prototype based on Codd’s ideas. This group created a simplified version of DSL/Alpha that they called SQUARE. Refinements to SQUARE led to a language called SEQUEL, which was, finally, renamed SQL.
SQL is now entering middle age (as is this author, alas), and it has undergone a great deal of change along the way. In the mid-1980s, the American National Standards Institute (ANSI) began working on the first standard for the SQL language, which was published in 1986. Subsequent refinements led to new releases of the SQL standard in 1989, 1992, 1999, 2003, and 2006. Along with refinements to the core language, new features have been added to the SQL language to incorporate object-oriented functionality, among other things. The latest standard, SQL:2006, focuses on the integration of SQL and XML and defines a language called XQuery which is used to query data in XML documents.
SQL goes hand in hand with the relational model because the result of an SQL query is a table (also called, in this context, a result set). Thus, a new permanent table can be created in a relational database simply by storing the result set of a query. Similarly, a query can use both permanent tables and the result sets from other queries as inputs (we explore this in detail in Chapter 9).
One final note: SQL is not an acronym for anything (although many people will insist it stands for “Structured Query Language”). When referring to the language, it is equally acceptable to say the letters individually (i.e., S. Q. L.) or to use the word sequel.
The SQL language is divided into several distinct parts: the parts that we
explore in this book include SQL schema statements, which
are used to define the data structures stored in the database; SQL
data statements, which are used to manipulate the data structures
previously defined using SQL schema statements; and SQL transaction
statements, which are used to begin, end, and roll back
transactions (covered in Chapter 12). For example, to create
a new table in your database, you would use the SQL schema statement
create table, whereas the process of populating
your new table with data would require the SQL data statement
To give you a taste of what these statements look like, here’s an SQL schema
statement that creates a table called
CREATE TABLE corporation (corp_id SMALLINT, name VARCHAR(30), CONSTRAINT pk_corporation PRIMARY KEY (corp_id) );
This statement creates a table with two columns,
name, with the
corp_id column identified as the primary
key for the table. We probe the finer details of this statement, such as the
different data types available with MySQL, in Chapter 2. Next, here’s an SQL data
statement that inserts a row into the
corporation table for Acme Paper Corporation:
INSERT INTO corporation (corp_id, name) VALUES (27, 'Acme Paper Corporation');
This statement adds a row to the
corporation table with a value of
27 for the
corp_id column and
a value of
Acme Paper Corporation for the
Finally, here’s a simple
to retrieve the data that was just created:
WHERE corp_id = 27;+------------------------+ | name | +------------------------+ | Acme Paper Corporation | +------------------------+
All database elements created via SQL schema statements are stored in a
special set of tables called the data dictionary. This
“data about the database” is known collectively as metadata
and is explored in Chapter 15. Just like tables that you create
yourself, data dictionary tables can be queried via a
select statement, thereby allowing you to discover the current
data structures deployed in the database at runtime. For example, if you are
asked to write a report showing the new accounts created last month, you could
either hardcode the names of the columns in the
account table that were known to you when you wrote the report,
or query the data dictionary to determine the current set of columns and
dynamically generate the report each time it is executed.
Most of this book is concerned with the data portion of the SQL language,
which consists of the
delete commands. SQL schema statements is
demonstrated in Chapter 2, where the
sample database used throughout this book is generated. In general, SQL schema
statements do not require much discussion apart from their syntax, whereas SQL
data statements, while few in number, offer numerous opportunities for detailed
study. Therefore, while I try to introduce you to many of the SQL schema
statements, most chapters in this book concentrate on the SQL data statements.
If you have worked with programming languages in the past, you are used to defining variables and data structures, using conditional logic (i.e., if-then-else) and looping constructs (i.e., do while ... end), and breaking your code into small, reusable pieces (i.e., objects, functions, procedures). Your code is handed to a compiler, and the executable that results does exactly (well, not always exactly) what you programmed it to do. Whether you work with Java, C#, C, Visual Basic, or some other procedural language, you are in complete control of what the program does.
A procedural language defines both the desired results and the mechanism, or process, by which the results are generated. Nonprocedural languages also define the desired results, but the process by which the results are generated is left to an external agent.
With SQL, however, you will need to give up some of the control you are used to, because SQL statements define the necessary inputs and outputs, but the manner in which a statement is executed is left to a component of your database engine known as the optimizer. The optimizer’s job is to look at your SQL statements and, taking into account how your tables are configured and what indexes are available, decide the most efficient execution path (well, not always the most efficient). Most database engines will allow you to influence the optimizer’s decisions by specifying optimizer hints, such as suggesting that a particular index be used; most SQL users, however, will never get to this level of sophistication and will leave such tweaking to their database administrator or performance expert.
With SQL, therefore, you will not be able to write complete applications. Unless you are writing a simple script to manipulate certain data, you will need to integrate SQL with your favorite programming language. Some database vendors have done this for you, such as Oracle’s PL/SQL language, MySQL’s stored procedure language, and Microsoft’s Transact-SQL language. With these languages, the SQL data statements are part of the language’s grammar, allowing you to seamlessly integrate database queries with procedural commands. If you are using a non-database-specific language such as Java, however, you will need to use a toolkit/API to execute SQL statements from your code. Some of these toolkits are provided by your database vendor, whereas others are created by third-party vendors or by open source providers. Table 1-2 shows some of the available options for integrating SQL into a specific language.
Table 1-2. SQL integration toolkits
JDBC (Java Database Connectivity; JavaSoft)
Rogue Wave SourcePro DB (third-party tool to connect to Oracle, SQL Server, MySQL, Informix, DB2, Sybase, and PostgreSQL databases)
Pro*C (Oracle), MySQL C API (open source), and DB2 Call Level Interface (IBM)
If you only need to execute SQL commands interactively, every database vendor
provides at least a simple command-line tool for submitting SQL commands to the
database engine and inspecting the results. Most vendors provide a graphical
tool as well that includes one window showing your SQL commands and another
window showing the results from your SQL commands. Since the examples in this
book are executed against a MySQL database, I use the
mysql command-line tool that is included as part of the MySQL
installation to run the examples and format the results.
SELECT t.txn_id, t.txn_type_cd, t.txn_date, t.amount FROM individual i INNER JOIN account a ON i.cust_id = a.cust_id INNER JOIN product p ON p.product_cd = a.product_cd INNER JOIN transaction t ON t.account_id = a.account_id WHERE i.fname = 'George' AND i.lname = 'Blake' AND p.name = 'checking account'; +--------+-------------+---------------------+--------+ | txn_id | txn_type_cd | txn_date | amount | +--------+-------------+---------------------+--------+ | 11 | DBT | 2008-01-05 00:00:00 | 100.00 | +--------+-------------+---------------------+--------+ 1 row in set (0.00 sec)
Without going into too much detail at this point, this query identifies the
row in the
individual table for George Blake
and the row in the
product table for the
“checking” product, finds the row in the
account table for this individual/product combination, and
returns four columns from the
table for all transactions posted to this account. If you happen to know that
George Blake’s customer ID is 8 and that checking accounts are designated by the
'CHK', then you can simply find George
Blake’s checking account in the
based on the customer ID and use the account ID to find the appropriate
SELECT t.txn_id, t.txn_type_cd, t.txn_date, t.amount FROM account a INNER JOIN transaction t ON t.account_id = a.account_id WHERE a.cust_id = 8 AND a.product_cd = 'CHK';
I cover all of the concepts in these queries (plus a lot more) in the following chapters, but I wanted to at least show what they would look like.
The previous queries contain three different clauses:
where. Almost every
query that you encounter will include at least these three clauses, although
there are several more that can be used for more specialized purposes. The role
of each of these three clauses is demonstrated by the following:
SELECT /* one or more things */ ... FROM /* one or more places */ ... WHERE /* one or more conditions apply */ ...
Most SQL implementations treat any text between the
tags as comments.
When constructing your query, your first task is generally to determine which
table or tables will be needed and then add them to your
from clause. Next, you will need to add conditions
where clause to filter out the data
from these tables that you aren’t interested in. Finally, you will decide which
columns from the different tables need to be retrieved and add them to your
select clause. Here’s a simple example
that shows how you would find all customers with the last name
SELECT cust_id, fname FROM individual WHERE lname = 'Smith';
This query searches the
for all rows whose
lname column matches the
'Smith' and returns the
fname columns from those rows.
Along with querying your database, you will most likely be involved with
populating and modifying the data in your database. Here’s a simple example of
how you would insert a new row into the
INSERT INTO product (product_cd, name) VALUES ('CD', 'Certificate of Depysit')
UPDATE product SET name = 'Certificate of Deposit' WHERE product_cd = 'CD';
Notice that the
update statement also
where clause, just like the
select statement. This is because an
update statement must identify the rows
to be modified; in this case, you are specifying that only those rows whose
product_cd column matches the string
'CD' should be modified. Since the
product_cd column is the primary key for
product table, you should expect your
update statement to modify exactly one
row (or zero, if the value doesn’t exist in the table). Whenever you execute an
SQL data statement, you will receive feedback from the database engine as to how
many rows were affected by your statement. If you are using an interactive tool
such as the
mysql command-line tool mentioned
earlier, then you will receive feedback concerning how many rows were
Returned by your
Created by your
Modified by your
Removed by your
If you are using a procedural language with one of the toolkits mentioned
earlier, the toolkit will include a call to ask for this information after your
SQL data statement has executed. In general, it’s a good idea to check this info
to make sure your statement didn’t do something unexpected (like when you forget
to put a
where clause on your
delete statement and delete every row in the
Oracle Database from Oracle Corporation
SQL Server from Microsoft
DB2 Universal Database from IBM
Sybase Adaptive Server from Sybase
All these database servers do approximately the same thing, although some are better equipped to run very large or very-high-throughput databases. Others are better at handling objects or very large files or XML documents, and so on. Additionally, all these servers do a pretty good job of complying with the latest ANSI SQL standard. This is a good thing, and I make it a point to show you how to write SQL statements that will run on any of these platforms with little or no modification.
Along with the commercial database servers, there has been quite a bit of activity
in the open source community in the past five years with the goal of creating a
viable alternative to the commercial database servers. Two of the most commonly used
open source database servers are PostgreSQL and MySQL. The MySQL website (http://www.mysql.com) currently claims over 10 million installations,
its server is available for free, and I have found its server to be extremely simple
to download and install. For these reasons, I have decided that all examples for
this book be run against a MySQL (version 6.0) database, and that the
mysql command-line tool be used to format query
results. Even if you are already using another server and never plan to use MySQL, I
urge you to install the latest MySQL server, load the sample schema and data, and
experiment with the data and examples in this book.
However, keep in mind the following caveat:
This is not a book about MySQL’s SQL implementation.
Rather, this book is designed to teach you how to craft SQL statements that will run on MySQL with no modifications, and will run on recent releases of Oracle Database, Sybase Adaptive Server, and SQL Server with few or no modifications.
To keep the code in this book as vendor-independent as possible, I will refrain from demonstrating some of the interesting things that the MySQL SQL language implementers have decided to do that can’t be done on other database implementations. Instead, Appendix B covers some of these features for readers who are planning to continue using MySQL.
The overall goal of the next four chapters is to introduce the SQL data
statements, with a special emphasis on the three main clauses of the
select statement. Additionally, you will see many
examples that use the bank schema (introduced in the next chapter), which will be
used for all examples in the book. It is my hope that familiarity with a single
database will allow you to get to the crux of an example without your having to stop
and examine the tables being used each time. If it becomes a bit tedious working
with the same set of tables, feel free to augment the sample database with
additional tables, or invent your own database with which to experiment.
After you have a solid grasp on the basics, the remaining chapters will drill deep into additional concepts, most of which are independent of each other. Thus, if you find yourself getting confused, you can always move ahead and come back later to revisit a chapter. When you have finished the book and worked through all of the examples, you will be well on your way to becoming a seasoned SQL practitioner.
For readers interested in learning more about relational databases, the history of computerized database systems, or the SQL language than was covered in this short introduction, here are a few resources worth checking out:
C.J. Date’s Database in Depth: Relational Theory for Practitioners (O’Reilly)
C.J. Date’s An Introduction to Database Systems, Eighth Edition (Addison-Wesley)
C.J. Date’s The Database Relational Model: A Retrospective Review and Analysis: A Historical Account and Assessment of E. F. Codd’s Contribution to the Field of Database Technology (Addison-Wesley)