Graeme Birchall (until Version 9.7, 2011)
Version 1.0, Rodney Krick, November 2019
Contributors: Andres Gomez Casanova, Robert Mala
Db2 SQL Cookbook by Graeme Birchall is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Based on a work at https://github.com/rodneykrick/db2-sql-cookbook.
Permissions beyond the scope of this license may be available at https://github.com/rodneykrick/db2-sql-cookbook
Introduction
Project in GitHub: https://github.com/rodneykrick/db2-sql-cookbook The PDF version can be found here: http://db2-sql-cookbook.org/pdf/Db2_SQL_Cookbook.pdf.
This version of the book is intended to be a community effort to help people learn SQL. Join, improve, enjoy, have fun! == Document history
(The history of the original book can bee seen at Graeme Birchall Book Editions Upload Dates)
Version | Date | Content |
---|---|---|
1.0 |
9.10.2019 |
Started new project in GitHub. Text transformed to asciidoc. |
1.1 |
24.11.2019 |
Added license terms |
1.2 |
06.12.2019 |
Update functions to Db2 version 11.1 |
1.3 |
18.01.2020 |
Added Time Travel chapter. |
1.4 |
30.08.2022 |
Personal notes removed from introduction. |
1.5 |
06.10.2022 |
VARCHAR_FORMAT explation corrected. New contributor added. |
1. Quick find
1.1. Index of Concepts
1.1.1. Join rows
To combine matching rows in multiple tables, use a join (see Joins).
EMP_NM
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
50 |
Hanes |
EMP_JB
ID | JOB |
---|---|
10 |
Sales |
20 |
Clerk |
SELECT nm.id
, nm.name
, jb.job
FROM emp_nm nm
, emp_jb jb
WHERE nm.id = jb.id
ORDER BY 1;
ANSWER
ID | NAME | JOB |
---|---|---|
10 |
Sanders |
Sales |
20 |
Pernal |
Clerk |
1.1.2. Outer Join
To get all of the rows from one table, plus the matching rows from another table (if there are any), use an outer join (see Join Types).
EMP_NM
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
50 |
Hanes |
EMP_JB
ID | JOB |
---|---|
10 |
Sales |
20 |
Clerk |
SELECT nm.id ,nm.name ,jb.job
FROM emp_nm nm
LEFT OUTER JOIN emp_jb jb
ON nm.id = jb.id
ORDER BY nm.id;
ANSWER
ID | NAME | JOB |
---|---|---|
10 |
Sanders |
Sales |
20 |
Pernal |
Clerk |
50 |
Hanes |
- |
To get rows from either side of the join, regardless of whether they match (the join) or not, use a full outer join (see Full Outer Joins).
1.1.3. Null Values & Replace
Use the COALESCE function (see COALESCE) to replace a null value (e.g. generated in an outer join) with a non-null value.
1.1.4. Select Where No Match
To get the set of the matching rows from one table where something is true or false in another table (e.g. no corresponding row), use a sub-query (see Sub-Query).
EMP_NM
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
50 |
Hanes |
EMP_JB
ID | JOB |
---|---|
10 |
Sales |
20 |
Clerk |
SELECT * FROM emp_nm nm
WHERE NOT EXISTS
(SELECT * FROM emp_jb jb
WHERE nm.id = jb.id)
ORDER BY id;
ANSWER
ID | NAME |
---|---|
50 |
Hanes |
1.1.5. Append Rows
To add (append) one set of rows to another set of rows, use a union (see Union, Intersect, and Except).
EMP_NM
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
50 |
Hanes |
EMP_JB
ID | JOB |
---|---|
10 |
Sales |
20 |
Clerk |
SELECT *
FROM emp_nm
WHERE emp_nm name < 'S'
UNION
SELECT *
FROM emp_jb
ORDER BY 1, 2;
ANSWER
ID | 2 |
---|---|
10 |
Sales |
20 |
Clerk |
20 |
Pernal |
50 |
Hanes |
1.1.6. Assign Output Numbers
To assign line numbers to SQL output, use the ROW_NUMBER function (see ROW_NUMBER).
EMP_JB
ID | JOB |
---|---|
10 |
Sales |
20 |
Clerk |
SELECT id
, job
, ROW_NUMBER() OVER(ORDER BY job) AS R
FROM emp_jb
ORDER BY job;
ANSWER
ID | JOB | R |
---|---|---|
20 |
Clerk |
1 |
10 |
Sales |
2 |
1.1.7. Assign Unique Key Numbers
To make each row inserted into a table automatically get a unique key value, use an identity column, or a sequence, when creating the table (see Identity Columns and Sequences).
1.1.8. If-Then-Else Logic
To include if-then-else logical constructs in SQL stmts, use the CASE phrase (see CASE Expression).
EMP_JB
ID | JOB |
---|---|
10 |
Sales |
20 |
Clerk |
SELECT id
, job
, CASE
WHEN job = 'Sales' THEN 'Fire'
ELSE 'Demote'
END AS STATUS
FROM emp_jb;
ANSWER
ID | JOB | STATUS |
---|---|---|
10 |
Sales |
Fire |
20 |
Clerk |
Demote |
1.1.9. Get Dependents
To get all of the dependents of some object, regardless of the degree of separation from the parent to the child, use recursion (see Recursive SQL).
FAMILY
PARNT | CHILD |
---|---|
GrDad |
Dad |
Dad |
Dghtr |
Dghtr |
GrSon |
Dghtr |
GrDtr |
WITH temp (persn, lvl)
AS (SELECT parnt, 1
FROM family
WHERE parnt = 'Dad'
UNION ALL
SELECT child, Lvl + 1
FROM temp, family
WHERE persn = parnt)
SELECT * FROM temp;
ANSWER
PERSN | LVL |
---|---|
Dad |
1 |
Dghtr |
2 |
GrSon |
3 |
GrDtr |
3 |
1.1.10. Convert String to Rows
To convert a (potentially large) set of values in a string (character field) into separate rows (e.g. one row per word), use recursion (see Recursive SQL).
INPUT DATA "Some silly text" Use Recursive SQL
ANSWER
TEXT | LINE# |
---|---|
Some |
1 |
silly |
2 |
text |
3 |
Be warned - in many cases, the code is not pretty.
1.1.11. Convert Rows to String
To convert a (potentially large) set of values that are in multiple rows into a single combined field, use recursion (see Recursive SQL).
INPUT DATA
TEXT | LINE# |
---|---|
Some |
1 |
silly |
2 |
text |
3 |
Use Recursive SQL
ANSWER: "Some silly text"
1.1.12. Fetch First "n" Rows
To fetch the first "n" matching rows, use the FETCH FIRST notation (see FETCH FIRST Clause).
EMP_NM
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
50 |
Hanes |
SELECT * FROM
emp_nm
ORDER BY id DESC
FETCH FIRST 2 ROWS ONLY;
ANSWER
ID | NAME |
---|---|
50 |
Hanes |
20 |
Pernal |
Another way to do the same thing is to assign row numbers to the output, and then fetch those rows where the row-number is less than "n" (see Selecting "n" Rows).
1.1.13. Fetch Subsequent "n" Rows
To the fetch the "n" through "n + m" rows, first use the ROW_NUMBER function to assign output numbers, then put the result in a nested-table-expression, and then fetch the rows with desired numbers.
1.1.14. Fetch Uncommitted Data
To retrieve data that may have been changed by another user, but which they have yet to commit, use the WITH UR (Uncommitted Read) notation.
EMP_NM
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
50 |
Hanes |
SELECT *
FROM emp_nm
WHERE name like 'S%'
WITH UR;
ANSWER
ID | NAME |
---|---|
10 |
Sanders |
Using this option can result in one fetching data that is subsequently rolled back, and so was never valid. Use with extreme care.
1.1.15. Summarize Column Contents
Use a column function (see Column Functions or Aggregate Functions) to summarize the contents of a column.
EMP_NM
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
50 |
Hanes |
SELECT AVG(id)AS avg
,MAX(name) AS maxn
,COUNT(*) AS #rows
FROM emp_nm;
ANSWER
AVG | MAXN | #ROWS |
---|---|---|
26 |
Sanders |
3 |
1.1.16. Subtotals and Grand Totals
To obtain subtotals and grand-totals, use the ROLLUP or CUBE statements (see ROLLUP Statement).
SELECT job
,dept
,SUM(salary) AS sum_sal
,COUNT(*) AS #emps
FROM staff
WHERE dept < 30
AND salary < 90000
AND job < 'S'
GROUP BY ROLLUP(job, dept)
ORDER BY job, dept;
ANSWER
JOB | DEPT | SUM_SAL | #EMPS |
---|---|---|---|
Clerk |
15 |
84766.70 |
2 |
Clerk |
20 |
77757.35 |
2 |
Clerk |
- |
162524.05 |
4 |
Mgr |
10 |
243453.45 |
3 |
Mgr |
15 |
80659.80 |
1 |
Mgr |
- |
324113.25 |
4 |
- |
- |
486637.30 |
8 |
1.1.17. Enforcing Data Integrity
When a table is created, various Db2 features can be used to ensure that the data entered in the table is always correct:
-
Uniqueness (of values) can be enforced by creating unique indexes.
-
Check constraints can be defined to limit the values that a column can have.
-
Default values (for a column) can be defined - to be used when no value is provided.
-
Identity columns (see Identity Columns and Sequences), can be defined to automatically generate unique numeric values (e.g. invoice numbers) for all of the rows in a table. Sequences can do the same thing over multiple tables.
-
Referential integrity rules can be created to enforce key relationships between tables.
-
Triggers can be defined to enforce more complex integrity rules, and also to do things (e.g. populate an audit trail) whenever data is changed.
See the Db2 manuals for documentation or Protecting Your Data for more information about the above.
1.1.18. Hide Complex SQL
One can create a view (see View) to hide complex SQL that is run repetitively. Be warned however that doing so can make it significantly harder to tune the SQL - because some of the logic will be in the user code, and some in the view definition.
1.2. Summary Table
Some queries that use a GROUP BY can be made to run much faster by defining a summary table (see Materialized Query Tables) that Db2 automatically maintains. Subsequently, when the user writes the original GROUP BY against the source-data table, the optimizer substitutes with a much simpler (and faster) query against the summary table.
2. Introduction to SQL
This chapter contains a basic introduction to Db2 SQL. It also has numerous examples illustrating how to use this language to answer particular business problems. However, it is not meant to be a definitive guide to the language. Please refer to the relevant IBM manuals for a more detailed description.
2.1. Syntax Diagram
The original book has lots of syntax diagrams in it. We decided not to put them here anymore. Syntax diagrams can be found in the SQL Reference of Db2 (or your prefered RDBMS). We only give you lots of examples how to use SQL. When you master this stuff, you should learn more and improve your knowledge using the "real" manuals.
2.2. SQL Comments
A comment in a SQL statement starts with two dashes and goes to the end of the line:
SELECT name
FROM staff
ORDER BY id;
-- this is a comment.
-- this is another comment.
Some Db2 command processors (e.g. db2batch on the PC, or SPUFI on the mainframe) can process intelligent comments. These begin the line with a --#SET
phrase, and then identify the value to be set. In the following example, the statement delimiter is changed using an intelligent comment:
--#SET DELIMITER !
SELECT name
FROM staff
WHERE id = 10!
--#SET DELIMITER ;
SELECT name
FROM staff
WHERE id = 20;
When using the Db2 Command Processor (batch) script, the default statement terminator can be set using the "-tdx" option, where "x" is the value have chosen.
See the section titled Special Character Usage for notes on how to refer to the statement delimiter in the SQL text. |
2.3. Statement Delimiter
Db2 SQL does not come with a designated statement delimiter (terminator), though a semicolon is often used. A semi-colon cannot be used when writing a compound SQL statement (see Compound SQL) because that character is used to terminate the various subcomponents of the statement.
2.4. SQL Components
2.4.1. Db2 Objects
Db2 is a relational database that supports a variety of object types. In this section we shall overview those items which one can obtain data from using SQL.
Table
A table is an organized set of columns and rows. The number, type, and relative position, of the various columns in the table is recorded in the Db2 catalogue. The number of rows in the table will fluctuate as data is inserted and deleted. The CREATE TABLE statement is used to define a table. The following example will define the EMPLOYEE table, which is found in the Db2 sample database.
CREATE TABLE employee
( empno CHARACTER(6) NOT NULL
, firstnme VARCHAR(12) NOT NULL
, midinit CHARACTER(1) NOT NULL
, lastname VARCHAR(15) NOT NULL
, workdept CHARACTER(3)
, phoneno CHARACTER(4)
, hiredate DATE
, job CHARACTER(8)
, edlevel SMALLINT NOT NULL
, sex CHARACTER(1)
, birthdate DATE
, salary DECIMAL(9,02)
, bonus DECIMAL(9,02)
, comm DECIMAL(9,02)
);
View
A view is another way to look at the data in one or more tables (or other views). For example, a user of the following view will only see those rows (and certain columns) in the EMPLOYEE table where the salary of a particular employee is greater than or equal to the average salary for their particular department.
CREATE VIEW employee_view AS
SELECT a.empno
, a.firstnme
, a.salary
, a.workdept
FROM employee a
WHERE a.salary >=
(SELECT AVG(b.salary)
FROM employee b
WHERE a.workdept = b.workdept
);
A view need not always refer to an actual table. It may instead contain a list of values:
CREATE VIEW silly (c1, c2, c3)
AS VALUES
(11, 'AAA', SMALLINT(22))
,(12, 'BBB', SMALLINT(33))
,(13, 'CCC', NULL);
Selecting from the above view works the same as selecting from a table:
SELECT c1, c2, c3
FROM silly
ORDER BY c1 ASC;
ANSWER
C1 | C2 | C3 |
---|---|---|
11 |
AAA |
22 |
12 |
BBB |
33 |
13 |
CCC |
- |
We can go one step further and define a view that begins with a single value that is then manipulated using SQL to make many other values. For example, the following view, when selected from, will return 10,000 rows. Note however that these rows are not stored anywhere in the database - they are instead created on the fly when the view is queried.
CREATE VIEW test_data AS
WITH temp1 (num1) AS
(VALUES (1)
UNION ALL
SELECT num1 + 1
FROM temp1
WHERE num1 < 10000)
SELECT *
FROM temp1;
Alias
An alias is an alternate name for a table or a view. Unlike a view, an alias can not contain any processing logic. No authorization is required to use an alias other than that needed to access to the underlying table or view.
CREATE ALIAS employee_al1 FOR employee;
COMMIT;
CREATE ALIAS employee_al2 FOR employee_al1;
COMMIT;
CREATE ALIAS employee_al3 FOR employee_al2;
COMMIT;
Neither a view, nor an alias, can be linked in a recursive manner (e.g. V1 points to V2, which points back to V1). Also, both views and aliases still exist after a source object (e.g. a table) has been dropped. In such cases, a view, but not an alias, is marked invalid.
Nickname
A nickname is the name that one provides to Db2 for either a remote table, or a non-relational object that one wants to query as if it were a table.
CREATE NICKNAME emp FOR unixserver.production.employee;
Tablesample
Use of the optional TABLESAMPLE reference enables one to randomly select (sample) some fraction of the rows in the underlying base table:
SELECT *
FROM staff
TABLESAMPLE BERNOULLI(10);
See Randomly Sample Data for information on using the TABLESAMPLE feature.
2.4.2. Db2 Data Types
Db2 comes with the following standard data types:
-
SMALLINT, INT, and BIGINT (i.e. integer numbers).
-
FLOAT, REAL, and DOUBLE (i.e. floating point numbers).
-
DECIMAL and NUMERIC (i.e. decimal numbers).
-
DECFLOAT (i.e. decimal floating-point numbers).
-
CHAR, VARCHAR, and LONG VARCHAR (i.e. character values).
-
GRAPHIC, VARGRAPHIC, and LONG VARGRAPHIC (i.e. graphical values).
-
BLOB, CLOB, and DBCLOB (i.e. binary and character long object values).
-
DATE, TIME, and TIMESTAMP (i.e. date/time values).
-
DATALINK (i.e. link to external object).
-
XML (i.e. contains well formed XML data).
Below is a simple table definition that uses some of the above data types:
CREATE TABLE sales_record
(sales# INTEGER NOT NULL
GENERATED ALWAYS AS IDENTITY (START WITH 1, INCREMENT BY 1)
, sale_ts TIMESTAMP NOT NULL
, num_items SMALLINT NOT NULL
, payment_type CHAR(2) NOT NULL
, sale_value DECIMAL(12,2) NOT NULL
, sales_tax DECIMAL(12,2)
, employee# INTEGER NOT NULL
, CONSTRAINT sales1 CHECK (payment_type IN ('CS','CR'))
, CONSTRAINT sales2 CHECK (sale_value > 0)
, CONSTRAINT sales3 CHECK (num_items > 0)
, CONSTRAINT sales4 FOREIGN KEY (employee#)
REFERENCES staff (id) ON DELETE RESTRICT
, PRIMARY KEY (sales#)
);
In the above table, we have listed the relevant columns, and added various checks to ensure that the data is always correct. In particular, we have included the following:
-
The sales# is automatically generated (see Identity Columns and Sequences for details). It is also the primary key of the table, and so must always be unique.
-
The payment-type must be one of two possible values.
-
Both the sales-value and the num-items must be greater than zero.
-
The employee# must already exist in the staff table. Furthermore, once a row has been inserted into this table, any attempt to delete the related row from the staff table will fail.
Default Lengths
The following table has two columns:
CREATE TABLE default_values
(c1 CHAR NOT NULL
,d1 DECIMAL NOT NULL);
The length has not been provided for either of the above columns. In this case, Db2 defaults to CHAR(1) for the first column and DECIMAL(5,0) for the second column.
Data Type Usage
In general, use the standard Db2 data types as follows:
-
Always store monetary data in a decimal field.
-
Store non-fractional numbers in one of the integer field types.
-
Use floating-point when absolute precision is not necessary.
A Db2 data type is not just a place to hold data. It also defines what rules are applied when the data in manipulated. For example, storing monetary data in a Db2 floating-point field is a no-no, in part because the data-type is not precise, but also because a floating-point number is not manipulated (e.g. during division) according to internationally accepted accounting rules.
DECFLOAT Arithmetic
DECFLOAT numbers have quite different processing characteristics from the other number types. For a start, they support more values:
-
Zero.
-
Negative and positive numbers (e.g. -1234.56).
-
Negative and positive infinity.
-
Negative and positive NaN (i.e. Not a Number).
-
Negative and positive sNaN (i.e. signaling Not a Number).
NaN Usage
The value NaN represents the result of an arithmetic operation that does not return a number (e.g. the square root of a negative number), but is also not infinity. For example, the expression 0/0 returns NaN, while 1/0 returns infinity.
The value NaN propagates through any arithmetic expression. Thus the final result is always either positive or negative NaN, as the following query illustrates:
SELECT DECFLOAT(+1.23) + NaN AS " NaN"
, DECFLOAT(-1.23) + NaN AS " NaN"
, DECFLOAT(-1.23) + -NaN AS " -NaN"
, DECFLOAT(+infinity) + NaN AS " NaN"
, DECFLOAT(+sNaN) + NaN AS " NaN"
, DECFLOAT(-sNaN) + NaN AS " -NaN"
, DECFLOAT(+NaN) + NaN AS " NaN"
, DECFLOAT(-NaN) + NaN AS " -NaN"
FROM sysibm.sysdummy1;
Any reference to a signaling NaN value in a statement (as above) will result in a warning message being generated. |
Infinity Usage
The value infinity works similar to NaN. Its reference in an arithmetic expression almost always returns either positive or negative infinity (assuming NaN is not also present). The one exception is division by infinity, which returns a really small, but still finite, number:
SELECT DECFLOAT(1) / +infinity AS " 0E-6176"
, DECFLOAT(1) * +infinity AS " Infinity"
, DECFLOAT(1) + +infinity AS " Infinity"
, DECFLOAT(1) - +infinity AS "-Infinity"
, DECFLOAT(1) / -infinity AS " -0E-6176"
, DECFLOAT(1) * -infinity AS "-Infinity"
, DECFLOAT(1) + -infinity AS "-Infinity"
, DECFLOAT(1) - -infinity AS " Infinity"
FROM sysibm.sysdummy1;
The next query shows some situations where either infinity or NaN is returned:
SELECT DECFLOAT(+1.23) / 0 AS " Infinity"
, DECFLOAT(-1.23) / 0 AS "-Infinity"
, DECFLOAT(+1.23) + infinity AS " Infinity"
, DECFLOAT(0) / 0 AS "NaN"
, DECFLOAT(infinity) + -infinity AS "NaN"
, LOG(DECFLOAT(0)) AS "-Infinity"
, LOG(DECFLOAT(-123)) AS "NaN"
, SQRT(DECFLOAT(-123)) AS "NaN"
FROM sysibm.sysdummy1;
DECFLOAT Value Order
The DECFLOAT values have the following order, from low to high:
-NaN -sNan -infinity -1.2 -1.20 0 1.20 1.2 infinity sNaN NaN
Please note that the numbers 1.2 and 1.200 are "equal", but they will be stored as different values, and will have a different value order. The TOTALORDER function can be used to illustrate this. It returns one of three values:
-
Zero if the two values have the same order.
-
+1 if the first value has a higher order (even if it is equal).
-
-1 if the first value has a lower order (even if it is equal).
WITH temp1 (d1, d2) AS
(VALUES (DECFLOAT(+1.0), DECFLOAT(+1.00))
,(DECFLOAT(-1.0), DECFLOAT(-1.00))
,(DECFLOAT(+0.0), DECFLOAT(+0.00))
,(DECFLOAT(-0.0), DECFLOAT(-0.00))
,(DECFLOAT(+0), DECFLOAT(-0)) )
SELECT TOTALORDER(d1,d2) AS TOTALORDER
FROM temp1;
ANSWER
TOTALORDER |
---|
1 |
-1 |
1 |
1 |
0 |
The NORMALIZE_DECFLOAT scalar function can be used to strip trailing zeros from a DECFLOAT value:
WITH temp1 (d1) AS
(VALUES (DECFLOAT(+0 ,16))
,(DECFLOAT(+0.0 ,16))
,(DECFLOAT(+0.00 ,16))
,(DECFLOAT(+0.000 ,16))
)
SELECT d1
, HEX(d1) AS hex_d1
, NORMALIZE_DECFLOAT(d1) AS d2
, HEX(NORMALIZE_DECFLOAT(d1)) AS hex_d2
FROM temp1;
ANSWER
D1 | HEX_D1 | D2 | HEX_D2 |
---|---|---|---|
0 |
0000000000003822 |
0 |
0000000000003822 |
0.0 |
0000000000003422 |
0 |
0000000000003822 |
0.00 |
0000000000003022 |
0 |
0000000000003822 |
0.000 |
0000000000002C22 |
0 |
0000000000003822 |
DECFLOAT Scalar Functions
The following scalar functions support the DECFLOAT data type:
-
COMPARE_DECFLOAT: Compares order of two DECFLOAT values.
-
DECFLOAT: Converts input value to DECFLOAT.
-
NORMALIZE_DECFLOAT: Removes trailing blanks from DECFLOAT value.
-
QUANTIZE: Converts number to DECFLOAT, using mask to define precision.
-
TOTALORDER: Compares order of two DECFLOAT values.
Date/Time Arithmetic
Manipulating date/time values can sometimes give unexpected results. What follows is a brief introduction to the subject. The basic rules are:
-
Multiplication and division is not allowed.
-
Subtraction is allowed using date/time values, date/time durations, or labeled durations.
-
Addition is allowed using date/time durations, or labeled durations.
The valid labeled durations are listed below:
LABELED DURATIONS | WORKS WITH DATE/TIME | ||||
---|---|---|---|---|---|
SINGULAR |
PLURAL |
ITEM FIXED SIZE |
DATE |
TIME |
TIMESTAMP |
YEAR |
YEARS |
N |
Y |
- |
Y |
MONTH |
MONTHS |
N |
Y |
- |
Y |
DAY |
DAYS |
Y |
Y |
- |
Y |
HOUR |
HOURS |
Y |
- |
Y |
Y |
MINUTE |
MINUTES |
Y |
- |
Y |
Y |
SECOND |
SECONDS |
Y |
- |
Y |
Y |
MICROSECOND |
MICROSECONDS |
Y |
- |
Y |
Y |
Usage Notes
-
It doesn’t matter if one uses singular or plural. One can add "4 day" to a date.
-
Some months and years are longer than others. So when one adds "2 months" to a date the result is determined, in part, by the date that you began with. More on this below.
-
One cannot add "minutes" to a date, or "days" to a time, etc.
-
One cannot combine labeled durations in parenthesis: "date - (1 day + 2 months)" will fail. One should instead say: "date - 1 day - 2 months".
-
Adding too many hours, minutes or seconds to a time will cause it to wrap around. The overflow will be lost.
-
Adding 24 hours to the time '00.00.00' will get '24.00.00'. Adding 24 hours to any other time will return the original value.
-
When a decimal value is used (e.g. 4.5 days) the fractional part is discarded. So to add (to a timestamp value) 4.5 days, add 4 days and 12 hours.
Now for some examples:
SELECT sales_date
, sales_date - 10 DAY AS d1
, sales_date + -1 MONTH AS d2
, sales_date + 99 YEARS AS d3
, sales_date + 55 DAYS
- 22 MONTHS AS d4
, sales_date + (4+6) DAYS AS d5
FROM sales
WHERE sales_person = 'GOUNOT'
AND sales_date = '1995-12-31';
ANSWER
sales_date | d1 | d2 | d3 | d4 | d5 |
---|---|---|---|---|---|
1995-12-31 |
1995-12-21 |
1995-11-30 |
2094-12-31 |
1994-04-24 |
1996-01-10 |
Adding or subtracting months or years can give somewhat odd results when the month of the beginning date is longer than the month of the ending date. For example, adding 1 month to '2004-01-31' gives '2004-02-29', which is not the same as adding 31 days, and is not the same result that one will get in 2005. Likewise, adding 1 month, and then a second 1 month to '2004-01-31' gives '2004-03-29', which is not the same as adding 2 months. Below are some examples of this issue:
SELECT sales_date
, sales_date + 2 MONTH AS d1
, sales_date + 3 MONTHS AS d2
, sales_date + 2 MONTH + 1 MONTH AS d3
, sales_date + (2+1) MONTHS AS d4
FROM sales
WHERE sales_person = 'GOUNOT'
AND sales_date = '1995-12-31';
ANSWER
sales_date | d1 | d2 | d3 | d4 |
---|---|---|---|---|
1995-12-31 |
1996-02-29 |
1996-03-31 |
1996-03-29 |
1996-03-31 |
Date/Time Duration Usage
When one date/time value is subtracted from another date/time value the result is a date, time,or timestamp duration. This decimal value expresses the difference thus:
DURATION-TYPE | FORMAT | NUMBER-REPRESENTS | USE-WITH-D-TYPE |
---|---|---|---|
DATE |
DECIMAL(8,0) |
yyyymmdd |
TIMESTAMP, DATE |
TIME |
DECIMAL(6,0) |
hhmmss |
TIMESTAMP, TIME |
TIMESTAMP |
DECIMAL(20,6) |
yyyymmddhhmmss.zzzzzz |
TIMESTAMP |
Below is an example of date duration generation:
SELECT empno
, hiredate
, birthdate
, hiredate - birthdate
FROM employee
WHERE workdept = 'D11'
AND lastname < 'L'
ORDER BY empno;
ANSWER
EMPNO | HIREDATE | BIRTHDATE | - |
---|---|---|---|
000150 |
1972-02-12 |
1947-05-17 |
240826 |
000200 |
1966-03-03 |
1941-05-29 |
240905 |
000210 |
1979-04-11 |
1953-02-23 |
260116 |
A date/time duration can be added to or subtracted from a date/time value, but it does not make for very pretty code:
SELECT hiredate
, hiredate - 12345678
, hiredate - 1234 years
- 56 months
- 78 days
FROM employee
WHERE empno = '000150';
ANSWER
HIREDATE | - | - |
---|---|---|
1972-02-12 |
0733-03-26 |
0733-03-26 |
Date/Time Subtraction
One date/time can be subtracted (only) from another valid date/time value. The result is a date/time duration value. Date Duration Generation above has an example.
Db2 Special Registers
A special register is a Db2 variable that contains information about the state of the system. The complete list follows:
Special Register | Updatable | Data type |
---|---|---|
CURRENT CLIENT_ACCTNG |
no |
VARCHAR(255) |
CURRENT CLIENT_APPLNAME |
no |
VARCHAR(255) |
CURRENT CLIENT_USERID |
no |
VARCHAR(255) |
CURRENT CLIENT_WRKSTNNAME |
no |
VARCHAR(255) |
CURRENT DATE |
no |
DATE |
CURRENT DBPARTITIONNUM |
no |
INTEGER |
CURRENT DECFLOAT ROUNDING MODE |
no |
VARCHAR(128) |
CURRENT DEFAULT TRANSFORM GROUP |
yes |
VARCHAR(18) |
CURRENT DEGREE |
yes |
CHAR(5) |
CURRENT EXPLAIN MODE |
yes |
VARCHAR(254) |
CURRENT EXPLAIN SNAPSHOT |
yes |
CHAR(8) |
CURRENT FEDERATED ASYNCHRONY |
yes |
INTEGER |
CURRENT IMPLICIT XMLPARSE OPTION |
yes |
VARCHAR(19) |
CURRENT ISOLATION |
yes |
CHAR(2) |
CURRENT LOCK TIMEOUT |
yes |
INTEGER |
CURRENT MAINTAINED TABLE TYPES FOR OPTIMIZATION |
yes |
VARCHAR(254) |
CURRENT MDC ROLLOUT MODE |
yes |
VARCHAR(9) |
CURRENT OPTIMIZATION PROFILE |
yes |
VARCHAR(261) |
CURRENT PACKAGE PATH |
yes |
VARCHAR(4096) |
CURRENT PATH |
yes |
VARCHAR(2048) |
CURRENT QUERY OPTIMIZATION |
yes |
INTEGER |
CURRENT REFRESH AGE |
yes |
DECIMAL(20,6) |
CURRENT SCHEMA |
yes |
VARCHAR(128) |
CURRENT SERVER |
no |
VARCHAR(128) |
CURRENT TIME |
no |
TIME |
CURRENT TIMESTAMP |
no |
TIMESTAMP |
CURRENT TIMEZONE |
no |
DECIMAL(6,0) |
CURRENT USER |
no |
VARCHAR(128) |
SESSION_USER |
yes |
VARCHAR(128) |
SYSTEM_USER |
no |
VARCHAR(128) |
USER |
yes |
VARCHAR(128) |
Usage Notes
-
Some special registers can be referenced using an underscore instead of a blank in the name - as in: CURRENT_DATE.
-
Some special registers can be updated using the SET command (see list above).
-
All special registers can be queried using the SET command. They can also be referenced in ordinary SQL statements.
-
Those special registers that automatically change over time (e.g. current timestamp) are always the same for the duration of a given SQL statement. So if one inserts a thousand rows in a single insert, all will get the same current timestamp.
-
One can reference the current timestamp in an insert or update, to record in the target table when the row was changed. To see the value assigned, query the DML statement. See Select DML Changes for details.
Refer to the Db2 SQL Reference Volume 1 for a detailed description of each register.
Sample SQL
SET CURRENT ISOLATION = RR;
SET CURRENT SCHEMA = 'ABC';
SELECT CURRENT TIME AS cur_TIME
, CURRENT ISOLATION AS cur_ISO
, CURRENT SCHEMA AS cur_ID
FROM sysibm.sysdummy1;
ANSWER
CUR_TIME | CUR_ISO | CUR_ID |
---|---|---|
12:15:16 |
RR |
ABC |
Distinct Types
A distinct data type is a field type that is derived from one of the base Db2 field types. It is used when one wants to prevent users from combining two separate columns that should never be manipulated together (e.g. adding US dollars to Japanese Yen).
The following source type do not support distinct types: XML, Array. |
The creation of a distinct type, under the covers, results in the creation of two implied functions that can be used to convert data to and from the source type and the distinct type. Support for the basic comparison operators ( =, <>, <, < =, >, and > =
) is also provided. Below is a typical create and drop statement:
CREATE DISTINCT TYPE JAP_YEN AS DECIMAL(15,2) WITH COMPARISONS;
DROP DISTINCT TYPE JAP_YEN;
A distinct type cannot be dropped if it is currently being used in a table. |
Usage Example
Imagine that we had the following customer table:
CREATE TABLE customer
( id INTEGER NOT NULL
, fname VARCHAR(10) NOT NULL WITH DEFAULT ''
, lname VARCHAR(15) NOT NULL WITH DEFAULT ''
, date_of_birth DATE
, citizenship CHAR(3) NOT NULL WITH DEFAULT ''
, usa_sales DECIMAL(9,2)
, eur_sales DECIMAL(9,2)
, sales_office# SMALLINT
, last_updated TIMESTAMP
, PRIMARY KEY(id));
One problem with the above table is that the user can add the American and European sales values, which if they are expressed in dollars and euros respectively, is silly:
SELECT id
, usa_sales + eur_sales AS tot_sales
FROM customer;
To prevent the above, we can create two distinct types:
CREATE DISTINCT TYPE USA_DOLLARS AS DECIMAL(9,2) WITH COMPARISONS;
CREATE DISTINCT TYPE EUROS AS DECIMAL(9,2) WITH COMPARISONS;
Now we can define the customer table thus:
CREATE TABLE customer
( id INTEGER NOT NULL
, fname VARCHAR(10) NOT NULL WITH DEFAULT ''
, lname VARCHAR(15) NOT NULL WITH DEFAULT ''
, date_of_birth DATE
, citizenship CHAR(3)
, usa_sales USA_DOLLARS
, eur_sales EUROS
, sales_office# SMALLINT
, last_updated TIMESTAMP
, PRIMARY KEY(id));
Now, when we attempt to run the following, it will fail:
SELECT id
, usa_sales + eur_sales AS tot_sales
FROM customer;
The creation of a distinct type, under the covers, results in the creation two implied functions that can be used to convert data to and from the source type and the distinct type. In the next example, the two monetary values are converted to their common decimal source type, and then added together:
SELECT id
, DECIMAL(usa_sales) + DECIMAL(eur_sales) AS tot_sales
FROM customer;
Fullselect, Subselect, & Common Table Expression
It is not the purpose of this book to give you detailed description of SQL terminology, but there are a few words that you should know. For example, the following diagram illustrates the various components of a query:
WITH get_matching_rows AS
(SELECT id
, name
, salary
FROM staff
WHERE id < 50
UNION ALL
SELECT id
, name
, salary
FROM staff
WHERE id = 100
)
SELECT *
FROM get_matching_rows
ORDER BY id
FETCH FIRST 10 ROWS ONLY
FOR FETCH ONLY
WITH UR;
-
The structure from WITH until the last parenthesis is a COMMON TABLE EXPRESSION.
-
Each select block is called a SUBSELECT.
-
The block inside the WITH with two SUBSELECTS is a FULLSELECT.
2.4.3. Query Components
-
SUBSELECT: A query that selects zero or more rows from one or more tables.
-
FULLSELECT: One or more subselects or VALUES clauses, connected using a UNION, INTERSECT, or EXCEPT, all enclosed in parenthesis.
-
COMMON TABLE EXPRESSION: A named fullselect that can be referenced one more times in another subselect. See Common Table Expression for a more complete definition.
2.5. SELECT Statement
A SELECT statement is used to query the database. It has the following components, not all of which need be used in any particular query:
-
SELECT clause. One of these is required, and it must return at least one item, be it a column, a literal, the result of a function, or something else. One must also access at least one table, be that a true table, a temporary table, a view, an alias or a table function.
-
WITH clause. This clause is optional. Use this phrase to include independent SELECT statements that are subsequently accessed in a final SELECT (see Common Table Expression).
-
ORDER BY clause. Optionally, order the final output (see Order By, Group By, and Having).
-
FETCH FIRST clause. Optionally, stop the query after "n" rows (see FETCH FIRST Clause). If an optimize-for value is also provided, both values are used independently by the optimizer.
-
READ-ONLY clause. Optionally, state that the query is read-only. Some queries are inherently read-only, in which case this option has no effect.
-
FOR UPDATE clause. Optionally, state that the query will be used to update certain columns that are returned during fetch processing.
-
OPTIMIZE FOR n ROWS clause. Optionally, tell the optimizer to tune the query assuming that not all of the matching rows will be retrieved. If a first-fetch value is also provided, both values are used independently by the optimizer.
Refer to the IBM manuals for a complete description of all of the above. Some of the more interesting options are described below.
SELECT Clause
Every query must have at least one SELECT statement, and it must return at least one item, and access at least one object.
2.5.1. SELECT Items
-
Column: A column in one of the table being selected from.
-
Literal: A literal value (e.g. "ABC"). Use the AS expression to name the literal.
-
Special Register: A special register (e.g. CURRENT TIME).
-
Expression: An expression result (e.g. MAX(COL1*10)).
-
Full Select: An embedded SELECT statement that returns a single row.
2.5.2. FROM Objects
-
Table: Either a permanent or temporary Db2 table.
-
View: A standard Db2 view.
-
Alias: A Db2 alias that points to a table, view, or another alias.
-
Full Select: An embedded SELECT statement that returns a set of rows.
-
Table function: A kind of function that returns a table.
Sample SQL
SELECT deptno
, admrdept
, 'ABC' AS abc
FROM department
WHERE deptname LIKE '%ING%'
ORDER BY 1;
ANSWER
DEPTNO | ADMRDEPT | ABC |
---|---|---|
B01 |
A00 |
ABC |
D11 |
D01 |
ABC |
To select all of the columns in a table (or tables) one can use the "*" notation:
SELECT *
FROM department
WHERE deptname LIKE '%ING%'
ORDER BY 1;
ANSWER (part of)
DEPTNO | etc… |
---|---|
B01 |
PLANNING |
D11 |
MANUFACTU |
To select both individual columns, and all of the columns (using the "" notation), in a single SELECT statement, one can still use the "", but it must fully-qualified using either the object name, or a correlation name:
SELECT deptno
, department.*
FROM department
WHERE deptname LIKE '%ING%'
ORDER BY 1;
ANSWER (part of)
DEPTNO | DEPTNO | etc… |
---|---|---|
B01 |
B01 |
PLANNING |
D11 |
D11 |
MANUFACTU |
Use the following notation to select all the fields in a table twice:
SELECT department.*
, department.*
FROM department
WHERE eptname LIKE '%NING%'
ORDER BY 1;
ANSWER (part of)
DEPTNO | etc… | … | DEPTNO | etc… | … |
---|---|---|---|---|---|
B01 |
PLANNING |
… |
B01 |
PLANNING |
… |
D11 |
MANUFACTU |
… |
D11 |
MANUFACTU |
… |
2.5.3. FETCH FIRST Clause
The fetch first clause limits the cursor to retrieving "n" rows. If the clause is specified and no number is provided, the query will stop after the first fetch. If this clause is used, and there is no ORDER BY, then the query will simply return a random set of matching rows, where the randomness is a function of the access path used and/or the physical location of the rows in the table:
SELECT years
, name
, id
FROM staff
FETCH FIRST 3 ROWS ONLY;
ANSWER
YEARS | NAME | ID |
---|---|---|
7 |
Sanders |
10 |
8 |
Pernal |
20 |
5 |
Marenghi |
30 |
Using the FETCH FIRST clause to get the first "n" rows can sometimes return an answer that is not what the user really intended. See below for details. |
If an ORDER BY is provided, then the FETCH FIRST clause can be used to stop the query after a certain number of what are, perhaps, the most desirable rows have been returned. However, the phrase should only be used in this manner when the related ORDER BY uniquely identifies each row returned. To illustrate what can go wrong, imagine that we wanted to query the STAFF table in order to get the names of those three employees that have worked for the firm the longest - in order to give them a little reward (or possibly to fire them). The following query could be run:
SELECT years
, name
, id
FROM staff
WHERE years IS NOT NULL
ORDER BY years DESC
FETCH FIRST 3 ROWS ONLY;
ANSWER
YEARS | NAME | ID |
---|---|---|
13 |
Graham |
310 |
12 |
Jones |
260 |
10 |
Hanes |
50 |
The above query answers the question correctly, but the question was wrong, and so the answer is wrong. The problem is that there are two employees that have worked for the firm for ten years, but only one of them shows, and the one that does show was picked at random by the query processor. This is almost certainly not what the business user intended. The next query is similar to the previous, but now the ORDER ID uniquely identifies each row returned (presumably as per the end-user’s instructions):
SELECT years
, name
, id
FROM staff
WHERE years IS NOT NULL
ORDER BY years DESC
, id DESC
FETCH FIRST 3 ROWS ONLY;
ANSWER
YEARS | NAME | ID |
---|---|---|
13 |
Graham |
310 |
12 |
Jones |
260 |
10 |
Quill |
290 |
Getting the first "n" rows from a query is actually quite a complicated problem. Refer to Selecting "n" or more Rows for a more complete discussion. |
2.5.4. Correlation Name
The correlation name is defined in the FROM clause and relates to the preceding object name. In some cases, it is used to provide a short form of the related object name. In other situations, it is required in order to uniquely identify logical tables when a single physical table is referred to twice in the same query. Some sample SQL follows:
SELECT a.empno
, a.lastname
, (SELECT MAX(empno)AS empno
FROM employee) AS b
FROM employee a
WHERE a.empno = b.empno;
ANSWER
EMPNO | LASTNAME |
---|---|
000340 |
GOUNOT |
SELECT a.empno
, a.lastname
, b.deptno AS dept
FROM employee a
, department b
WHERE a.workdept = b.deptno
AND a.job <> 'SALESREP'
AND b.deptname = 'OPERATIONS'
AND a.sex IN ('M','F')
AND b.location IS NULL
ORDER BY 1;
ANSWER
EMPNO | LASTNAME | DEPT |
---|---|---|
000090 |
HENDERSON |
E11 |
000280 |
SCHNEIDER |
E11 |
000290 |
PARKER |
E11 |
000300 |
SMITH |
E11 |
000310 |
SETRIGHT |
E11 |
2.5.5. Renaming Fields
The AS phrase can be used in a SELECT list to give a field a different name. If the new name is an invalid field name (e.g. contains embedded blanks), then place the name in quotes:
SELECT empno AS e_num
, midinit AS "m int"
, phoneno AS "..."
FROM employee
WHERE empno < '000030'
ORDER BY 1;
ANSWER
E_NUM | M INT | … |
---|---|---|
000010 |
I |
3978 |
000020 |
L |
3476 |
The new field name must not be qualified (e.g. A.C1), but need not be unique. Subsequent usage of the new name is limited as follows:
-
It can be used in an order by clause.
-
It cannot be used in other part of the select (where-clause, group-by, or having).
-
It cannot be used in an update clause.
-
It is known outside of the fullselect of nested table expressions, common table expressions, and in a view definition.
CREATE view emp2
AS SELECT empno AS e_num
, midinit AS "m int"
, phoneno AS "..."
FROM employee;
SELECT * FROM emp2 WHERE "..." = '3978';
ANSWER
E_NUM | M INT | … |
---|---|---|
000010 |
I |
3978 |
2.5.6. Working with Nulls
In SQL something can be true, false, or NULL. This three-way logic has to always be considered when accessing data. To illustrate, if we first select all the rows in the STAFF table where the SALARY is < $10,000, then all the rows where the SALARY is >= $10,000, we have not necessarily found all the rows in the table because we have yet to select those rows where the SALARY is null. The presence of null values in a table can also impact the various column functions. For example, the AVG function ignores null values when calculating the average of a set of rows. This means that a user-calculated average may give a different result from a Db2 calculated equivalent:
SELECT AVG(comm) AS a1
, SUM(comm) / COUNT(*) AS a2
FROM staff
WHERE id < 100;
ANSWER
A1 | A2 |
---|---|
796.025 |
530.68 |
Null values can also pop in columns that are defined as NOT NULL. This happens when a field is processed using a column function and there are no rows that match the search criteria:
SELECT COUNT(*) AS num
, MAX(lastname) AS max
FROM employee
WHERE firstnme = 'FRED';
ANSWER
NUM | MAX |
---|---|
0 |
- |
Why Null Exist
NULL values can represent two kinds of data. In first case, the value is unknown (e.g. we do not know the name of the person’s spouse). Alternatively, the value is not relevant to the situation (e.g. the person does not have a spouse). Many people prefer not to have to bother with nulls, so they use instead a special value when necessary (e.g. an unknown employee name is blank). This trick works OK with character data, but it can lead to problems when used on numeric values (e.g. an unknown salary is set to zero).
Locating Null Values
One can not use an equal predicate to locate those values that are null because a null value does not actually equal anything, not even null, it is simply null. The IS NULL or IS NOT NULL phrases are used instead. The following example gets the average commission of only those rows that are not null. Note that the second result differs from the first due to rounding loss.
SELECT AVG(comm) AS a1
, SUM(comm) / COUNT(*) AS a2
FROM staff
WHERE id < 100
AND comm IS NOT NULL;
ANSWER
A1 | A2 |
---|---|
796.025 |
796.02 |
2.5.7. Quotes and Double-quotes
To write a string, put it in quotes. If the string contains quotes, each quote is represented by a pair of quotes:
SELECT 'JOHN' AS J1
, 'JOHN''S' AS J2
, '''JOHN''S''' AS J3
, '"JOHN''S"' AS J4
FROM staff
WHERE id = 10;
ANSWER
J1 | J2 | J3 | J4 |
---|---|---|---|
JOHN |
JOHN’S |
'JOHN’S' |
"JOHN’S" |
Double quotes can be used to give a name to an output field that would otherwise not be valid. To put a double quote in the name, use a pair of quotes:
SELECT id AS "USER ID"
, dept AS "D#"
, years AS "#Y"
, 'ABC' AS "'TXT'"
, '"' AS """quote"" fld"
FROM staff s
WHERE id < 40
ORDER BY "USER ID";
ANSWER
USER ID | D# | #Y | 'TXT' | "quote" fld |
---|---|---|---|---|
10 |
20 |
7 |
ABC |
" |
20 |
20 |
8 |
ABC |
" |
30 |
38 |
5 |
ABC |
" |
2.6. SQL Predicates
A predicate is used in either the WHERE or HAVING clauses of a SQL statement. It specifies a condition that true, false, or unknown about a row or a group.
2.6.1. Predicate Precedence
As a rule, a query will return the same result regardless of the sequence in which the various predicates are specified. However, note the following:
-
Predicates separated by an OR may need parenthesis - see AND/OR Precedence.
-
Checks specified in a CASE statement are done in the order written - see CASE Expression.
Basic Predicate
A basic predicate compares two values. If either value is null, the result is unknown. Otherwise the result is either true or false.
SELECT id, job, dept
FROM staff
WHERE job = 'Mgr'
AND NOT job <> 'Mgr'
AND NOT job = 'Sales'
AND id <> 100
AND id >= 0
AND id <= 150
AND NOT dept = 50
ORDER BY id;
ANSWER
ID | JOB | DEPT |
---|---|---|
10 |
Mgr |
20 |
30 |
Mgr |
38 |
50 |
Mgr |
15 |
140 |
Mgr |
51 |
A variation of this predicate type can be used to compare sets of columns/values. Everything on both sides must equal in order for the expressions to match:
SELECT id, dept, job
FROM staff
WHERE (id,dept) = (30,28)
OR (id,years) = (90, 7)
OR (dept,job) = (38,'Mgr')
ORDER BY 1;
ANSWER
ID | DEPT | JOB |
---|---|---|
30 |
38 |
Mgr |
Below is the same query written the old fashioned way:
SELECT id, dept, job
FROM staff
WHERE (id = 30 AND dept = 28)
OR (id = 90 AND years = 7)
OR (dept = 38 AND job = 'Mgr')
ORDER BY 1;
ANSWER
ID | DEPT | JOB |
---|---|---|
30 |
38 |
Mgr |
2.6.2. Quantified Predicate
A quantified predicate compares one or more values with a collection of values.
SELECT id, job
FROM staff
WHERE job = ANY (SELECT job FROM staff)
AND id <= ALL (SELECT id FROM staff)
ORDER BY id;
ANSWER
ID | JOB |
---|---|
10 |
Mgr |
SELECT id, dept, job
FROM staff
WHERE (id,dept) = ANY
(SELECT dept, id
FROM staff
)
ORDER BY 1;
ANSWER
ID | DEPT | JOB |
---|---|---|
20 |
20 |
Sales |
See the sub-query chapter on Sub-Query for more data on this predicate type.
2.6.3. BETWEEN Predicate
The BETWEEN predicate compares a value within a range of values.
The between check always assumes that the first value in the expression is the low value and the second value is the high value. For example, BETWEEN 10 AND 12 may find data, but BETWEEN 12 AND 10 never will.
SELECT id, job
FROM staff
WHERE id BETWEEN 10 AND 30
AND id NOT BETWEEN 30 AND 10
AND NOT id NOT BETWEEN 10 AND 30
ORDER BY id;
ANSWER
ID | JOB |
---|---|
10 |
Mgr |
20 |
Sales |
30 |
Mgr |
2.6.4. EXISTS Predicate
An EXISTS predicate tests for the existence of matching rows.
SELECT id, job
FROM staff a
WHERE EXISTS
(SELECT *
FROM staff b
WHERE b.id = a.id
AND b.id < 50
)
ORDER BY id;
ANSWER
ID | JOB |
---|---|
10 |
Mgr |
20 |
Sales |
30 |
Mgr |
40 |
Sales |
See the sub-query chapter on Sub-Query for more data on this predicate type. |
2.6.5. IN Predicate
The IN predicate compares one or more values with a list of values.
The list of values being compared in the IN statement can either be a set of in-line expressions (e.g. ID in (10,20,30)), or a set rows returned from a sub-query. Either way, Db2 simply goes through the list until it finds a match.
SELECT id, job
FROM staff a
WHERE id IN (10,20,30)
AND id IN
(SELECT id
FROM staff
)
AND id NOT IN 99
ORDER BY id;
ANSWER
ID | JOB |
---|---|
10 |
Mgr |
20 |
Sales |
30 |
Mgr |
The IN statement can also be used to compare multiple fields against a set of rows returned from a sub-query. A match exists when all fields equal. This type of statement is especially useful when doing a search against a table with a multi-columns key.
Be careful when using the NOT IN expression against a sub-query result. If any one row in the sub-query returns null, the result will be no match. See Sub-Query for more details. |
SELECT empno, lastname
FROM employee
WHERE (empno, 'AD3113') IN
(SELECT empno, projno
FROM emp_act
WHERE emptime > 0.5
)
ORDER BY 1;
ANSWER
EMPNO | LASTNAME |
---|---|
000260 |
JOHNSON |
000270 |
PEREZ |
See the sub-query chapter on Sub-Query for more data on this statement type. |
2.6.6. LIKE Predicate
The LIKE predicate does partial checks on character strings.
The percent and underscore characters have special meanings. The first means skip a string of any length (including zero) and the second means skip one byte. For example:
-
LIKE 'AB_D%' Finds 'ABCD' and 'ABCDE', but not 'ABD', nor 'ABCCD'.
-
LIKE '_X' Finds 'XX' and 'DX', but not 'X', nor 'ABX', nor 'AXB'.
-
LIKE '%X' Finds 'AX', 'X', and 'AAX', but not 'XA'.
SELECT id
, name
FROM staff
WHERE name LIKE 'S%n'
OR name LIKE '_a_a%'
OR name LIKE '%r_%a'
ORDER BY id;
ANSWER
ID | NAME |
---|---|
130 |
Yamaguchi |
200 |
Scoutten |
The ESCAPE Phrase
The escape character in a LIKE statement enables one to check for percent signs and/or underscores in the search string. When used, it precedes the '%' or '_' in the search string indicating that it is the actual value and not the special character which is to be checked for. When processing the LIKE pattern, Db2 works thus: Any pair of escape characters is treated as the literal value (e.g. "+" means the string ""). Any single occurrence of an escape character followed by either a "%" or a "\_" means the literal "%" or "_" (e.g. "+%" means the string "%"). Any other "%" or "\_" is used as in a normal LIKE pattern.
LIKE STATEMENT TEXT | WHAT VALUES MATCH |
---|---|
LIKE 'AB%' |
Finds AB, any string |
LIKE 'AB%' ESCAPE '+' |
Finds AB, any string |
LIKE 'AB+%' ESCAPE '+' |
Finds AB% |
LIKE 'AB+' ESCAPE '' |
Finds AB+ |
LIKE 'AB+%%' ESCAPE '+' |
Finds AB%, any string |
LIKE 'AB+%' ESCAPE '' |
Finds AB+, any string |
LIKE 'AB++%' ESCAPE '' |
Finds AB+% |
LIKE 'AB++%%' ESCAPE '' |
Finds AB+%, any string |
LIKE 'AB+%%%' ESCAPE '' |
Finds AB%%, any string |
LIKE 'AB' ESCAPE '+' |
Finds AB++ |
LIKE 'AB%' ESCAPE '' |
Finds AB++% |
LIKE 'AB%' ESCAPE '+' |
Finds AB++, any string |
LIKE 'AB+%+%' ESCAPE '' |
Finds AB%+, any string |
Now for sample SQL:
SELECT id
FROM staff
WHERE id = 10
AND 'ABC' LIKE 'AB%'
AND 'A%C' LIKE 'A/%C' ESCAPE '/'
AND 'A_C' LIKE 'A\_C' ESCAPE '\'
AND 'A_$' LIKE 'A$_$$' ESCAPE '$';
ANSWER
ID |
---|
10 |
2.6.7. LIKE_COLUMN Function
The LIKE predicate cannot be used to compare one column against another. One may need to do this when joining structured to unstructured data. For example, imagine that one had a list of SQL statements (in a table) and a list of view names in a second table. One might want to scan the SQL text (using a LIKE predicate) to find those statements that referenced the views. The LOCATE function can be used to do a simple equality check. The LIKE predicate allows a more sophisticated search. The following code creates a scalar function and dependent procedure that can compare one column against another (by converting both column values into input variables). The function is just a stub. It passes the two input values down to the procedure where they are compared using a LIKE predicate. If there is a match, the function returns one, else zero.
These examples use an "!" as the stmt delimiter. |
--#SET DELIMITER !
CREATE PROCEDURE LIKE_COLUMN
( IN instr1 VARCHAR(4000)
, IN instr2 VARCHAR(4000)
, OUT outval SMALLINT)
LANGUAGE SQL
CONTAINS SQL
DETERMINISTIC
NO EXTERNAL ACTION
BEGIN
SET outval =
CASE
WHEN instr1 LIKE instr2
THEN 1
ELSE 0
END;
RETURN;
END!
CREATE FUNCTION LIKE_COLUMN
( instr1 VARCHAR(4000)
, instr2 VARCHAR(4000))
RETURNS SMALLINT
LANGUAGE SQL
CONTAINS SQL
DETERMINISTIC
NO EXTERNAL ACTION
BEGIN ATOMIC
DECLARE outval SMALLINT;
CALL LIKE_COLUMN(instr1, instr2, outval);
RETURN outval;
END!
Below is an example of the above function being used to compare to the contents of one column against another:
WITH temp1 (jtest) AS
(VALUES ('_gr%')
, ('S_le%')
)
SELECT
s.id
, s.name
, s.job
, t.jtest
FROM staff s
, temp1 t
WHERE LIKE_COLUMN(s.job , t.jtest) = 1
AND s.id < 70
ORDER BY s.id;
ANSWER
ID | NAME | JOB | JTEST |
---|---|---|---|
10 |
Sanders |
Mgr |
_gr% |
20 |
Pernal |
Sales |
S_le% |
30 |
Marenghi |
Mgr |
_gr% |
40 |
O’Brien |
Sales |
S_le% |
50 |
Hanes |
Mgr |
_gr% |
60 |
Quigley |
Sales |
S_le% |
2.6.8. NULL Predicate
The NULL predicate checks for null values. The result of this predicate cannot be unknown. If the value of the expression is null, the result is true. If the value of the expression is not null, the result is false.
SELECT id, comm
FROM staff
WHERE id < 100
AND id IS NOT NULL
AND comm IS NULL
AND NOT comm IS NOT NULL
ORDER BY id;
ANSWER
ID | COMM |
---|---|
10 |
- |
30 |
- |
50 |
- |
Use the COALESCE function to convert null values into something else. |
2.6.9. Special Character Usage
To refer to a special character in a predicate, or anywhere else in a SQL statement, use the "X" notation to substitute with the ASCII hex value. For example, the following query will list all names in the STAFF table that have an "a" followed by a semicolon:
SELECT id
, name
FROM staff
WHERE name LIKE '%a' || X'3B' || '%'
ORDER BY id;
2.6.10. Precedence Rules
Expressions within parentheses are done first, then prefix operators (e.g. -1), then multiplication and division, then addition and subtraction. When two operations of equal precedence are together (e.g. 1 * 5 / 4) they are done from left to right.
Example: 555 + -22 / (12 - 3) * 66 ^ ^ ^ ^ ^ 5th 2nd 3rd 1st 4th ANSWER: 423
Be aware that the result that you get depends very much on whether you are doing integer or decimal arithmetic. Below is the above done using integer numbers:
SELECT (12 - 3) AS int1
, -22 / (12 - 3) AS int2
, -22 / (12 - 3) * 66 AS int3
, 555 + -22 / (12 - 3) * 66 AS int4
FROM sysibm.sysdummy1;
ANSWER
INT1 | INT2 | INT3 | INT4 |
---|---|---|---|
9 |
-2 |
-132 |
423 |
Db2 truncates, not rounds, when doing integer arithmetic. |
Here is the same done using decimal numbers:
SELECT (12.0 - 3) AS dec1
, -22 / (12.0 - 3) AS dec2
, -22 / (12.0 - 3) * 66 AS dec3
, 555 + -22 / (12.0 - 3) * 66 AS dec4
FROM sysibm.sysdummy1;
ANSWER
DEC1 | DEC2 | DEC3 | DEC4 |
---|---|---|---|
9.0 |
-2.4 |
-161.3 |
393.6 |
2.6.11. AND/OR Precedence
AND operations are done before OR operations. This means that one side of an OR is fully processed before the other side is begun. To illustrate:
TABLE1
col1 | col2 |
---|---|
A |
AA |
B |
BB |
C |
CC |
SELECT *
FROM table1
WHERE col1 = 'C'
AND col1 >= 'A'
OR col2 >= 'AA'
ORDER BY col1;
ANSWER
COL1 | COL2 |
---|---|
A |
AA |
B |
BB |
C |
CC |
SELECT *
FROM table1
WHERE (col1 = 'C'
AND col1 >= 'A')
OR col2 >= 'AA'
ORDER BY col1;
ANSWER
COL1 | COL2 |
---|---|
A |
AA |
B |
BB |
C |
CC |
SELECT *
FROM table1
WHERE col1 = 'C'
AND (col1 >= 'A'
OR col2 >= 'AA')
ORDER BY col1;
ANSWER
COL1 | COL2 |
---|---|
C |
CC |
The omission of necessary parenthesis surrounding OR operators is a very common mistake. The result is usually the wrong answer. One symptom of this problem is that many more rows are returned (or updated) than anticipated. |
2.6.12. Processing Sequence
The various parts of a SQL statement are always executed in a specific sequence in order to avoid semantic ambiguity:
-
FROM clause.
-
JOIN ON clause.
-
WHERE clause.
-
GROUP BY and aggregate.
-
HAVING clause.
-
SELECT list.
-
ORDER BY clause.
-
FETCH FIRST.
Observe that ON predicates (e.g. in an outer join) are always processed before any WHERE predicates (in the same join) are applied. Ignoring this processing sequence can cause what looks like an outer join to run as an inner join (see ON and WHERE Usage). Likewise, a function that is referenced in the SELECT section of a query (e.g. row-number) is applied after the set of matching rows has been identified, but before the data has been ordered.
2.7. CAST Expression
The CAST expression is used to convert one data type to another. It is similar to the various field-type functions (e.g. CHAR, SMALLINT) except that it can also handle null values and host-variable parameter markers.
2.7.1. Input vs. Output Rules
-
EXPRESSION: If the input is neither null, nor a parameter marker, the input data-type is converted to the output data-type. Truncation and/or padding with blanks occur as required. An error is generated if the conversion is illegal.
-
NULL: If the input is null, the output is a null value of the specified type.
-
PARAMETER MAKER: This option is only used in programs and need not concern us here. See the Db2 SQL Reference for details.
Examples
Use the CAST expression to convert the SALARY field from decimal to integer:
SELECT id
, salary
, CAST(salary AS INTEGER) AS sal2
FROM staff
WHERE id < 30
ORDER BY id;
ANSWER
ID | SALARY | SAL2 |
---|---|---|
10 |
98357.50 |
98357 |
20 |
78171.25 |
78171 |
Use the CAST expression to truncate the JOB field. A warning message will be generated for the second line of output because non-blank truncation is being done.
SELECT id
, job
, CAST(job AS CHAR(3)) AS job2
FROM staff
WHERE id < 30
ORDER BY id;
ANSWER
ID | JOB | JOB2 |
---|---|---|
10 |
Mgr |
Mgr |
20 |
Sales |
Sal |
Use the CAST expression to make a derived field called JUNK of type SMALLINT where all of the values are null.
SELECT id
, CAST(NULL AS SMALLINT) AS junk
FROM staff
WHERE id < 30
ORDER BY id;
ANSWER
ID | JUNK |
---|---|
10 |
- |
20 |
- |
The CAST expression can also be used in a join, where the field types being matched differ:
SELECT stf.id
, emp.empno
FROM staff stf
LEFT OUTER JOIN employee emp
ON stf.id = CAST(emp.empno AS INTEGER)
AND emp.job = 'MANAGER'
WHERE stf.id < 60
ORDER BY stf.id;
ANSWER
ID | EMPNO |
---|---|
10 |
- |
20 |
000020 |
30 |
000030 |
40 |
- |
50 |
000050 |
Of course, the same join can be written using the raw function:
SELECT stf.id
, emp.empno
FROM staff stf
LEFT OUTER JOIN employee emp
ON stf.id = INTEGER(emp.empno)
AND emp.job = 'MANAGER'
WHERE stf.id < 60
ORDER BY stf.id;
ANSWER
ID | EMPNO |
---|---|
10 |
- |
20 |
000020 |
30 |
000030 |
40 |
- |
50 |
000050 |
2.8. VALUES Statement
The VALUES clause is used to define a set of rows and columns with explicit values. The clause is commonly used in temporary tables, but can also be used in view definitions. Once defined in a table or view, the output of the VALUES clause can be grouped by, joined to, and otherwise used as if it is an ordinary table - except that it can not be updated.
Each column defined is separated from the next using a comma. Multiple rows (which may also contain multiple columns) are separated from each other using parenthesis and a comma. When multiple rows are specified, all must share a common data type. Some examples follow:
VALUES 6 <== 1 row, 1 column VALUES(6) <== 1 row, 1 column VALUES 6, 7, 8 <== 1 row, 3 columns VALUES (6), (7), (8) <== 3 rows, 1 column VALUES (6,66), (7,77), (8,NULL) <== 3 rows, 2 column
2.8.1. Sample SQL
The VALUES clause can be used by itself as a very primitive substitute for the SELECT statement. One key difference is that output columns cannot be named. But they can be ordered, and fetched, and even named externally, as the next example illustrates:
PLAIN VALUES
VALUES
(1,2)
, (2,3)
, (3,4)
ORDER BY 2 DESC;
VALUES + WITH
WITH temp (c1,c2) AS
( VALUES (1,2)
, (2,3)
, (3,4)
)
SELECT *
FROM temp
ORDER BY 2 DESC;
VALUES + SELECT
SELECT *
FROM (VALUES (1,2)
, (2,3)
, (3,4)
) temp (c1,c2)
ORDER BY 2 DESC;
ANSWER
1 | 2 |
---|---|
3 |
4 |
2 |
3 |
1 |
2 |
The VALUES clause can encapsulate several independent queries:
VALUES
(
(SELECT COUNT(*) FROM employee)
, (SELECT AVG(salary) FROM staff)
, (SELECT MAX(deptno) FROM department)
)
FOR FETCH ONLY
WITH UR;
ANSWER
1 | 2 | 3 |
---|---|---|
42 |
67932.78 |
J22 |
The next statement defines a temporary table containing two columns and three rows. The first column defaults to type integer and the second to type varchar.
WITH temp1 (col1, col2) AS
(VALUES
(0, 'AA')
, (1, 'BB')
, (2, NULL)
)
SELECT *
FROM temp1;
ANSWER
COL1 | COL2 |
---|---|
0 |
AA |
1 |
BB |
2 |
- |
If we wish to explicitly control the output field types we can define them using the appropriate function. This trick does not work if even a single value in the target column is null.
WITH temp1 (col1, col2) AS
(VALUES
(DECIMAL(0 ,3, 1), 'AA')
, (DECIMAL(1 ,3, 1), 'BB')
, (DECIMAL(2 ,3, 1), NULL)
)
SELECT *
FROM temp1;
ANSWER
COL1 | COL2 |
---|---|
0.0 |
AA |
1.0 |
BB |
2.0 |
- |
If any one of the values in the column that we wish to explicitly define has a null value, we have to use the CAST expression to set the output field type:
WITH temp1 (col1,col2) AS
(VALUES
(0, CAST('AA' AS CHAR(1)))
, (1, CAST('BB' AS CHAR(1)))
, (2, CAST(NULL AS CHAR(1)))
)
SELECT *
FROM temp1;
ANSWER
COL1 | COL2 |
---|---|
0 |
A |
1 |
B |
2 |
- |
Alternatively, we can set the output type for all of the not-null rows in the column. Db2 will then use these rows as a guide for defining the whole column:
WITH temp1 (col1,col2) AS
(VALUES
(0, CHAR('AA', 1))
, (1, CHAR('BB', 1))
, (2, NULL)
)
SELECT *
FROM temp1;
ANSWER
COL1 | COL2 |
---|---|
0 |
A |
1 |
B |
2 |
- |
2.8.2. More Sample SQL
Temporary tables, or (permanent) views, defined using the VALUES expression can be used much like a Db2 table. They can be joined, unioned, and selected from. They can not, however, be updated, or have indexes defined on them. Temporary tables can not be used in a sub-query.
WITH temp1 (col1, col2,col3) AS
(VALUES
(0, 'AA', 0.00)
, (1, 'BB', 1.11)
, (2, 'CC', 2.22)
)
, temp2 (col1b, colx) AS
(SELECT col1
, col1 + col3
FROM temp1
)
SELECT *
FROM temp2;
ANSWER
COL1B | COLX |
---|---|
0 |
0.00 |
1 |
2.11 |
2 |
4.22 |
CREATE VIEW silly (c1, c2, c3)
AS VALUES
(11, 'AAA', SMALLINT(22))
, (12, 'BBB', SMALLINT(33))
, (13, 'CCC', NULL);
COMMIT;
WITH temp1 (col1) AS
(VALUES 0
UNION ALL
SELECT col1 + 1
FROM temp1
WHERE col1 + 1 < 100
)
SELECT *
FROM temp1;
ANSWER
COL1 |
---|
0 |
1 |
2 |
3 |
etc |
All of the above examples have matched a VALUES statement up with a prior WITH expression, so as to name the generated columns. One doesn’t have to use the latter, but if you don’t, you get a table with unnamed columns, which is pretty useless:
SELECT *
FROM
(VALUES
(123, 'ABC')
, (234, 'DEF')
) AS ttt
ORDER BY 1 DESC;
ANSWER
- | - |
---|---|
234 |
DEF |
123 |
ABC |
Combine Columns
The VALUES statement can be used inside a TABLE function to combine separate columns into one. In the following example, three columns in the STAFF table are combined into a single column – with one row per item:
SELECT id
, salary AS sal
, comm AS com
, combo
, typ
FROM staff
, TABLE( VALUES(salary , 'SAL')
, (comm , 'COM')
) AS tab(combo, typ)
WHERE id < 40
ORDER BY id
, typ;
ANSWER
ID | SAL | COM | COMBO | TYP |
---|---|---|---|---|
10 |
98357.50 |
- |
COM |
|
10 |
98357.50 |
- |
98357.50 |
SAL |
20 |
78171.25 |
612.45 |
612.45 |
COM |
20 |
78171.25 |
612.45 |
78171.25 |
SAL |
30 |
77506.75 |
- |
COM |
|
30 |
77506.75 |
- |
77506.75 |
SAL |
The above query works as follows:
-
The set of matching rows are obtained from the STAFF table.
-
For each matching row, the TABLE function creates two rows, the first with the salary value, and the second with the commission.
-
Each new row as gets a second literal column – indicating the data source.
-
Finally, the "AS" expression assigns a correlation name to the table output, and also defines two column names.
The TABLE function is resolved row-by-row, with the result being joined to the current row in the STAFF table. This explains why we do not get a Cartesian product, even though no join criteria are provided.
The keyword LATERAL can be used instead of TABLE in the above query. |
2.9. CASE Expression
CASE expressions enable one to do if-then-else type processing inside of SQL statements.
The sequence of the CASE conditions can affect the answer. The first WHEN check that matches is the one used. |
2.9.1. CASE Syntax Styles
There are two general flavors of the CASE expression. In the first kind, each WHEN statement does its own independent check. In the second kind, all of the WHEN conditions do similar "equal" checks against a common reference expression.
SELECT Lastname
, sex AS sx
, CASE sex
WHEN 'F' THEN 'FEMALE'
WHEN 'M' THEN 'MALE'
ELSE NULL
END AS sexx
FROM employee
WHERE lastname LIKE 'J%'
ORDER BY 1;
ANSWER
LASTNAME | SX | SEXX |
---|---|---|
JEFFERSON |
M |
MALE |
JOHN |
F |
FEMALE |
JOHNSON |
F |
FEMALE |
JONES |
M |
MALE |
SELECT lastname
, sex AS sx
, CASE WHEN sex = 'F' THEN 'FEMALE'
WHEN sex = 'M' THEN 'MALE'
ELSE NULL
END AS sexx
FROM employee
WHERE lastname LIKE 'J%'
ORDER BY 1;
ANSWER
LASTNAME | SX | SEXX |
---|---|---|
JEFFERSON |
M |
MALE |
JOHN |
F |
FEMALE |
JOHNSON |
F |
FEMALE |
JONES |
M |
MALE |
Notes & Restrictions
-
If more than one WHEN condition is true, the first one processed that matches is used.
-
If no WHEN matches, the value in the ELSE clause applies. If no WHEN matches and there is no ELSE clause, the result is NULL.
-
There must be at least one non-null result in a CASE statement. Failing that, one of the NULL results must be inside of a CAST expression.
-
All result values must be of the same type.
-
Functions that have an external action (e.g. RAND) can not be used in the expression part of a CASE statement.
2.9.2. Sample SQL
SELECT lastname
, midinit AS mi
, sex AS sx
, CASE WHEN midinit > SEX THEN midinit
ELSE sex
END AS mx
FROM employee
WHERE lastname LIKE 'J%'
ORDER BY 1;
ANSWER
LASTNAME | MI | SX | MX |
---|---|---|---|
JEFFERSON |
J |
M |
M |
JOHN |
K |
K |
K |
JOHNSON |
P |
F |
P |
JONES |
T |
M |
T |
SELECT COUNT(*) AS tot
, SUM(CASE sex WHEN 'F' THEN 1 ELSE 0 END) AS #f
, SUM(CASE sex WHEN 'M' THEN 1 ELSE 0 END) AS #m
FROM employee
WHERE lastname LIKE 'J%';
ANSWER
TOT | #F | #M |
---|---|---|
4 |
2 |
2 |
SELECT lastname
, LENGTH(RTRIM(lastname)) AS len
, SUBSTR(lastname , 1 ,
CASE WHEN LENGTH(RTRIM(lastname)) > 6 THEN 6
ELSE LENGTH(RTRIM(lastname))
END
) AS lastnm
FROM employee
WHERE lastname LIKE 'J%'
ORDER BY 1;
ANSWER
LASTNAME | LEN | LASTNM |
---|---|---|
JEFFERSON |
9 |
JEFFER |
JOHN |
4 |
JOHN |
JOHNSON |
7 |
JOHNSO |
JONES |
5 |
JONES |
The CASE expression can also be used in an UPDATE statement to do any one of several alternative updates to a particular field in a single pass of the data:
UPDATE staff
SET comm =
CASE dept
WHEN 15 THEN comm * 1.1
WHEN 20 THEN comm * 1.2
WHEN 38 THEN
CASE
WHEN years < 5 THEN comm * 1.3
WHEN years >= 5 THEN comm * 1.4
ELSE NULL
END
ELSE comm
END
WHERE comm IS NOT NULL
AND dept < 50;
In the next example a CASE expression is used to avoid a divide-by-zero error:
WITH temp1 (c1, c2) AS
(VALUES
(88, 9),(44, 3),(22, 0),(0, 1))
SELECT c1
, c2
, CASE c2
WHEN 0 THEN NULL
ELSE c1/c2
END AS c3
FROM temp1;
ANSWER
C1 | C2 | C3 |
---|---|---|
88 |
9 |
9 |
44 |
3 |
14 |
22 |
0 |
0 |
1 |
0 |
- |
At least one of the results in a CASE expression must be a value (i.e. not null). This is so that Db2 will know what output type to make the result.
2.9.3. Problematic CASE Statements
The case WHEN checks are always processed in the order that they are found. The first one that matches is the one used. This means that the answer returned by the query can be affected by the sequence on the WHEN checks. To illustrate this, the next statement uses the SEX field (which is always either "F" or "M") to create a new field called SXX. In this particular example, the SQL works as intended.
SELECT lastname
, sex
, CASE
WHEN sex >= 'M' THEN 'MAL'
WHEN sex >= 'F' THEN 'FEM'
END AS sxx
FROM employee
WHERE lastname LIKE 'J%'
ORDER BY 1;
ANSWER
LASTNAME | SX | SXX |
---|---|---|
JEFFERSON |
M |
MAL |
JOHN |
F |
FEM |
JOHNSON |
F |
FEM |
JONES |
M |
MAL |
In the example below all of the values in SXX field are "FEM". This is not the same as what happened above, yet the only difference is in the order of the CASE checks.
SELECT lastname
, sex
, CASE
WHEN sex >= 'F' THEN 'FEM'
WHEN sex >= 'M' THEN 'MAL'
END AS sxx
FROM employee
WHERE lastname LIKE 'J%'
ORDER BY 1;
ANSWER
LASTNAME | SX | SXX |
---|---|---|
JEFFERSON |
M |
FEM |
JOHN |
F |
FEM |
JOHNSON |
F |
FEM |
JONES |
M |
FEM |
In the prior statement the two WHEN checks overlap each other in terms of the values that they include. Because the first check includes all values that also match the second, the latter never gets invoked. Note that this problem can not occur when all of the WHEN expressions are equality checks.
2.9.4. CASE in Predicate
The result of a CASE expression can be referenced in a predicate:
SELECT id
, dept
, salary
, comm
FROM staff
WHERE CASE
WHEN comm < 70 THEN 'A'
WHEN name LIKE 'W%' THEN 'B'
WHEN salary < 11000 THEN 'C'
WHEN salary < 18500 AND dept <> 33 THEN 'D'
WHEN salary < 19000 THEN 'E'
END IN ('A','C','E')
ORDER BY id;
ANSWER
ID | DEPT | SALARY | COMM |
---|---|---|---|
130 |
42 |
10505.90 |
75.60 |
270 |
66 |
18555.50 |
|
330 |
66 |
10988.00 |
55.50 |
The above query is arguably more complex than it seems at first glance, because unlike in an ordinary query, the CASE checks are applied in the sequence they are defined. So a row will only match "B" if it has not already matched "A". In order to rewrite the above query using standard AND/OR predicates, we have to reproduce the CASE processing sequence. To this end, the three predicates in the next example that look for matching rows also apply any predicates that preceded them in the CASE statement:
SELECT id
, name
, salary
, comm
FROM staff
WHERE (comm < 70)
OR (salary < 11000 AND NOT name LIKE 'W%')
OR (salary < 19000 AND NOT (name LIKE 'W%'
OR (salary < 18500 AND dept <> 33)
)
)
ORDER BY id;
ANSWER
ID | DEPT | SALARY | COMM |
---|---|---|---|
130 |
42 |
10505.90 |
75.60 |
270 |
66 |
18555.50 |
|
330 |
66 |
10988.00 |
55.50 |
2.10. Miscellaneous SQL Statements
This section will briefly discuss several miscellaneous SQL statements. See the Db2 manuals for more details.
2.10.1. Cursor
A cursor is used in an application program to retrieve and process individual rows from a result set. To use a cursor, one has to do the following:
-
DECLARE the cursor. The declare statement has the SQL text that the cursor will run. If the cursor is declared "with hold", it will remain open after a commit, otherwise it will be closed at commit time.
The declare cursor statement is not actually executed when the program is run. It simply defines the query that will be run. |
-
OPEN the cursor. This is when the contents of on any host variables referenced by the cursor (in the predicate part of the query) are transferred to Db2.
-
FETCH rows from the cursor. One does as many fetches as is needed. If no row is found, the SQLCODE from the fetch will be 100.
-
CLOSE the cursor.
Syntax Notes
-
The cursor-name must be unique with the application program.
-
The WITH HOLD phrase indicates that the cursor will remain open if the unit of work ends with a commit. The cursor will be closed if a rollback occurs.
-
The WITH RETURN phrase is used when the cursor will generate the result set returned by a stored procedure. If the cursor is open when the stored procedure ends the result set will be return either to the calling procedure, or directly to the client application.
-
The FOR phrase can either refer to a select statement, the text for which will follow, or to the name of a statement has been previously prepared.
Usage notes
-
Cursors that require a sort (e.g. to order the output) will obtain the set of matching rows at open time, and then store them in an internal temporary table. Subsequent fetches will be from the temporary table.
-
Cursors that do not require a sort are resolved as each row is fetched from the data table.
-
All references to the current date, time, and timestamp will return the same value (i.e. as of when the cursor was opened) for all fetches in a given cursor invocation.
-
One does not have to close a cursor, but one cannot reopen it until it is closed. All open cursors are automatically closed when the thread terminates, or when a rollback occurs, or when a commit is done - except if the cursor is defined "with hold".
-
One can both update and delete "where current of cursor". In both cases, the row most recently fetched is updated or deleted. An update can only be used when the cursor being referenced is declared "for update of".
Examples
DECLARE fred CURSOR FOR
WITH RETURN TO CALLER
SELECT id
, name
, salary
, comm
FROM staff
WHERE id < :id-var
AND salary > 1000
ORDER BY id ASC
FETCH FIRST 10 ROWS ONLY
OPTIMIZE FOR 10 ROWS
FOR FETCH ONLY
WITH UR
DECLARE fred CURSOR WITH HOLD FOR
SELECT name
, salary
FROM staff
WHERE id > :id-var
FOR UPDDATE OF salary, comm
OPEN fred
DO UNTIL SQLCODE = 100
FETCH fred INTO :name-var
, :salary-var
IF salary < 1000 THEN
DO
UPDATE staff
SET salary = :new-salary-var
WHERE CURRENT OF fred
END-IF
END-DO
CLOSE fred
2.10.2. Select Into
A SELECT-INTO statement is used in an application program to retrieve a single row. If more than one row matches, an error is returned. The statement text is the same as any ordinary query, except that there is an INTO section (listing the output variables) between the SELECT list and the FROM section.
Example
SELECT name
, salary
INTO :name-var
, :salary-var
FROM staff
WHERE id = :id-var
2.10.3. Prepare
The PREPARE statement is used in an application program to dynamically prepare a SQL statement for subsequent execution.
Syntax Notes
-
The statement name names the statement. If the name is already in use, it is overridden.
-
The OUTPUT descriptor will contain information about the output parameter markers.
-
The DESCRIBE statement may be used instead of this clause.
-
The INPUT descriptor will contain information about the input parameter markers.
-
The FROM phrase points to the host-variable which contains the SQL statement text.
Prepared statement can be used by the following:
STATEMENT CAN BE USED BY | STATEMENT TYPE |
---|---|
DESCRIBE |
Any statement |
DECLARE CURSOR |
Must be SELECT |
EXECUTE |
Must not be SELECT |
2.10.4. Describe
The DESCRIBE statement is typically used in an application program to get information about a prepared statement. It can also be used in the Db2 command processor (but not in Db2BATCH) to get a description of a table, or the output columns in a SQL statement:
Below are some examples of using the statement:
DESCRIBE OUTPUT SELECT * FROM staff SQLDA Information sqldaid : SQLDA sqldabc: 896 sqln: 20 sqld: 7 Column Information sqltype sqllen sqlname.data sqlname.length ----------------- ----- ------------------------- ------------- 500 SMALLINT 2 ID 2 449 VARCHAR 9 NAME 4 501 SMALLINT 2 DEPT 4 453 CHARACTER 5 JOB 3 501 SMALLINT 2 YEARS 5 485 DECIMAL 7, 2 SALARY 6 485 DECIMAL 7, 2 COMM 4
DESCRIBE TABLE staff Column name Type schema Type name Length Scale Nulls ---------------------- ------- ----------- ------ ---- ---- ID SYSIBM SMALLINT 2 0 No NAME SYSIBM VARCHAR 9 0 Yes DEPT SYSIBM SMALLINT 2 0 Yes JOB SYSIBM CHARACTER 5 0 Yes YEARS SYSIBM SMALLINT 2 0 Yes SALARY SYSIBM DECIMAL 7 2 Yes COMM SYSIBM DECIMAL 7 2 Yes
2.10.5. Execute
The EXECUTE statement is used in an application program to execute a prepared statement. The statement can not be a select.
2.10.6. Execute Immediate
The EXECUTE IMMEDIATE statement is used in an application program to prepare and execute a statement. Only certain kinds of statement (e.g. insert, update, delete, commit) can be run this way. The statement can not be a select.
2.10.7. Set Variable
The SET statement is used in an application program to set one or more program variables to values that are returned by Db2.
Examples
SET :host-var = CURRENT TIMESTAMP
SET :host-v1 = CURRENT TIME
, :host-v2 = CURRENT DEGREE
, :host-v3 = NULL
The SET statement can also be used to get the result of a select, as long as the select only returns a single row:
SET
( :hv1
, :hv2
, :hv3) =
(SELECT id
, name
, salary
FROM staff
WHERE id = :id-var)
2.10.8. Set Db2 Control Structures
In addition to setting a host-variable, one can also set various Db2 control structures:
SET CONNECTION SET CURRENT DEFAULT TRANSFORM GROUP SET CURRENT DEGREE SET CURRENT EXPLAIN MODE SET CURRENT EXPLAIN SNAPSHOT SET CURRENT ISOLATION SET CURRENT LOCK TIMEOUT SET CURRENT MAINTAINED TABLE TYPES FOR OPTIMIZATION SET CURRENT PACKAGE PATH SET CURRENT PACKAGESET SET CURRENT QUERY OPTIMIZATION SET CURRENT REFRESH AGE SET ENCRYPTION PASSWORD SET EVENT MONITOR STATE SET INTEGRITY SET PASSTHRU SET PATH SET SCHEMA SET SERVER OPTION SET SESSION AUTHORIZATION
2.11. Unit-of-Work Processing
No changes that you make are deemed to be permanent until they are committed. This section briefly lists the commands one can use to commit or rollback changes.
2.11.1. Commit
The COMMIT statement is used to commit whatever changes have been made. Locks that were taken as a result of those changes are freed. If no commit is specified, an implicit one is done when the thread terminates.
2.11.2. Savepoint
The SAVEPOINT statement is used in an application program to set a savepoint within a unit of work. Subsequently, the program can be rolled back to the savepoint, as opposed to rolling back to the start of the unit of work.
Notes
-
If the savepoint name is the same as a savepoint that already exists within the same level, it overrides the prior savepoint - unless the latter was defined a being unique, in which case an error is returned.
-
The RETAIN CURSORS phrase tells Db2 to, if possible, keep open any active cursors.
-
The RETAIN LOCKS phrase tells Db2 to retain any locks that were obtained subsequent to the savepoint. In other words, the changes are rolled back, but the locks that came with those changes remain.
Savepoint Levels
Savepoints exist within a particular savepoint level, which can be nested within another level. A new level is created whenever one of the following occurs:
-
A new unit of work starts.
-
A procedure defined with NEW SAVEPOINT LEVEL is called.
-
An atomic compound SQL statement starts.
A savepoint level ends when the process that caused its creation finishes. When a savepoint level ends, all of the savepoints created within it are released. The following rules apply to savepoint usage:
-
Savepoints can only be referenced from within the savepoint level in which they were created. Active savepoints in prior levels are not accessible.
-
The uniqueness of savepoint names is only enforced within a given savepoint level. The same name can exist in multiple active savepoint levels.
Example
Savepoints are especially useful when one has multiple SQL statements that one wants to run or rollback as a whole, without affecting other statements in the same transaction. For example, imagine that one is transferring customer funds from one account to another. Two updates will be required - and if one should fail, both should fail:
INSERT INTO transaction_audit_table;
SAVEPOINT before_updates ON ROLLBACK RETAIN CURSORS;
UPDATE savings_account
SET balance = balance - 100
WHERE cust# = 1234;
IF SQLCODE <> 0 THEN
ROLLBACK TO SAVEPOINT before_updates;
ELSE
UPDATE checking_account
SET balance = balance + 100
WHERE cust# = 1234;
IF SQLCODE <> 0 THEN
ROLLBACK TO SAVEPOINT before_updates;
END
END
COMMIT;
In the above example, if either of the update statements fail, the transaction is rolled back to the predefined savepoint. And regardless of what happens, there will still be a row inserted into the transaction-audit table.
2.11.3. Savepoints vs. Commits
Savepoints differ from commits in the following respects:
-
One cannot rollback changes that have been committed.
-
Only a commit guarantees that the changes are stored in the database. If the program subsequently fails, the data will still be there.
-
Once a commit is done, other users can see the changed data. After a savepoint, the data is still not visible to other users.
2.11.4. Release Savepoint
The RELEASE SAVEPOINT statement will remove the named savepoint. Any savepoints nested within the named savepoint are also released. Once run, the application can no longer rollback to any of the released savepoints.
2.11.5. Rollback
The ROLLBACK statement is used to rollback any database changes since the beginning of the unit of work, or since the named savepoint - if one is specified.
3. Data Manipulation Language
The chapter has a very basic introduction to the DML (Data Manipulation Language) statements. See the Db2 manuals for more details.
3.1. Insert
The INSERT statement is used to insert rows into a table, view, or fullselect. To illustrate how it is used, this section will use a copy of the EMP_ACT sample table:
CREATE TABLE emp_act_copy
( empno CHARACTER (00006) NOT NULL
, projno CHARACTER (00006) NOT NULL
, actno SMALLINT NOT NULL
, emptime DECIMAL (05,02)
, emstdate DATE
, emendate DATE);
3.1.1. Target Objects
One can insert into a table, view, nickname, or SQL expression. For views and SQL expressions, the following rules apply:
-
The list of columns selected cannot include a column function (e.g. MIN).
-
There must be no GROUP BY or HAVING acting on the select list.
-
The list of columns selected must include all those needed to insert a new row.
-
The list of columns selected cannot include one defined from a constant, expression, or a scalar function.
-
Sub-queries, and other predicates, are fine, but are ignored (see Insert into a fullselect).
-
The query cannot be a join, nor (plain) union.
-
A "union all" is permitted - as long as the underlying tables on either side of the union have check constraints such that a row being inserted is valid for one, and only one, of the tables in the union.
All bets are off if the insert is going to a table that has an INSTEAD OF trigger defined.
Usage Notes
-
One has to provide a list of the columns (to be inserted) if the set of values provided does not equal the complete set of columns in the target table, or are not in the same order as the columns are defined in the target table.
-
The columns in the INCLUDE list are not inserted. They are intended to be referenced in a SELECT statement that encompasses the INSERT (see Select DML Changes).
-
The input data can either be explicitly defined using the VALUES statement, or retrieved from some other table using a fullselect.
3.1.2. Direct Insert
To insert a single row, where all of the columns are populated, one lists the input the values in the same order as the columns are defined in the table:
INSERT INTO emp_act_copy VALUES
('100000' ,'ABC' ,10 ,1.4 ,'2003-10-22', '2003-11-24');
To insert multiple rows in one statement, separate the row values using a comma:
INSERT INTO emp_act_copy VALUES
('200000' ,'ABC' ,10 ,1.4 ,'2003-10-22', '2003-11-24')
, ('200000' ,'DEF' ,10 ,1.4 ,'2003-10-22', '2003-11-24')
, ('200000' ,'IJK' ,10 ,1.4 ,'2003-10-22', '2003-11-24');
If multiple rows are inserted in one statement, and one of them violates a unique index check, all of the rows are rejected. |
The NULL and DEFAULT keywords can be used to assign these values to columns. One can also refer to special registers, like the current date and current time:
INSERT INTO emp_act_copy VALUES
('400000' ,'ABC' ,10 ,NULL ,DEFAULT, CURRENT DATE);
To leave some columns out of the insert statement, one has to explicitly list the columns that are included. When this is done, one can refer to the columns used in any order:
INSERT INTO emp_act_copy (projno, emendate, actno, empno)
VALUES
('ABC' ,DATE(CURRENT TIMESTAMP) ,123 ,'500000');
3.1.3. Insert into Full-Select
The next statement inserts a row into a fullselect that just happens to have a predicate which, if used in a subsequent query, would not find the row inserted. The predicate has no impact on the insert itself:
INSERT INTO
(SELECT *
FROM emp_act_copy
WHERE empno < '1'
)
VALUES
('510000' ,'ABC' ,10 ,1.4 ,'2003-10-22', '2003-11-24');
One can insert rows into a view (with predicates in the definition) that are outside the bounds of the predicates. To prevent this, define the view WITH CHECK OPTION. |
3.1.4. Insert from Select
One can insert a set of rows that is the result of a query using the following notation:
INSERT INTO emp_act_copy
SELECT LTRIM(CHAR(id + 600000))
, SUBSTR(UCASE(name),1,6)
, salary / 229
, 123
, CURRENT DATE
, '2003-11-11'
FROM staff
WHERE id < 50;
In the above example, the fractional part of the SALARY value is eliminated when the data is inserted into the ACTNO field, which only supports integer values. |
If only some columns are inserted using the query, they need to be explicitly listed:
INSERT INTO emp_act_copy (empno, actno, projno)
SELECT LTRIM(CHAR(id + 700000))
, MINUTE(CURRENT TIME)
, 'DEF'
FROM staff
WHERE id < 40;
One reason why tables should always have unique indexes is to stop stupid SQL statements like the following, which will double the number of rows in the table:
INSERT INTO emp_act_copy
SELECT *
FROM emp_act_copy;
The select statement using the insert can be as complex as one likes. In the next example, it contains the union of two queries:
INSERT INTO emp_act_copy (empno, actno, projno)
SELECT LTRIM(CHAR(id + 800000))
, 77
, 'XYZ'
FROM staff
WHERE id < 40
UNION
SELECT LTRIM(CHAR(id + 900000))
, SALARY / 100
, 'DEF'
FROM staff
WHERE id < 50;
The select can also refer to a common table expression. In the following example, six values are first generated, each in a separate row. These rows are then selected during the insert:
INSERT INTO emp_act_copy (empno, actno, projno, emptime)
WITH temp1 (col1) AS
( VALUES (1),(2),(3),(4),(5),(6))
SELECT LTRIM(CHAR(col1 + 910000))
, col1
, CHAR(col1)
, col1 / 2
FROM temp1;
The next example inserts multiple rows - all with an EMPNO beginning "92". Three rows are found in the STAFF table, and all three are inserted, even though the sub-query should get upset once the first row has been inserted. This doesn’t happen because all of the matching rows in the STAFF table are retrieved and placed in a work-file before the first insert is done:
INSERT INTO emp_act_copy (empno, actno, projno)
SELECT LTRIM(CHAR(id + 920000))
, id
, 'ABC'
FROM staff
WHERE id < 40
AND NOT EXISTS
( SELECT *
FROM emp_act_copy
WHERE empno LIKE '92%'
);
3.1.5. Insert into Multiple Tables
Below are two tables that hold data for US and international customers respectively:
CREATE TABLE us_customer
( cust# INTEGER NOT NULL
, cname CHAR(10) NOT NULL
, country CHAR(03) NOT NULL
, CHECK (country = 'USA')
, PRIMARY KEY (cust#));
CREATE TABLE intl_customer
( cust# INTEGER NOT NULL
, cname CHAR(10) NOT NULL
, country CHAR(03) NOT NULL
, CHECK (country <> 'USA')
, PRIMARY KEY (cust#));
One can use a single insert statement to insert into both of the above tables because they have mutually exclusive check constraints. This means that a new row will go to one table or the other, but not both, and not neither. To do so one must refer to the two tables using a "union all" phrase - either in a view, or a query, as is shown below:
INSERT INTO
(SELECT *
FROM us_customer
UNION ALL
SELECT *
FROM intl_customer
)
VALUES
(111,'Fred','USA')
,(222,'Dave','USA')
,(333,'Juan','MEX');
The above statement will insert two rows into the table for US customers, and one row into the table for international customers.
3.2. Update
The UPDATE statement is used to change one or more columns/rows in a table, view, or fullselect. Each column that is to be updated has to specified. Here is an example:
UPDATE emp_act_copy
SET emptime = NULL
, emendate = DEFAULT
, emstdate = CURRENT DATE + 2 DAYS
, actno = ACTNO / 2
, projno = 'ABC'
WHERE empno = '100000';
Usage Notes
-
One can update rows in a table, view, or fullselect. If the object is not a table, then it must be updateable (i.e. refer to a single table, not have any column functions, etc).
-
The correlation name is optional, and is only needed if there is an expression or predicate that references another table.
-
The columns in the INCLUDE list are not updated. They are intended to be referenced in a SELECT statement that encompasses the UPDATE (see Select DML Changes).
-
The SET statement lists the columns to be updated, and the new values they will get.
-
Predicates are optional. If none are provided, all rows in the table are updated.
-
Usually, all matching rows are updated. To update some fraction of the matching rows, use a fullselect (see Use Full-Select).
Update Examples
To update all rows in a table, leave off all predicates:
UPDATE emp_act_copy
SET actno = actno / 2;
In the next example, both target columns get the same values. This happens because the result for both columns is calculated before the first column is updated:
UPDATE emp_act_copy ac1
SET actno = actno * 2
, emptime = actno * 2
WHERE empno LIKE '910%';
One can also have an update refer to the output of a select statement - as long as the result of the select is a single row:
UPDATE emp_act_copy
SET actno =
( SELECT MAX(salary) / 10
FROM staff)
WHERE empno = '200000';
The following notation lets one update multiple columns using a single select:
UPDATE emp_act_copy
SET (actno, emstdate, projno) =
( SELECT MAX(salary) / 10
, CURRENT DATE + 2 DAYS
, MIN(CHAR(id))
FROM staff
WHERE id <> 33
)
WHERE empno LIKE '600%';
Multiple rows can be updated using multiple different values, as long as there is a one-to-one relationship between the result of the select, and each row to be updated.
UPDATE emp_act_copy ac1
SET (actno, emptime) =
(SELECT ac2.actno + 1
, ac1.emptime / 2
FROM emp_act_copy ac2
WHERE ac2.empno LIKE '60%'
AND SUBSTR(ac2.empno,3) = SUBSTR(ac1.empno,3)
)
WHERE EMPNO LIKE '700%';
3.2.1. Use Full-Select
An update statement can be run against a table, a view, or a fullselect. In the next example, the table is referred to directly:
UPDATE emp_act_copy
SET emptime = 10
WHERE empno = '000010'
AND projno = 'MA2100';
Below is a logically equivalent update that pushes the predicates up into a fullselect:
UPDATE
(SELECT *
FROM emp_act_copy
WHERE empno = '000010'
AND projno = 'MA2100'
) AS ea
SET emptime = 20;
3.2.2. Update First "n" Rows
An update normally changes all matching rows. To update only the first "n" matching rows do the following:
-
In a fullselect, retrieve the first "n" rows that you want to update.
-
Update all rows returned by the fullselect.
In the next example, the first five staff with the highest salary get a nice fat commission:
UPDATE
(SELECT *
FROM staff
ORDER BY salary DESC
FETCH FIRST 5 ROWS ONLY
) AS xxx
SET comm = 10000;
The above statement may update five random rows – if there is more than one row with the ordering value. To prevent this, ensure that the list of columns provided in the ORDER BY identify each matching row uniquely. |
3.2.3. Use OLAP Function
Imagine that we want to set the employee-time for a particular row in the EMP_ACT table to the MAX time for that employee. Below is one way to do it:
UPDATE emp_act_copy ea1
SET emptime =
(SELECT MAX(emptime)
FROM emp_act_copy ea2
WHERE ea1.empno = ea2.empno
)
WHERE empno = '000010'
AND projno = 'MA2100';
The same result can be achieved by calling an OLAP function in a fullselect, and then updating the result. In next example, the MAX employee-time per employee is calculated (for each row), and placed in a new column. This column is then used to do the final update:
UPDATE
(SELECT ea1.*
, MAX(emptime) OVER(PARTITION BY empno) AS maxtime
FROM emp_act_copy ea1
) AS ea2
SET emptime = maxtime
WHERE empno = '000010'
AND projno = 'MA2100';
The above statement has the advantage of only accessing the EMP_ACT table once. If there were many rows per employee, and no suitable index (i.e. on EMPNO and EMPTIME), it would be much faster than the prior update. The next update is similar to the prior - but it does the wrong update! In this case, the scope of the OLAP function is constrained by the predicate on PROJNO, so it no longer gets the MAX time for the employee:
UPDATE emp_act_copy
SET emptime = MAX(emptime) OVER(PARTITION BY empno)
WHERE empno = '000010'
AND projno = 'MA2100';
3.2.4. Correlated and Uncorrelated Update
In the next example, regardless of the number of rows updated, the ACTNO will always come out as one. This is because the sub-query that calculates the row-number is correlated, which means that it is resolved again for each row to be updated in the "AC1" table. At most, one "AC2" row will match, so the row-number must always equal one:
UPDATE emp_act_copy ac1
SET (actno, emptime)
= (SELECT ROW_NUMBER() OVER()
, ac1.emptime / 2
FROM emp_act_copy ac2
WHERE ac2.empno LIKE '60%'
AND SUBSTR(ac2.empno,3) = SUBSTR(ac1.empno,3)
)
WHERE EMPNO LIKE '800%';
In the next example, the ACTNO will be updated to be values 1, 2, 3, etc, in order that the rows are updated. In this example, the sub-query that calculates the row-number is uncorrelated, so all of the matching rows are first resolved, and then referred to in the next, correlated, step:
UPDATE emp_act_copy ac1
SET (actno, emptime) =
(SELECT c1
, c2
FROM (SELECT ROW_NUMBER() OVER() AS c1
, actno / 100 AS c2
, empno
FROM emp_act_copy
WHERE empno LIKE '60%'
) AS ac2
WHERE SUBSTR(ac2.empno,3) = SUBSTR(ac1.empno,3)
)
WHERE empno LIKE '900%';
3.3. Delete
The DELETE statement is used to remove rows from a table, view, or fullselect. The set of rows deleted depends on the scope of the predicates used. The following example would delete a single row from the EMP_ACT sample table:
DELETE
FROM emp_act_copy
WHERE empno = '000010'
AND projno = 'MA2100'
AND actno = 10;
Usage Notes
-
One can delete rows from a table, view, or fullselect. If the object is not a table, then it must be deletable (i.e. refer to a single table, not have any column functions, etc).
-
The correlation name is optional, and is only needed if there is a predicate that references another table.
-
The columns in the INCLUDE list are not updated. They are intended to be referenced in a SELECT statement that encompasses the DELETE (see Select DML Changes).
-
Predicates are optional. If none are provided, all rows are deleted.
-
Usually, all matching rows are deleted. To delete some fraction of the matching rows, use a fullselect (see Use Full-Select).
3.3.1. Basic Delete
This statement would delete all rows in the EMP_ACT table:
DELETE
FROM emp_act_copy;
This statement would delete all the matching rows in the EMP_ACT:
DELETE
FROM emp_act_copy
WHERE empno LIKE '00%'
AND projno >= 'MA';
3.3.2. Correlated Delete
The next example deletes all the rows in the STAFF table - except those that have the highest ID in their respective department:
DELETE
FROM staff s1
WHERE id NOT IN
(SELECT MAX(id)
FROM staff s2
WHERE s1.dept = s2.dept
);
Here is another way to write the same:
DELETE
FROM staff s1
WHERE EXISTS
(SELECT *
FROM staff s2
WHERE s2.dept = s1.dept
AND s2.id > s1.id
);
The next query is logically equivalent to the prior two, but it works quite differently. It uses a fullselect and an OLAP function to get, for each row, the ID, and also the highest ID value in the current department. All rows where these two values do not match are then deleted:
DELETE
FROM
(SELECT id
, MAX(id) OVER(PARTITION BY dept) AS max_id
FROM staff
) AS ss
WHERE id <> max_id;
3.3.3. Delete First "n" Rows
A delete removes all encompassing rows. Sometimes this is not desirable - usually because an unknown, and possibly undesirably large, number rows is deleted. One can write a delete that stops after "n" rows using the following logic:
-
In a fullselect, retrieve the first "n" rows that you want to delete.
-
Delete all rows returned by the fullselect.
In the following example, those five employees with the highest salary are deleted:
DELETE
FROM
(SELECT *
FROM staff
ORDER BY salary DESC
FETCH FIRST 5 ROWS ONLY
) AS xxx;
The above statement may delete five random rows – if there is more than one row with the same salary. To prevent this, ensure that the list of columns provided in the ORDER BY identify each matching row uniquely. |
3.4. Select DML Changes
A special kind of SELECT statement (see Select DML Changes) can encompass an INSERT, UPDATE, or DELETE statement to get the before or after image of whatever rows were changed (e.g. select the list of rows deleted). This kind of SELECT can be very useful when the DML statement is internally generating a value that one needs to know (e.g. an INSERT automatically creates a new invoice number using a sequence column), or get the set of rows that were removed by a delete. All of this can be done by coding a special kind of select.
3.4.1. Table Types
-
OLD: Returns the state of the data prior to the statement being run. This is allowed for an update and a delete.
-
NEW: Returns the state of the data prior to the application of any AFTER triggers or referential constraints. Data in the table will not equal what is returned if it is subsequently changed by AFTER triggers or R.I. This is allowed for an insert and an update.
-
FINAL: Returns the final state of the data. If there AFTER triggers that alter the target table after running of the statement, an error is returned. Ditto for a view that is defined with an INSTEAD OF trigger. This is allowed for an insert and an update.
Usage Notes
-
Only one of the above tables can be listed in the FROM statement.
-
The table listed in the FROM statement cannot be given a correlation name.
-
No other table can be listed (i.e. joined to) in the FROM statement. One can reference another table in the SELECT list (see example Join result to another table), or by using a sub-query in the predicate section of the statement.
-
The SELECT statement cannot be embedded in a nested-table expression.
-
The SELECT statement cannot be embedded in an insert statement.
-
To retrieve (generated) columns that are not in the target table, list them in an INCLUDE phrase in the DML statement. This technique can be used to, for example, assign row numbers to the set of rows entered during an insert.
-
Predicates (on the select) are optional. They have no impact on the underlying DML.
-
The INPUT SEQUENCE phrase can be used in the ORDER BY to retrieve the rows in the same sequence as they were inserted. It is not valid in an update or delete.
-
The usual scalar functions, OLAP functions, and column functions, plus the GROUP BY phrase, can be applied to the output - as desired.
3.4.2. Insert Examples
The example below selects from the final result of the insert:
SELECT empno
, projno AS prj
, actno AS act
FROM FINAL TABLE
(INSERT INTO emp_act_copy
VALUES
('200000', 'ABC', 10, 1, '2003-10-22', '2003-11-24')
, ('200000', 'DEF', 10, 1, '2003-10-22', '2003-11-24')
)
ORDER BY 1,2,3;
ANSWER
EMPNO | PRJ | ACT |
---|---|---|
200000 |
ABC |
10 |
200000 |
DEF |
10 |
One way to retrieve the new rows in the order that they were inserted is to include a column in the insert statement that is a sequence number:
SELECT empno
, projno AS prj
, actno AS act
, row#
AS r#
FROM FINAL TABLE
(INSERT INTO emp_act_copy (empno, projno, actno)
INCLUDE (row# SMALLINT)
VALUES
('300000','ZZZ',999,1)
,('300000','VVV',111,2)
)
ORDER BY row#;
ANSWER
EMPNO | PRJ | ACT | R# |
---|---|---|---|
300000 |
ZZZ |
999 |
1 |
300000 |
VVV |
111 |
2 |
The next example uses the INPUT SEQUENCE phrase to select the new rows in the order that they were inserted. Row numbers are assigned to the output:
SELECT empno
, projno AS prj
, actno AS act
, ROW_NUMBER() OVER() AS r#
FROM FINAL TABLE
(INSERT INTO emp_act_copy (empno, projno, actno)
VALUES
('400000','ZZZ',999)
, ('400000','VVV',111)
)
ORDER BY INPUT SEQUENCE;
ANSWER
EMPNO | PRJ | ACT | R# |
---|---|---|---|
400000 |
ZZZ |
999 |
1 |
400000 |
VVV |
111 |
2 |
The INPUT SEQUENCE phrase only works in an insert statement. It can be listed in the ORDER BY part of the statement, but not in the SELECT part. The only way to display the row number of each row inserted is to explicitly assign row numbers. |
In the next example, the only way to know for sure what the insert has done is to select from the result. This is because the select statement (in the insert) has the following unknowns:
-
We do not, or may not, know what ID values were selected, and thus inserted.
-
The project-number is derived from the current-time special register.
-
The action-number is generated using the RAND function.
Now for the insert:
SELECT empno
,projno AS prj
,actno AS act
,ROW_NUMBER() OVER() AS r#
FROM NEW TABLE
(INSERT INTO emp_act_copy (empno, actno, projno)
SELECT LTRIM(CHAR(id + 600000))
, SECOND(CURRENT TIME)
, CHAR(SMALLINT(RAND(1) * 1000))
FROM staff
WHERE id < 40
)
ORDER BY INPUT SEQUENCE;
ANSWER
EMPNO | PRJ | ACT | R# |
---|---|---|---|
600010 |
1 |
59 |
1 |
600020 |
563 |
59 |
2 |
600030 |
193 |
59 |
3 |
3.4.3. Update Examples
The statement below updates the matching rows by a fixed amount. The select statement gets the old EMPTIME values:
SELECT empno
, projno AS prj
, emptime AS etime
FROM OLD TABLE
(UPDATE emp_act_copy
SET emptime = emptime * 2
WHERE empno = '200000')
ORDER BY projno;
ANSWER
EMPNO | PRJ | ETIME |
---|---|---|
200000 |
ABC |
1.00 |
200000 |
DEF |
1.00 |
The next statement updates the matching EMPTIME values by random amount. To find out exactly what the update did, we need to get both the old and new values. The new values are obtained by selecting from the NEW table, while the old values are obtained by including a column in the update which is set to them, and then subsequently selected:
SELECT projno AS prj
, old_t AS old_t
, emptime AS new_t
FROM NEW TABLE
(UPDATE emp_act_copy
INCLUDE (old_t DECIMAL(5,2))
SET emptime = emptime * RAND(1) * 10
, old_t = emptime
WHERE empno = '200000'
)
ORDER BY 1;
ANSWER
PRJ | OLD_T | NEW_T |
---|---|---|
ABC |
2.00 |
0.02 |
DEF |
2.00 |
11.27 |
3.4.4. Delete Examples
The following example lists the rows that were deleted:
SELECT projno AS prj
, actno AS act
FROM OLD TABLE
(DELETE
FROM emp_act_copy
WHERE empno = '300000'
)
ORDER BY 1,2;
ANSWER
PRJ | ACT |
---|---|
VVV |
111 |
ZZZ |
999 |
The next query deletes a set of rows, and assigns row-numbers (to the included field) as the rows are deleted. The subsequent query selects every second row:
SELECT empno
, projno
, actno AS act
, row# AS r#
FROM OLD TABLE
(DELETE
FROM emp_act_copy
INCLUDE (row# SMALLINT)
SET row# = ROW_NUMBER() OVER()
WHERE empno = '000260'
)
WHERE row# = row# / 2 * 2
ORDER BY 1, 2, 3;
ANSWER
EMPNO | PROJNO | ACT | R# |
---|---|---|---|
000260 |
AD3113 |
70 |
2 |
000260 |
AD3113 |
80 |
4 |
000260 |
AD3113 |
180 |
6 |
Predicates (in the select result phrase) have no impact on the range of rows changed by the underlying DML, which is determined by its own predicates. |
One cannot join the table generated by a DML statement to another table, nor include it in a nested table expression, but one can join in the SELECT phrase. The following delete illustrates this concept by joining to the EMPLOYEE table:
SELECT empno
, (SELECT lastname
FROM
(SELECT empno AS e#
, lastname
FROM employee
) AS xxx
WHERE empno = e#)
, projno AS projno
, actno AS act
FROM OLD TABLE
(DELETE
FROM emp_act_copy
WHERE empno < '0001')
ORDER BY 1, 2, 3
FETCH FIRST 5 ROWS ONLY;
ANSWER
EMPNO | LASTNAME | PROJNO | ACT |
---|---|---|---|
000010 |
HAAS |
AD3100 |
10 |
000010 |
HAAS |
MA2100 |
10 |
000010 |
HAAS |
MA2110 |
10 |
000020 |
THOMPSON |
PL2100 |
30 |
000030 |
KWAN |
IF1000 |
10 |
Observe above that the EMPNO field in the EMPLOYEE table was be renamed (before doing the join) using a nested table expression. This was necessary because one cannot join on two fields that have the same name, without using correlation names. A correlation name cannot be used on the OLD TABLE, so we had to rename the field to get around this problem.
3.5. Merge
The MERGE statement is a combination insert and update, or delete, statement on steroids. It can be used to take the data from a source table, and combine it with the data in a target table.
The qualifying rows in the source and target tables are first matched by unique key value, and then evaluated:
-
If the source row is already in the target, the latter can be either updated or deleted.
-
If the source row is not in the target, it can be inserted.
-
If desired, a SQL error can also be generated.
Usage Rules
The following rules apply to the merge statement:
-
Correlation names are optional, but are required if the field names are not unique.
-
If the target of the merge is a fullselect or a view, it must allow updates, inserts, and deletes - as if it were an ordinary table.
-
At least one ON condition must be provided.
-
The ON conditions must uniquely identify the matching rows in the target table.
-
Each individual WHEN check can only invoke a single modification statement.
-
When a MATCHED search condition is true, the matching target row can be updated, deleted, or an error can be flagged.
-
When a NOT MATCHED search condition is true, the source row can be inserted into the target table, or an error can be flagged.
-
When more than one MATCHED or NOT MATCHED search condition is true, the first one that matches (for each type) is applied. This prevents any target row from being updated or deleted more than once. Ditto for any source row being inserted.
-
The ELSE IGNORE phrase specifies that no action be taken if no WHEN check evaluates to true.
-
If an error is encountered, all changes are rolled back.
-
Row-level triggers are activated for each row merged, depending on the type of modification that is made. So if the merge initiates an insert, all insert triggers are invoked. If the same input initiates an update, all update triggers are invoked.
-
Statement-level triggers are activated, even if no rows are processed. So if a merge does either an insert, or an update, both types of statement triggers are invoked, even if all of the input is inserted.
3.5.1. Sample Tables
To illustrate the merge statement, the following test tables were created and populated:
CREATE TABLE old_staff AS
(SELECT id
, job
, salary
FROM staff
)
WITH NO DATA;
CREATE TABLE new_staff AS
(SELECT id
, salary
FROM staff
)
WITH NO DATA;
INSERT INTO old_staff
SELECT id
, job
, salary
FROM staff
WHERE id BETWEEN 20 and 40;
OLD_STAFF
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
30 |
Mgr |
77506.75 |
40 |
Sales |
78006.00 |
INSERT INTO new_staff
SELECT id, salary / 10
FROM staff
WHERE id BETWEEN 30 and 50;
NEW_STAFF
ID | SALARY |
---|---|
30 |
7750.67 |
40 |
7800.60 |
50 |
8065.98 |
3.5.2. Update or Insert Merge
The next statement merges the new staff table into the old, using the following rules:
-
The two tables are matched on common ID columns.
-
If a row matches, the salary is updated with the new value.
-
If there is no matching row, a new row is inserted.
Now for the code:
MERGE INTO old_staff oo
USING new_staff nn
ON oo.id = nn.id
WHEN MATCHED THEN
UPDATE
SET oo.salary = nn.salary
WHEN NOT MATCHED THEN
INSERT
VALUES (nn.id,'?',nn.salary);
OLD_STAFF
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
30 |
Mgr |
77506.75 |
40 |
Sales |
78006.00 |
NEW_STAFF
ID | SALARY |
---|---|
30 |
7750.67 |
40 |
7800.60 |
50 |
8065.98 |
AFTER-MERGE
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
30 |
Mgr |
7750.67 |
40 |
Sales |
7800.60 |
50 |
? |
8065.98 |
3.5.3. Delete-only Merge
The next statement deletes all matching rows:
MERGE INTO old_staff oo
USING new_staff nn
ON oo.id = nn.id
WHEN MATCHED THEN
DELETE;
AFTER-MERGE
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
3.5.4. Complex Merge
The next statement has the following options:
-
The two tables are matched on common ID columns.
-
If a row matches, and the old salary is < 18,000, it is updated.
-
If a row matches, and the old salary is > 18,000, it is deleted.
-
If no row matches, and the new ID is > 10, the new row is inserted.
-
If no row matches, and (by implication) the new ID is ⇐ 10, an error is flagged.
Now for the code:
MERGE INTO old_staff oo
USING new_staff nn
ON oo.id = nn.id
WHEN MATCHED
AND oo.salary < 78000 THEN
UPDATE
SET oo.salary = nn.salary
WHEN MATCHED
AND oo.salary > 78000 THEN
DELETE
WHEN NOT MATCHED
AND nn.id > 10 THEN
INSERT
VALUES (nn.id,'?',nn.salary)
WHEN NOT MATCHED THEN
SIGNAL SQLSTATE '70001'
SET MESSAGE_TEXT = 'New ID <= 10';
OLD_STAFF
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
30 |
Mgr |
77506.75 |
40 |
Sales |
78006.00 |
NEW_STAFF
ID | SALARY |
---|---|
30 |
7750.67 |
40 |
7800.60 |
50 |
8065.98 |
AFTER-MERGE
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
30 |
Mgr |
7750.67 |
50 |
? |
8065.98 |
The merge statement is like the case statement (see CASE Expression) in that the sequence in which one writes the WHEN checks determines the processing logic. In the above example, if the last check was written before the prior, any non-match would generate an error.
3.5.5. Using a Fullselect
The following merge generates an input table (i.e. fullselect) that has a single row containing the MAX value of every field in the relevant table. This row is then inserted into the table:
MERGE INTO old_staff
USING
(SELECT MAX(id) + 1 AS max_id
, MAX(job) AS max_job
, MAX(salary) AS max_sal
FROM old_staff
) AS mx
ON id = max_id
WHEN NOT MATCHED THEN
INSERT
VALUES (max_id, max_job, max_sal);
AFTER-MERGE
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
30 |
Mgr |
77506.75 |
40 |
Sales |
78006.00 |
41 |
Sales |
78171.25 |
Here is the same thing written as a plain on insert:
INSERT INTO old_staff
SELECT MAX(id) + 1 AS max_id
, MAX(job) AS max_job
, MAX(salary) AS max_sal
FROM old_staff;
Use a fullselect on the target and/or source table to limit the set of rows that are processed during the merge:
MERGE INTO
(SELECT *
FROM old_staff
WHERE id < 40
) AS oo
USING
(SELECT *
FROM new_staff
WHERE id < 50
) AS nn
ON oo.id = nn.id
WHEN MATCHED THEN
DELETE
WHEN NOT MATCHED THEN
INSERT
VALUES (nn.id,'?',nn.salary);
OLD_STAFF
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
30 |
Mgr |
77506.75 |
40 |
Sales |
78006.00 |
NEW_STAFF
ID | SALARY |
---|---|
30 |
7750.67 |
40 |
7800.60 |
50 |
8065.98 |
AFTER-MERGE
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
40 |
? |
7800.60 |
40 |
Sales |
78006.00 |
Observe that the above merge did the following:
-
The target row with an ID of 30 was deleted - because it matched.
-
The target row with an ID of 40 was not deleted, because it was excluded in the fullselect that was done before the merge.
-
The source row with an ID of 40 was inserted, because it was not found in the target fullselect. This is why the base table now has two rows with an ID of 40.
-
The source row with an ID of 50 was not inserted, because it was excluded in the fullselect that was done before the merge.
3.5.6. Listing Columns
The next example explicitly lists the target fields in the insert statement - so they correspond to those listed in the following values phrase:
MERGE INTO old_staff oo
USING new_staff nn
ON oo.id = nn.id
WHEN MATCHED THEN
UPDATE
SET (salary, job) = (1234, '?')
WHEN NOT MATCHED THEN
INSERT (id,salary,job)
VALUES (id,5678.9,'?');
AFTER-MERGE
ID | JOB | SALARY |
---|---|---|
20 |
Sales |
78171.25 |
30 |
? |
1234.00 |
40 |
? |
1234.00 |
50 |
? |
5678.90 |
4. Compound SQL
A compound statement groups multiple independent SQL statements into a single executable. In addition, simple processing logic can be included to create what is, in effect, a very basic program. Such statements can be embedded in triggers, SQL functions, SQL methods, and dynamic SQL statements.
4.1. Introduction
A compound SQL statement begins with an (optional) name, followed by the variable declarations, followed by the procedural logic.
Below is a compound statement that reads a set of rows from the STAFF table and, for each row fetched, updates the COMM field to equal the current fetch number.
BEGIN ATOMIC
DECLARE cntr SMALLINT DEFAULT 1;
FOR V1 AS
SELECT id as idval
FROM staff
WHERE id < 80
ORDER BY id
DO
UPDATE staff
SET comm = cntr
WHERE id = idval;
SET cntr = cntr + 1;
END FOR;
END
4.1.1. Statement Delimiter
Db2 SQL does not come with a designated statement delimiter (terminator), though a semicolon is typically used. However, a semi-colon cannot be used in a compound SQL statement because that character is used to differentiate the sub-components of the statement. In Db2BATCH, one can run the SET DELIMITER command (intelligent comment) to use something other than a semi-colon. The following script illustrates this usage:
--#SET DELIMITER !
SELECT NAME FROM STAFF WHERE id = 10!
--#SET DELIMITER ;
SELECT NAME FROM STAFF WHERE id = 20;
In the Db2 command processor one can do the same thing using the terminator keyword:
--#SET TERMINATOR !
SELECT NAME FROM STAFF WHERE id = 10!
--#SET TERMINATOR ;
SELECT NAME FROM STAFF WHERE id = 20;
4.1.2. SQL Statement Usage
When used in dynamic SQL, the following control statements can be used:
-
FOR statement
-
GET DIAGNOSTICS statement
-
IF statement
-
ITERATE statement
-
LEAVE statement
-
SIGNAL statement
-
WHILE statement
There are many more PSM (persistent stored modules) control statements than what is shown above. But only these ones can be used in Compound SQL statements. |
The following SQL statements can be issued:
-
fullselect
-
UPDATE
-
DELETE
-
INSERT
-
SET variable statement
4.1.3. DECLARE Variables
All variables have to be declared at the start of the compound statement. Each variable must be given a name and a type and, optionally, a default (start) value.
BEGIN ATOMIC
DECLARE aaa, bbb, ccc SMALLINT DEFAULT 1;
DECLARE ddd CHAR(10) DEFAULT NULL;
DECLARE eee INTEGER;
SET eee = aaa + 1;
UPDATE staff
SET comm = aaa
, salary = bbb
, years = eee
WHERE id = 10;
END
FOR Statement
The FOR statement executes a group of statements for each row fetched from a query.
In the next example one row is fetched per year of service (for selected years) in the STAFF table. That row is then used to do two independent updates to the three matching rows:
BEGIN ATOMIC
FOR V1 AS
SELECT years AS yr_num
, max(id) AS max_id
FROM staff
WHERE years < 4
GROUP BY years
ORDER BY years
DO
UPDATE staff
SET salary = salary / 10
WHERE id = max_id;
UPDATE staff
set comm = 0
WHERE years = yr_num;
END FOR;
END
BEFORE
ID | SALARY | COMM |
---|---|---|
180 |
37009.75 |
236.50 |
230 |
83369.80 |
189.65 |
330 |
49988.00 |
55.50 |
AFTER
ID | SALARY | COMM |
---|---|---|
180 |
37009.75 |
0.00 |
230 |
8336.98 |
0.00 |
330 |
4998.80 |
0.00 |
4.1.4. GET DIAGNOSTICS Statement
The GET DIAGNOSTICS statement returns information about the most recently run SQL statement. One can either get the number of rows processed (i.e. inserted, updated, or deleted), or the return status (for an external procedure call).
In the example below, some number of rows are updated in the STAFF table. Then the count of rows updated is obtained, and used to update a row in the STAFF table:
BEGIN ATOMIC
DECLARE numrows INT DEFAULT 0;
UPDATE staff
SET salary = 12345
WHERE id < 100;
GET DIAGNOSTICS numrows = ROW_COUNT;
UPDATE staff
SET salary = numrows
WHERE id = 10;
END
4.1.5. IF Statement
The IF statement is used to do standard if-then-else branching logic. It always begins with an IF THEN statement and ends with and END IF statement.
The next example uses if-then-else logic to update one of three rows in the STAFF table, depending on the current timestamp value:
BEGIN ATOMIC
DECLARE cur INT;
SET cur = MICROSECOND(CURRENT TIMESTAMP);
IF cur > 600000 THEN
UPDATE staff
SET name = CHAR(cur)
WHERE id = 10;
ELSEIF cur > 300000 THEN
UPDATE staff
SET name = CHAR(cur)
WHERE id = 20;
ELSE
UPDATE staff
SET name = CHAR(cur)
WHERE id = 30;
END IF;
END
4.1.6. ITERATE Statement
The ITERATE statement causes the program to return to the beginning of the labeled loop.
In next example, the second update statement will never get performed because the ITERATE will always return the program to the start of the loop:
BEGIN ATOMIC
DECLARE cntr INT DEFAULT 0;
whileloop:
WHILE cntr < 60 DO
SET cntr = cntr + 10;
UPDATE staff
SET salary = cntr
WHERE id = cntr;
ITERATE whileloop;
UPDATE staff
SET comm = cntr + 1
WHERE id = cntr;
END WHILE;
END
4.1.7. LEAVE Statement
The LEAVE statement exits the labeled loop.
In the next example, the WHILE loop would continue forever, if left to its own devices. But after some random number of iterations, the LEAVE statement will exit the loop:
BEGIN ATOMIC
DECLARE cntr INT DEFAULT 1;
whileloop:
WHILE 1 <> 2 DO
SET cntr = cntr + 1;
IF RAND() > 0.99 THEN
LEAVE whileloop;
END IF;
END WHILE;
UPDATE staff
SET salary = cntr
WHERE id = 10;
END
4.1.8. SIGNAL Statement
The SIGNAL statement is used to issue an error or warning message.
The next example loops a random number of times, and then generates an error message using the SIGNAL command, saying how many loops were done:
BEGIN ATOMIC
DECLARE cntr INT DEFAULT 1;
DECLARE emsg CHAR(20);
whileloop:
WHILE RAND() < .99 DO
SET cntr = cntr + 1;
END WHILE;
SET emsg = '#loops: ' || CHAR(cntr);
SIGNAL SQLSTATE '75001' SET MESSAGE_TEXT = emsg;
END
4.1.9. WHILE Statement
The WHILE statement repeats one or more statements while some condition is true.
The next statement has two nested WHILE loops, and then updates the STAFF table:
BEGIN ATOMIC
DECLARE c1, C2 INT DEFAULT 1;
WHILE c1 < 10 DO
WHILE c2 < 20 DO
SET c2 = c2 + 1;
END WHILE;
SET c1 = c1 + 1;
END WHILE;
UPDATE staff
SET salary = c1
, comm = c2
WHERE id = 10;
END
4.1.10. Other Usage
The following Db2 objects also support the language elements described above:
-
Triggers
-
Stored procedures
-
User-defined functions
-
Embedded compound SQL (in programs).
Some of the above support many more language elements. For example stored procedures that are written in SQL also allow the following:
ASSOCIATE, CASE, GOTO, LOOP, REPEAT, RESIGNAL
, and RETURN
.
4.1.11. Test Query
To illustrate some of the above uses of compound SQL, we are going to get from the STAFF table a complete list of departments, and the number of rows in each department. Here is the basic query, with the related answer:
SELECT dept
, count(*) as #rows
FROM staff
GROUP BY dept
ORDER BY dept;
ANSWER
DEPT | #ROWS |
---|---|
10 |
4 |
15 |
4 |
20 |
4 |
38 |
5 |
42 |
4 |
51 |
5 |
66 |
5 |
84 |
4 |
If all you want to get is this list, the above query is the way to go. But we will get the same answer using various other methods, just to show how it can be done using compound SQL statements.
Trigger
One cannot get an answer using a trigger. All one can do is alter what happens during an insert, update, or delete. With this in mind, the following example does the following:
-
Sets the statement delimiter to an "!". Because we are using compound SQL inside the trigger definition, we cannot use the usual semi-colon.
-
Creates a new table (note: triggers are not allowed on temporary tables).
-
Creates an INSERT trigger on the new table. This trigger gets the number of rows per department in the STAFF table - for each row (department) inserted.
-
Inserts a list of departments into the new table.
-
Selects from the new table.
Now for the code:
--#SET DELIMITER !
CREATE TABLE dpt
( dept SMALLINT NOT NULL
, #names SMALLINT
, PRIMARY KEY(dept))!
COMMIT!
CREATE TRIGGER dpt1 AFTER INSERT ON dpt
REFERENCING NEW AS NNN
FOR EACH ROW
MODE Db2SQL
BEGIN ATOMIC
DECLARE namecnt SMALLINT DEFAULT 0;
FOR getnames AS
SELECT COUNT(*) AS #n
FROM staff
WHERE dept = nnn.dept
DO
SET namecnt = #n;
END FOR;
UPDATE dpt
SET #names = namecnt
WHERE dept = nnn.dept;
END!
COMMIT!
INSERT INTO dpt (dept)
SELECT DISTINCT dept
FROM staff!
COMMIT!
SELECT *
FROM dpt
ORDER BY dept!
This example uses an "!" as the stmt delimiter. |
ANSWER
DEPT | #NAMES |
---|---|
10 |
4 |
15 |
4 |
20 |
4 |
38 |
5 |
42 |
4 |
51 |
5 |
66 |
5 |
84 |
4 |
The above code was designed to be run in Db2BATCH. The "set delimiter" notation will probably not work in other environments. |
Scalar Function
One can do something very similar to the above that is almost as stupid using a user-defined scalar function, that calculates the number of rows in a given department. The basic logic will go as follows:
-
Set the statement delimiter to an "!".
-
Create the scalar function.
-
Run a query that first gets a list of distinct departments, then calls the function.
Here is the code:
--#SET DELIMITER !
CREATE FUNCTION dpt1 (deptin SMALLINT)
RETURNS SMALLINT
BEGIN ATOMIC
DECLARE num_names SMALLINT;
FOR getnames AS
SELECT COUNT(*) AS #n
FROM staff
WHERE dept = deptin
DO
SET num_names = #n;
END FOR;
RETURN num_names;
END!
COMMIT!
SELECT XXX.*
, dpt1(dept) as #names
FROM
(SELECT dept
FROM staff
GROUP BY dept
) AS XXX
ORDER BY dept!
This example uses an "!" as the stmt delimiter. |
ANSWER
DEPT | #NAMES |
---|---|
10 |
4 |
15 |
4 |
20 |
4 |
38 |
5 |
42 |
4 |
51 |
5 |
66 |
5 |
84 |
4 |
Because the query used in the above function will only ever return one row, we can greatly simplify the function definition thus:
--#SET DELIMITER !
CREATE FUNCTION dpt1 (deptin SMALLINT)
RETURNS SMALLINT
BEGIN ATOMIC
RETURN
SELECT COUNT(*)
FROM staff
WHERE dept = deptin;
END!
COMMIT!
This example uses an "!" as the stmt delimiter. |
SELECT XXX.* , dpt1(dept) as #names FROM (SELECT dept FROM staff GROUP BY dept ) AS XXX ORDER BY dept!
In the above example, the RETURN statement is directly finding the one matching row, and then returning it to the calling statement.
Table Function
Below is almost exactly the same logic, this time using a table function:
Table Function with compound SQL
--#SET DELIMITER !
CREATE FUNCTION dpt2 ()
RETURNS TABLE ( dept SMALLINT
, #names SMALLINT)
BEGIN ATOMIC
RETURN
SELECT dept
, count(*)
FROM staff
GROUP BY dept
ORDER BY dept;
END!
COMMIT!
--#SET DELIMITER ;
SELECT *
FROM TABLE(dpt2()) T1
ORDER BY dept;
This example uses an "!" as the stmt delimiter. |
ANSWER
DEPT | #NAMES |
---|---|
10 |
4 |
15 |
4 |
20 |
4 |
38 |
5 |
42 |
4 |
51 |
5 |
66 |
5 |
84 |
4 |
BEGIN ATOMIC
DECLARE cntr SMALLINT DEFAULT 1;
FOR V1 AS
SELECT id as idval
FROM staff
WHERE id < 80
ORDER BY id
DO
UPDATE staff SET comm = cntr
WHERE id = idval;
SET cntr = cntr + 1;
END FOR;
END
5. Column Functions or Aggregate Functions
By themselves, column functions work on the complete set of matching rows. One can use a GROUP BY expression to limit them to a subset of matching rows. One can also use them in an OLAP function to treat individual rows differently.
Be very careful when using either a column function, or the DISTINCT clause, in a join. If the join is incorrectly coded, and does some form of Cartesian Product, the column function may get rid of the all the extra (wrong) rows so that it becomes very hard to confirm that the answer is incorrect. Likewise, be appropriately suspicious whenever you see that someone (else) has used a DISTINCT statement in a join. Sometimes, users add the DISTINCT clause to get rid of duplicate rows that they didn’t anticipate and don’t understand. |
5.1. Column Functions, Definitions
5.1.1. ARRAY_AGG
Aggregate the set of elements in an array. If an ORDER BY is provided, it determines the order in which the elements are entered into the array.
5.1.2. AVG
Get the average (mean) value of a set of non-null rows. The columns(s) must be numeric. ALL is the default. If DISTINCT is used duplicate values are ignored. If no rows match, the null value is returned.
SELECT AVG(dept) AS a1
, AVG(ALL dept) AS a2
, AVG(DISTINCT dept) AS a3
, AVG(dept/10) AS a4
, AVG(dept)/10 AS a5
FROM staff
HAVING AVG(dept) > 40;
ANSWER
A1 | A2 | A3 | A4 | A5 |
---|---|---|---|---|
41 |
41 |
40 |
3 |
4 |
Observe columns A4 and A5 above. Column A4 has the average of each value divided by 10. Column A5 has the average of all of the values divided by 10. In the former case, precision has been lost due to rounding of the original integer value and the result is arguably incorrect. This problem also occurs when using the SUM function. |
Averaging Null and Not-Null Values
Some database designers have an intense and irrational dislike of using nullable fields. What they do instead is define all columns as not-null and then set the individual fields to zero (for numbers) or blank (for characters) when the value is unknown. This solution is reasonable in some situations, but it can cause the AVG function to give what is arguably the wrong answer. One solution to this problem is some form of counseling or group therapy to overcome the phobia. Alternatively, one can use the CASE expression to put null values back into the answer-set being processed by the AVG function. The following SQL statement uses a modified version of the IBM sample STAFF table (all null COMM values were changed to zero) to illustrate the technique:
UPDATE staff
SET comm = 0
WHERE comm IS NULL;
SELECT AVG(salary) AS salary
, AVG(comm) AS comm1
, AVG(CASE comm
WHEN 0 THEN NULL
ELSE comm
END) AS comm2
FROM staff;
ANSWER
SALARY | COMM1 | COMM2 |
---|---|---|
67932.78 |
351.98 |
513.31 |
UPDATE staff
SET comm = NULL
WHERE comm = 0;
The COMM2 field above is the correct average. The COMM1 field is incorrect because it has factored in the zero rows with really represent null values. Note that, in this particular query, one cannot use a WHERE to exclude the "zero" COMM rows because it would affect the average salary value.
Dealing with Null Output
The AVG, MIN, MAX, and SUM functions almost always return a null value when there are no matching rows (see No Rows Match for exceptions). One can use the COALESCE function, or a CASE expression, to convert the null value into a suitable substitute. Both methodologies are illustrated below:
SELECT COUNT(*) AS c1
, AVG(salary) AS a1
, COALESCE(AVG(salary),0) AS a2
, CASE
WHEN AVG(salary) IS NULL THEN 0
ELSE AVG(salary)
END AS a3
FROM staff
WHERE id < 10;
ANSWER
C1 | A1 | A2 | A3 |
---|---|---|---|
0 |
- |
0 |
0 |
AVG Date/Time Values
The AVG function only accepts numeric input. However, one can, with a bit of trickery, also use the AVG function on a date field. First convert the date to the number of days since the start of the Current Era, then get the average, then convert the result back to a date. Please be aware that, in many cases, the average of a date does not really make good business sense. Having said that, the following SQL gets the average birth-date of all employees:
SELECT AVG(DAYS(birthdate))
, DATE(AVG(DAYS(birthdate)))
FROM employee;
ANSWER
1 | 2 |
---|---|
721092 |
1975-04-14 |
Time data can be manipulated in a similar manner using the MIDNIGHT_SECONDS function. If one is really desperate (or silly), the average of a character field can also be obtained using the ASCII and CHR functions.
Average of an Average
In some cases, getting the average of an average gives an overflow error. Inasmuch as you shouldn’t do this anyway, it is no big deal:
SELECT AVG(avg_sal) AS avg_avg
FROM (SELECT dept
, AVG(salary) AS avg_sal
FROM staff
GROUP BY dept
) AS xxx;
ANSWER: Overflow error
5.1.3. CORRELATION
I don’t know a thing about statistics, so I haven’t a clue what this function does. But I do know that the SQL Reference is wrong - because it says the value returned will be between 0 and 1. I found that it is between -1 and +1 (see below). The output type is float.
WITH temp1(col1, col2, col3, col4) AS
(VALUES (0, 0, 0, RAND(1))
UNION ALL
SELECT col1 + 1
, col2 - 1
, RAND()
, RAND()
FROM temp1
WHERE col1 < = 1000
)
SELECT DEC(CORRELATION(col1, col1), 5, 3) AS cor11
, DEC(CORRELATION(col1, col2), 5, 3) AS cor12
, DEC(CORRELATION(col2, col3), 5, 3) AS cor23
, DEC(CORRELATION(col3, col4), 5, 3) AS cor34
FROM temp1;
ANSWER
COR11 | COR12 | COR23 | COR34 |
---|---|---|---|
1.000 |
-1.000 |
-0.017 |
-0.005 |
5.1.4. COUNT
Get the number of values in a set of rows. The result is an integer. The value returned depends upon the options used:
-
COUNT(*) gets a count of matching rows.
-
COUNT(expression) gets a count of rows with a non-null expression value.
-
COUNT(ALL expression) is the same as the COUNT(expression) statement.
-
COUNT(DISTINCT expression) gets a count of distinct non-null expression values.
SELECT COUNT(*) AS c1
, COUNT(INT(comm/10)) AS c2
, COUNT(ALL INT(comm/10)) AS c3
, COUNT(DISTINCT INT(comm/10)) AS c4
, COUNT(DISTINCT INT(comm)) AS c5
, COUNT(DISTINCT INT(comm))/10 AS c6
FROM staff;
ANSWER
C1 | C2 | C3 | C4 | C5 | C6 |
---|---|---|---|---|---|
35 |
24 |
24 |
19 |
24 |
2 |
There are 35 rows in the STAFF table (see C1 above), but only 24 of them have non-null commission values (see C2 above). If no rows match, the COUNT returns zero - except when the SQL statement also contains a GROUP BY. In this latter case, the result is no row.
SELECT 'NO GP-BY' AS c1
, COUNT(*) AS c2
FROM staff
WHERE id = -1
UNION
SELECT 'GROUP-BY' AS c1
, COUNT(*) AS c2
FROM staff
WHERE id = -1
GROUP BY dept;
ANSWER
C1 | C2 |
---|---|
NO GP-BY |
0 |
5.1.5. COUNT_BIG
Get the number of rows or distinct values in a set of rows. Use this function if the result is too large for the COUNT function. The result is of type decimal 31. If the DISTINCT option is used both duplicate and null values are eliminated. If no rows match, the result is zero.
SELECT COUNT_BIG(*) AS c1
, COUNT_BIG(dept) AS c2
, COUNT_BIG(DISTINCT dept) AS c3
, COUNT_BIG(DISTINCT dept/10) AS c4
, COUNT_BIG(DISTINCT dept)/10 AS c5
FROM STAFF;
ANSWER
C1 | C2 | C3 | C4 | C5 |
---|---|---|---|---|
35. |
35. |
8. |
7. |
0. |
5.1.6. COVARIANCE
Returns the covariance of a set of number pairs. The output type is float.
WITH temp1(c1, c2, c3, c4) AS
(VALUES (0 , 0 , 0 , RAND(1))
UNION ALL
SELECT c1 + 1
, c2 - 1
, RAND()
, RAND()
FROM temp1
WHERE c1 <= 1000
)
SELECT DEC(COVARIANCE(c1,c1),6,0) AS cov11
, DEC(COVARIANCE(c1,c2),6,0) AS cov12
, DEC(COVARIANCE(c2,c3),6,4) AS cov23
, DEC(COVARIANCE(c3,c4),6,4) AS cov34
FROM temp1;
ANSWER
COV11 | COV12 | COV23 | COV34 |
---|---|---|---|
83666. |
-83666. |
-1.4689 |
-0.0004 |
5.1.7. COVARIANCE_SAMP
Returns the sample covariance of a set of number pairs.
5.1.8. CUME_DIST
Returns the cumulative distribution of a row that is hypothetically inserted into a group of rows.
5.1.9. GROUPING
The GROUPING function is used in CUBE, ROLLUP, and GROUPING SETS statements to identify what rows come from which particular GROUPING SET. A value of 1 indicates that the corresponding data field is null because the row is from of a GROUPING SET that does not involve this row. Otherwise, the value is zero.
SELECT dept
, AVG(salary) AS salary
, GROUPING(dept) AS df
FROM staff
GROUP BY ROLLUP(dept)
ORDER BY dept;
ANSWER
DEPT | SALARY | DF |
---|---|---|
10 |
83365.86 |
0 |
15 |
60482.33 |
0 |
20 |
63571.52 |
0 |
38 |
60457.11 |
0 |
42 |
49592.26 |
0 |
51 |
83218.16 |
0 |
66 |
73015.24 |
0 |
84 |
66536.75 |
0 |
- |
67932.78 |
1 |
See the section titled "Group By and Having" for more information on this function. |
5.1.10. LISTAGG
Aggregates a set of string elements into one string by concatenating the strings. Optionally, a separator string can be provided which is inserted between contiguous input strings.
5.1.11. MAX
Get the maximum value of a set of rows. The use of the DISTINCT option has no affect. If no rows match, the null value is returned.
SELECT MAX(dept)
, MAX(ALL dept)
, MAX(DISTINCT dept)
, MAX(DISTINCT dept/10)
FROM staff;
ANSWER
1 | 2 | 3 | 4 |
---|---|---|---|
84 |
84 |
84 |
8 |
MAX and MIN usage with Scalar Functions
Several Db2 scalar functions convert a value from one format to another, for example from numeric to character. The function output format will not always shave the same ordering sequence as the input. This difference can affect MIN, MAX, and ORDER BY processing.
SELECT MAX(hiredate)
, CHAR(MAX(hiredate),USA)
, MAX(CHAR(hiredate,USA))
FROM employee;
ANSWER
1 | 2 | 3 |
---|---|---|
2006-12-15 |
12/15/2006 |
12/15/2006 |
In the above the SQL, the second field gets the MAX before doing the conversion to character whereas the third field works the other way round. In most cases, the later is wrong. In the next example, the MAX function is used on a small integer value that has been converted to character. If the CHAR function is used for the conversion, the output is left justified, which results in an incorrect answer. The DIGITS output is correct (in this example).
SELECT MAX(id) AS id
, MAX(CHAR(id)) AS chr
, MAX(DIGITS(id)) AS dig
FROM staff;
ANSWER
ID | CHR | DIG |
---|---|---|
350 |
90 |
00350 |
The DIGITS function can also give the wrong answer - if the input data is part positive and part negative. This is because this function does not put a sign indicator in the output.
SELECT MAX(id - 250) AS id
, MAX(CHAR(id - 250)) AS chr
, MAX(DIGITS(id - 250)) AS dig
FROM staff;
ANSWER
D | CHR | DIG |
---|---|---|
100 |
90 |
0000000240 |
Be careful when using a column function on a field that has been converted from number to character, or from date/time to character. The result may not be what you intended. |
5.1.12. MEDIAN
Returns the median value in a set of values.
5.1.13. MIN
Get the minimum value of a set of rows. The use of the DISTINCT option has no affect. If no rows match, the null value is returned.
SELECT MIN(dept)
, MIN(ALL dept)
, MIN(DISTINCT dept)
, MIN(DISTINCT dept/10)
FROM staff;
ANSWER
1 | 2 | 3 | 4 |
---|---|---|---|
10 |
10 |
10 |
1 |
5.1.14. PERCENTILE_CONT
Returns the value that corresponds to the specified percentile given a sort specification by using a continuous distribution model.
5.1.15. PERCENTILE_DISC
Returns the value that corresponds to the specified percentile given a sort specification by using a discrete distribution model.
5.1.16. PERCENT_RANK
Returns the relative percentile rank of a row that is hypothetically inserted into a group of rows.
5.1.17. Regression Functions
The various regression functions support the fitting of an ordinary-least-squares regression line of the form y = a * x + b to a set of number pairs.
REGR_AVGX returns a quantity that than can be used to compute the validity of the regression model. The output is of type float.
REGR_AVGY (see REGR_AVGX).
REGR_COUNT returns the number of matching non-null pairs. The output is integer.
REGR_INTERCEPT returns the y-intercept of the regression line.
REGR_R2 returns the coefficient of determination for the regression.
REGR_SLOPE returns the slope of the line.
REGR_SXX (see REGR_AVGX).
REGR_SXY (see REGR_AVGX).
REGR_SYY (see REGR_AVGX).
See the IBM SQL Reference for more details on the above functions.
SELECT DEC(REGR_SLOPE(bonus,salary),7,5) AS r_slope
, DEC(REGR_INTERCEPT(bonus,salary),7,3) AS r_icpt
, INT(REGR_COUNT(bonus,salary)) AS r_count
, INT(REGR_AVGX(bonus,salary)) AS r_avgx
, INT(REGR_AVGY(bonus,salary)) AS r_avgy
, DEC(REGR_SXX(bonus,salary),10) AS r_sxx
, INT(REGR_SXY(bonus,salary)) AS r_sxy
, INT(REGR_SYY(bonus,salary)) AS r_syy
FROM employee
WHERE workdept = 'A00';
ANSWERS
r_slope | r_icpt | r_count | r_avgx | r_avgy | r_sxx | r_sxy | r_syy |
---|---|---|---|---|---|---|---|
0.00247 |
644.862 |
5 |
70850 |
820 |
8784575000 |
21715000 |
168000 |
5.1.18. STDDEV
Get the standard deviation of a set of numeric values. If DISTINCT is used, duplicate values are ignored. If no rows match, the result is null. The output format is double.
SELECT AVG(dept) AS a1
,STDDEV(dept) AS s1
,DEC(STDDEV(dept),3,1) AS s2
,DEC(STDDEV(ALL dept),3,1) AS s3
,DEC(STDDEV(DISTINCT dept),3,1) AS s4
FROM staff;
ANSWER
A1 | S1 | S2 | S3 | S4 |
---|---|---|---|---|
41 |
+2.3522355E+1 |
23.5 |
23.5 |
24.1 |
5.1.19. STDDEV_SAMP
The STDDEV_SAMP function returns the sample standard deviation (division by [n-1]) of a set of numbers.
5.1.20. SUM
Get the sum of a set of numeric values. If DISTINCT is used, duplicate values are ignored. Null values are always ignored. If no rows match, the result is null.
SELECT SUM(dept) AS s1
, SUM(ALL dept) AS s2
, SUM(DISTINCT dept) AS s3
, SUM(dept/10) AS s4
, SUM(dept)/10 AS s5
FROM staff;
ANSWER
S1 | S2 | S3 | S4 | S5 |
---|---|---|---|---|
1459 |
1459 |
326 |
134 |
145 |
The answers S4 and S5 above are different. This is because the division is done before the SUM in column S4, and after in column S5. In the former case, precision has been lost due to rounding of the original integer value and the result is arguably incorrect. When in doubt, use the S5 notation. |
5.1.21. VAR or VARIANCE
Get the variance of a set of numeric values. If DISTINCT is used, duplicate values are ignored. If no rows match, the result is null. The output format is double.
SELECT AVG(dept) AS a1
, VARIANCE(dept) AS s1
, DEC(VARIANCE(dept),4,1) AS s2
, DEC(VARIANCE(ALL dept),4,1) AS s3
, DEC(VARIANCE(DISTINCT dept),4,1) AS s4
FROM staff;
ANSWER
A1 | V1 | V2 | V3 | V4 |
---|---|---|---|---|
41 |
+5.533012244E+2 |
553 |
553 |
582 |
5.1.22. VARIANCE_SAMP
Returns the sample variance (division by [n-1]) of a set of numbers.
5.1.23. XMLAGG
Returns an XML sequence containing an item for each non-null value in a set of XML values.
5.1.24. XMLGROUP
The XMLGROUP function returns an XML value with a single XQuery document node containing one top-level element node. This is an aggregate expression that will return a single-rooted XML document from a group of rows where each row is mapped to a row subelement.
6. OLAP Functions
=== Introduction
Online Analytical Processing (OLAP) functions enable one to sequence and rank query rows. They are especially useful when the calling program is very simple.
6.1. The Bad Old Days
To really appreciate the value of the OLAP functions, one should try to do some seemingly trivial task without them. To illustrate this point, consider the following query:
SELECT s1.job
, s1.id
, s1.salary
FROM staff s1
WHERE s1.name LIKE '%s%'
AND s1.id < 90
ORDER BY s1.job
, s1.id;
ANSWER
JOB | ID | SALARY |
---|---|---|
Clerk |
80 |
43504.60 |
Mgr |
10 |
98357.50 |
Mgr |
50 |
80659.80 |
Let us now add two fields to this query:
-
A running sum of the salaries selected.
-
A running count of the rows retrieved.
Adding these fields is easy - when using OLAP functions:
SELECT s1.job
, s1.id
, s1.salary
, SUM(salary) OVER(ORDER BY job, id) AS sumsal
, ROW_NUMBER() OVER(ORDER BY job, id) AS r
FROM staff s1
WHERE s1.name LIKE '%s%'
AND s1.id < 90
ORDER BY s1.job
, s1.id;
ANSWER
JOB | ID | SALARY | SUMSAL | R |
---|---|---|---|---|
Clerk |
80 |
43504.60 |
43504.60 |
1 |
Mgr |
10 |
98357.50 |
141862.10 |
2 |
Mgr |
50 |
80659.80 |
222521.90 |
3 |
6.1.1. Write Query without OLAP Functions
If one does not have OLAP functions, one can still get the required answer, but the code is quite tricky. The problem is that this seemingly simple query contains two nasty tricks:
-
Not all of the rows in the table are selected.
-
The output is ordered on two fields, the first of which is not unique.
Below is the arguably the most elegant way to write the above query without using OLAP functions. There query has the following basic characteristics:
-
Define a common-table-expression with the set of matching rows.
-
Query from this common-table-expression.
-
For each row fetched, do two nested select statements. The first gets a running sum of the salaries, and the second gets a running count of the rows retrieved.
Now for the code:
WITH temp1 AS
(SELECT *
FROM staff s1
WHERE s1.name LIKE '%s%'
AND s1.id < 90
)
SELECT s1.job
, s1.id
, s1.salary
, (SELECT SUM(s2.salary)
FROM temp1 s2
WHERE (s2.job < s1.job)
OR (s2.job = s1.job AND s2.id <= s1.id)
) AS sumsal
, (SELECT COUNT(*)
FROM temp1 s2
WHERE (s2.job < s1.job)
OR (s2.job = s1.job AND s2.id <= s1.id)
) AS r
FROM temp1 s1
ORDER BY s1.job
, s1.id;
ANSWER
JOB | ID | SALARY | SUMSAL | R |
---|---|---|---|---|
Clerk |
80 |
43504.60 |
43504.60 |
1 |
Mgr |
10 |
98357.50 |
141862.10 |
2 |
Mgr |
50 |
80659.80 |
222521.90 |
3 |
6.1.2. Concepts
Below are some of the basic characteristics of OLAP functions:
-
OLAP functions are column functions that work (only) on the set of rows that match the predicates of the query.
-
Unlike ordinarily column functions, (e.g. SUM), OLAP functions do not require that the whole answer-set be summarized. In fact, OLAP functions never change the number of rows returned by the query.
-
OLAP functions work on sets of values, but the result is always a single value.
-
OLAP functions are used to return individual rows from a table (e.g. about each staff member), along with related summary data (e.g. average salary in department).
-
OLAP functions are often applied on some set (i.e. of a moving window) of rows that is defined relative to the current row being processed. These matching rows are classified using an ORDER BY as being one of three types:
-
Preceding rows are those that have already been processed.
-
Following rows are those that have yet to be processed.
-
Current row is the one currently being processed.
-
-
The ORDER BY used in an OLAP function is not related to the ORDER BY expression used to define the output order of the final answer set.
-
OLAP functions can summarize the matching rows by a subset (i.e. partition). When this is done, it is similar to the use of a GROUP BY in an ordinary column function.
Below is a query that illustrates these concepts. It gets some individual rows from the STAFF table, while using an OLAP function to calculate a running average salary within the DEPT of the current row. The average is calculated using one preceding row (in ID order), the current row, and two following rows:
SELECT dept
, id
, salary
, DEC(AVG(salary) OVER(PARTITION BY dept
ORDER BY id
ROWS BETWEEN 1 PRECEDING
AND 2 FOLLOWING)
, 8, 2) AS avb_sal
FROM staff
WHERE dept IN (20, 38)
ORDER BY dept
, id;
ANSWER
DEPT | ID | SALARY | AVG_SAL |
---|---|---|---|
20 |
10 |
98357.50 |
73344.45 |
20 |
20 |
78171.25 |
63571.52 |
20 |
80 |
43504.60 |
51976.20 |
20 |
190 |
34252.75 |
38878.67 |
38 |
30 |
77506.75 |
74107.01 |
38 |
40 |
78006.00 |
66318.95 |
38 |
60 |
66808.30 |
56194.70 |
38 |
120 |
42954.75 |
48924.26 |
38 |
180 |
37009.75 |
39982.25 |
TABLE
DEPT | ID | SALARY | Matching? | Partition | Relation to row [38 60 66808.30 56194.70] |
---|---|---|---|---|---|
15 |
110 |
42508.20 |
N |
- |
- |
15 |
170 |
42258.50 |
N |
- |
- |
20 |
10 |
98357.50 |
Y |
1 |
- |
20 |
20 |
78171.25 |
Y |
1 |
- |
20 |
80 |
43504.60 |
Y |
1 |
- |
20 |
190 |
34252.75 |
Y |
1 |
- |
38 |
30 |
77506.75 |
Y |
2 |
Preceding row |
38 |
40 |
78006.00 |
Y |
2 |
Preceding row |
38 |
60 |
66808.30 |
Y |
2 |
Current row |
38 |
120 |
42954.75 |
Y |
2 |
Following row |
38 |
180 |
37009.75 |
Y |
2 |
Following row |
42 |
90 |
38001.75 |
N |
- |
- |
42 |
100 |
78352.80 |
N |
- |
- |
42 |
130 |
40505.90 |
N |
- |
- |
Below is another query that calculates various running averages:
SELECT dept
, id
, salary
, DEC(AVG(salary) OVER() ,8,2) AS avg1
, DEC(AVG(salary) OVER(PARTITION BY dept) ,8,2) AS avg2
, DEC(AVG(salary) OVER(PARTITION BY dept
ORDER BY id
ROWS UNBOUNDED PRECEDING)
, 8, 2) AS avg3
, DEC(AVG(salary) OVER(PARTITION BY dept
ORDER BY id
ROWS BETWEEN 1 PRECEDING AND 2 FOLLOWING)
, 8, 2) AS avg4
FROM staff
WHERE dept IN (15,20)
AND id > 20
ORDER BY dept
, id;
ANSWER
DEPT | ID | SALARY | AVG1 | AVG2 | AVG3 | AVG4 |
---|---|---|---|---|---|---|
15 |
50 |
80659.80 |
53281.11 |
60482.33 |
80659.80 |
66556.94 |
15 |
70 |
76502.83 |
53281.11 |
60482.33 |
78581.31 |
60482.33 |
15 |
110 |
42508.20 |
53281.11 |
60482.33 |
66556.94 |
53756.51 |
15 |
170 |
42258.50 |
53281.11 |
60482.33 |
60482.33 |
42383.35 |
20 |
80 |
43504.60 |
53281.11 |
38878.67 |
43504.60 |
38878.67 |
20 |
190 |
34252.75 |
53281.11 |
38878.67 |
38878.67 |
38878.67 |
-
AVG1: An average of all matching rows
-
AVG2: An average of all matching rows within a department.
-
AVG3: An average of matching rows within a department, from the first matching row (ordered by ID), up to and including the current row.
-
AVG4: An average of matching rows within a department, starting with one preceding row (i.e. the highest, ordered by ID), the current row, and the next two following rows.
6.1.3. PARTITION Expression
The PARTITION BY expression, which is optional, defines the set of rows that are used in each OLAP function calculation.
Below is a query that uses different partitions to average sets of rows:
SELECT id
, dept
, job
, years
, salary
, DEC(AVG(salary) OVER(PARTITION BY dept) ,7,2) AS dpt_avg
, DEC(AVG(salary) OVER(PARTITION BY job) ,7,2) AS job_avg
, DEC(AVG(salary) OVER(PARTITION BY years/2) ,7,2) AS yr2_avg
, DEC(AVG(salary) OVER(PARTITION BY dept, job) ,7,2) AS d_j_avg
FROM staff
WHERE dept IN (15,20)
AND id > 20
ORDER BY id;
ANSWER
ID | DEPT | JOB | YEARS | SALARY | DPT_AVG | JOB_AVG | YR2_AVG | D_J_AVG |
---|---|---|---|---|---|---|---|---|
50 |
15 |
Mgr |
10 |
80659.80 |
60482.33 |
80659.80 |
80659.80 |
80659.80 |
70 |
15 |
Sales |
7 |
76502.83 |
60482.33 |
76502.83 |
76502.83 |
76502.83 |
80 |
20 |
Clerk |
- |
43504.60 |
38878.67 |
40631.01 |
43504.60 |
38878.67 |
110 |
15 |
Clerk |
5 |
42508.20 |
60482.33 |
40631.01 |
42383.35 |
42383.35 |
170 |
15 |
Clerk |
4 |
42258.50 |
60482.33 |
40631.01 |
42383.35 |
42383.35 |
190 |
20 |
Clerk |
8 |
34252.75 |
38878.67 |
40631.01 |
34252.75 |
38878.67 |
6.1.4. PARTITION vs. GROUP BY
The PARTITION clause, when used by itself, returns a very similar result to a GROUP BY, except that like all OLAP functions, it does not remove the duplicate rows. To illustrate, below is a simple query that does a GROUP BY:
SELECT dept
, SUM(years) AS sum
, AVG(years) AS avg
, COUNT(*) AS row
FROM staff
WHERE id BETWEEN 40 AND 120
AND years IS NOT NULL
GROUP BY dept;
ANSWER
DEPT | SUM | AVG | ROW |
---|---|---|---|
15 |
22 |
7 |
3 |
38 |
6 |
6 |
1 |
42 |
13 |
6 |
2 |
Below is a similar query that uses a PARTITION phrase. Observe that each value calculated is the same, but duplicate rows have not been removed:
SELECT dept
, SUM(years) OVER(PARTITION BY dept) AS sum
, AVG(years) OVER(PARTITION BY dept) AS avg
, COUNT(*) OVER(PARTITION BY dept) AS row
FROM staff
WHERE id BETWEEN 40 AND 120
AND years IS NOT NULL
ORDER BY dept;
ANSWER
DEPT | SUM | AVG | ROW |
---|---|---|---|
15 |
22 |
7 |
3 |
15 |
22 |
7 |
3 |
15 |
22 |
7 |
3 |
38 |
6 |
6 |
1 |
42 |
13 |
6 |
2 |
42 |
13 |
6 |
2 |
Below is a similar query that uses the PARTITION phrase, and then uses a DISTINCT clause to remove the duplicate rows:
SELECT DISTINCT dept
, SUM(years) OVER(PARTITION BY dept) AS sum
, AVG(years) OVER(PARTITION BY dept) AS avg
, COUNT(*) OVER(PARTITION BY dept) AS row
FROM staff
WHERE id BETWEEN 40 AND 120
AND years IS NOT NULL
ORDER BY dept;
ANSWER
DEPT | SUM | AVG | ROW |
---|---|---|---|
15 |
22 |
7 |
3 |
38 |
6 |
6 |
1 |
42 |
13 |
6 |
2 |
Even though the above statement gives the same answer as the prior GROUP BY example, it is not the same internally. Nor is it (probably) as efficient, and it is certainly not as easy to understand. Therefore, when in doubt, use the GROUP BY syntax. |
6.1.5. Window Definition
An OLAP function works on a "window" of matching rows. This window can be defined as:
-
All matching rows.
-
All matching rows within a partition.
-
Some moving subset of the matching rows (within a partition, if defined).
A moving window has to have an ORDER BY clause so that the set of matching rows can be determined.
Window Size Partitions
-
UNBOUNDED PRECEDING: All of the preceding rows.
-
Number PRECEDING: The "n" preceding rows.
-
UNBOUNDED FOLLOWING: All of the following rows.
-
Number FOLLOWING: The "n" following rows.
-
CURRENT ROW: Only the current row.
Defaults
-
No ORDER BY: UNBOUNDED PRECEDING to UNBOUNDED FOLLOWING.
-
ORDER BY only: UNBOUNDED PRECEDING to CURRENT ROW.
-
No BETWEEN: CURRENT ROW to "n" preceding/following row or rank.
-
BETWEEN stmt: From "n" to "n" preceding/following row or rank. The end-point must be greater than or equal to the starting point.
6.1.6. Sample Queries
Below is a query that illustrates some of the above concepts:
SELECT id
, salary
, DEC(AVG(salary) OVER() ,7,2) AS avg_all
, DEC(AVG(salary) OVER(ORDER BY id) ,7,2) AS avg_odr
, DEC(AVG(salary)
OVER(ORDER BY id
ROWS BETWEEN UNBOUNDED PRECEDING
AND UNBOUNDED FOLLOWING) ,7,2) AS avg_p_f
, DEC(AVG(salary)
OVER(ORDER BY id
ROWS BETWEEN UNBOUNDED PRECEDING
AND CURRENT ROW) ,7,2) AS avg_p_c
, DEC(AVG(salary)
OVER(ORDER BY id
ROWS BETWEEN CURRENT ROW
AND UNBOUNDED FOLLOWING) ,7,2) AS avg_c_f
, DEC(AVG(salary)
OVER(ORDER BY id
ROWS BETWEEN 2 PRECEDING
AND 1 FOLLOWING) ,7,2) AS avg_2_1
FROM staff
WHERE dept IN (15,20)
AND id > 20
ORDER BY id;
ANSWER
ID | SALARY | AVG_ALL | AVG_ODR | AVG_P_F | AVG_P_C | AVG_C_F | AVG_2_1 |
---|---|---|---|---|---|---|---|
50 |
80659.80 |
53281.11 |
80659.80 |
53281.11 |
80659.80 |
53281.11 |
78581.31 |
70 |
76502.83 |
53281.11 |
78581.31 |
53281.11 |
78581.31 |
47805.37 |
66889.07 |
80 |
43504.60 |
53281.11 |
66889.07 |
53281.11 |
66889.07 |
40631.01 |
60793.85 |
110 |
42508.20 |
53281.11 |
60793.85 |
53281.11 |
60793.85 |
39673.15 |
51193.53 |
170 |
42258.50 |
53281.11 |
57086.78 |
53281.11 |
57086.78 |
38255.62 |
40631.01 |
190 |
34252.75 |
53281.11 |
53281.11 |
53281.11 |
53281.11 |
34252.75 |
39673.15 |
When the BETWEEN syntax is used, the start of the range/rows must be less than or equal to the end of the range/rows. |
When no BETWEEN is used, the set of rows to be evaluated goes from the current row up or down to the end value:
SELECT id
, SUM(id) OVER(ORDER BY id) AS sum1
, SUM(id) OVER(ORDER BY id ROWS 1 PRECEDING) AS sum2
, SUM(id) OVER(ORDER BY id ROWS UNBOUNDED PRECEDING) AS sum3
, SUM(id) OVER(ORDER BY id ROWS CURRENT ROW) AS sum4
, SUM(id) OVER(ORDER BY id ROWS2 FOLLOWING) AS sum6
, SUM(id) OVER(ORDER BY id ROWS UNBOUNDED FOLLOWING) AS sum6
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SUM1 | SUM2 | SUM3 | SUM4 | SUM5 | SUM6 |
---|---|---|---|---|---|---|
10 |
10 |
10 |
10 |
10 |
60 |
60 |
20 |
30 |
30 |
30 |
20 |
50 |
50 |
30 |
60 |
50 |
60 |
30 |
30 |
30 |
6.1.7. ROWS vs. RANGE
(OLAP, ROWS A moving window of rows to be evaluated (relative to the current row) can be defined using either the ROW or RANGE expressions. These differ as follows:
-
ROWS: Refers to the "n" rows before and/or after (within the partition), as defined by the ORDER BY.
-
RANGE: Refers to those rows before and/or after (within the partition) that are within an arithmetic range of the current row, as defined by the ORDER BY.
The next query compares the ROW and RANGE expressions:
SELECT id
, SMALLINT(SUM(id)
OVER(ORDER BY id
RANGE BETWEEN 10 PRECEDING AND 10 FOLLOWING)) AS rng1
, SMALLINT(SUM(id)
OVER(ORDER BY id
ROWS BETWEEN 1 PRECEDING AND 1 FOLLOWING)) AS row1
, SMALLINT(SUM(id)
OVER(ORDER BY id
RANGE BETWEEN 10 PRECEDING AND CURRENT ROW)) AS rng2
, SMALLINT(SUM(id)
OVER(ORDER BY id
ROWS BETWEEN 3 PRECEDING AND 1 PRECEDING)) AS row2
, SMALLINT(SUM(id)
OVER(ORDER BY id DESC
ROWS BETWEEN 3 PRECEDING AND 1 PRECEDING)) AS row3
, SMALLINT(SUM(id)
OVER(ORDER BY id
RANGE BETWEEN UNBOUNDED PRECEDING AND 20 FOLLOWING)) AS rng3
FROM staff
WHERE id < 60
ORDER BY id;
ANSWER
ID | RNG1 | ROW1 | RNG2 | ROW2 | ROW3 | RNG3 |
---|---|---|---|---|---|---|
10 |
30 |
30 |
10 |
- |
90 |
60 |
20 |
60 |
60 |
30 |
10 |
120 |
100 |
30 |
90 |
90 |
50 |
30 |
90 |
150 |
40 |
120 |
120 |
70 |
60 |
50 |
150 |
50 |
90 |
90 |
90 |
90 |
- |
150 |
Usage Notes
-
An ORDER BY statement is required when using either expression.
-
If no RANGE or ROWS expression was provided, the default range (assuming there was an ORDER BY) is all preceding rows – up to the current row.
-
When using the RANGE expression, only one expression can be specified in the ORDER BY, and that expression must be numeric.
6.1.8. ORDER BY Expression
The ORDER BY phrase has several purposes:
-
It defines the set of rows that make up a moving window.
-
It provides a set of values to do aggregations on. Each distinct value gets a new result.
-
It gives a direction to the aggregation function processing (i.e. ASC or DESC).
An ORDER BY expression is required for the RANK and DENSE_RANK functions. It is optional for all others (except of using ROWS or RANGE).
Usage Notes
-
ASC: Sorts the values in ascending order. This is the default.
-
DESC: Sorts the values in descending order.
-
NULLS: Determines whether null values are sorted high or low, relative to the non-null values present. Note that the default option differs for ascending and descending order.
-
Sort Expression: The sort-key expression can be any valid column, or any scalar expression is deterministic, and has no external action.
-
ORDER BY ORDER OF table-designator: The table designator refers to a subselect or fullselect in the query and any ordering defined on columns in that subselect or fullselect (note: if there is no explicit ordering the results are unpredictable). If the subselect or fullselect ORDER BY is changed, the ordering sequence will automatically change to match. Note that the final query may have an ordering that differs from that in the subselect or fullselect.
When the table designator refers to a table in the current subselect or fullselect, as opposed to the results of a nested subselect or fullselect, the values are unpredictable. |
6.1.9. Sample Query
In the next query, various aggregations are done on a variety of fields, and on a nested-tableexpression that contains an ORDER BY. Observe that the ascending fields sum or count up, while the descending fields sum down. Also observe that each aggregation field gets a separate result for each new set of rows, as defined in the ORDER BY phrase:
SELECT dept
, name
, salary
, DEC(SUM(salary) OVER(ORDER BY dept) ,8,2) AS sum1
, DEC(SUM(salary) OVER(ORDER BY dept DESC) ,8,2) AS sum2
, DEC(SUM(salary) OVER(ORDER BY ORDER OF s1) ,8,2) AS sum3
, SMALLINT(RANK() OVER(ORDER BY salary, name, dept) ) AS r1
, SMALLINT(RANK() OVER(ORDER BY ORDER OF s1) AS r2
, ROW_NUMBER() OVER(ORDER) AS w1
, COUNT(*) OVER(ORDER BY salary) AS w2
FROM (SELECT *
FROM staff
WHERE id < 60
ORDER BY dept
, name
) AS s1
ORDER BY 1, 2;
ANSWER
DEPT | NAME | SALARY | SUM1 | SUM2 | SUM3 | R1 | R2 | W1 | W2 |
---|---|---|---|---|---|---|---|---|---|
15 |
Hanes |
80659.80 |
80659.80 |
412701.30 |
80659.80 |
4 |
1 |
4 |
4 |
20 |
Pernal |
78171.25 |
257188.55 |
332041.50 |
158831.05 |
3 |
2 |
3 |
3 |
20 |
Sanders |
98357.50 |
257188.55 |
332041.50 |
257188.55 |
5 |
3 |
5 |
5 |
38 |
Marenghi |
77506.75 |
412701.30 |
155512.75 |
334695.30 |
1 |
4 |
1 |
1 |
38 |
O’Brien |
78006.00 |
412701.30 |
155512.75 |
412701.30 |
2 |
5 |
2 |
2 |
There is no relationship between the ORDER BY used in an OLAP function, and the final ordering of the answer. Both are calculated independently. |
6.1.10. Table Designator
The next two queries illustrate referencing a table designator in a subselect. Observe that as the ORDER BY in the subselect changes, the ordering sequence changes. Note that the final query output order does match that of the subselect:
SELECT id
, name
, ROW_NUMBER()
OVER(ORDER BY ORDER OF s) od
FROM (SELECT *
FROM staff
WHERE id < 50
ORDER BY name ASC
) AS s
ORDER BY id ASC;
ANSWER
ID | NAME | OD |
---|---|---|
10 |
Sanders |
4 |
20 |
Pernal |
3 |
30 |
Marenghi |
1 |
40 |
O’Brien |
2 |
SELECT id
, name
, ROW_NUMBER()
OVER(ORDER BY ORDER OF s) od
FROM (SELECT *
FROM staff
WHERE id < 50
ORDER BY name DESC
) AS s
ORDER BY id ASC;
ANSWER
ID | NAME | OD |
---|---|---|
10 |
Sanders |
1 |
20 |
Pernal |
2 |
30 |
Marenghi |
4 |
40 |
O’Brien |
3 |
6.1.11. Nulls Processing
When writing the ORDER BY, one can optionally specify whether or not null values should be counted as high or low. The default, for an ascending field is that they are counted as high (i.e. come last), and for a descending field, that they are counted as low:
SELECT id
, years AS yr
, salary
, DENSE_RANK() OVER(ORDER BY years ASC) AS a
, DENSE_RANK() OVER(ORDER BY years ASC NULLS FIRST) AS af
, DENSE_RANK() OVER(ORDER BY years ASC NULLS LAST ) AS al
, DENSE_RANK() OVER(ORDER BY years DESC) AS d
, DENSE_RANK() OVER(ORDER BY years DESC NULLS FIRST) AS df
, DENSE_RANK() OVER(ORDER BY years DESC NULLS LAST ) AS dl
FROM staff
WHERE id < 100
ORDER BY years
, salary;
ANSWER
ID | YR | SALARY | A | AF | AL | D | DF | DL |
---|---|---|---|---|---|---|---|---|
30 |
5 |
77506.75 |
1 |
2 |
1 |
6 |
6 |
5 |
90 |
6 |
38001.75 |
2 |
3 |
2 |
5 |
5 |
4 |
40 |
6 |
78006.00 |
2 |
3 |
2 |
5 |
5 |
4 |
70 |
7 |
76502.83 |
3 |
4 |
3 |
4 |
4 |
3 |
10 |
7 |
98357.50 |
3 |
4 |
3 |
4 |
4 |
3 |
20 |
8 |
78171.25 |
4 |
5 |
4 |
3 |
3 |
2 |
50 |
10 |
80659.80 |
5 |
6 |
5 |
2 |
2 |
1 |
80 |
- |
43504.60 |
6 |
1 |
6 |
1 |
1 |
6 |
60 |
- |
66808.30 |
6 |
1 |
6 |
1 |
1 |
6 |
In general, one null value does not equal another null value. But, as is illustrated above, for purposes of assigning rank, all null values are considered equal. |
6.1.12. Counting Nulls
The DENSE RANK and RANK functions include null values when calculating rankings. By contrast the COUNT DISTINCT statement excludes null values when counting values. Thus, as is illustrated below, the two methods will differ (by one) when they are used get a count of distinct values - if there are nulls in the target data:
SELECT COUNT(DISTINCT years) AS y#1
, MAX(y#) AS y#2
FROM (SELECT years
, DENSE_RANK() OVER(ORDER BY years) AS y#
FROM staff
WHERE id < 100
) AS xxx
ORDER BY 1;
ANSWER
Y#1 | Y#2 |
---|---|
5 |
6 |
6.1.13. OLAP Functions
RANK and DENSE_RANK
The RANK and DENSE_RANK functions enable one to rank the rows returned by a query. The result is of type BIGINT.
The ORDER BY phrase, which is required, is used to both sequence the values, and to tell Db2 when to generate a new value. |
RANK vs. DENSE_RANK
The two functions differ in how they handle multiple rows with the same value:
-
The RANK function returns the number of proceeding rows, plus one. If multiple rows have equal values, they all get the same rank, while subsequent rows get a ranking that counts all of the prior rows. Thus, there may be gaps in the ranking sequence.
-
The DENSE_RANK function returns the number of proceeding distinct values, plus one. If multiple rows have equal values, they all get the same rank. Each change in data value causes the ranking number to be incremented by one.
Usage Notes
-
The ORDER BY expression is mandatory.
-
The PARTITION BY expression is optional.
Compare Functions
The following query illustrates the use of the two functions:
SELECT id
, years
, salary
, RANK()
OVER(ORDER BY years) AS rank#
, DENSE_RANK()
OVER(ORDER BY years) AS dense#
, ROW_NUMBER()
OVER(ORDER BY years) AS row#
FROM staff
WHERE id < 100
AND years < 10
ORDER BY years;
ANSWER
ID | YEARS | SALARY | RANK# | DENSE# | ROW# |
---|---|---|---|---|---|
30 |
5 |
77506.75 |
1 |
1 |
1 |
40 |
6 |
78006.00 |
2 |
2 |
2 |
90 |
6 |
38001.75 |
2 |
2 |
3 |
10 |
7 |
98357.50 |
4 |
3 |
4 |
70 |
7 |
76502.83 |
4 |
3 |
5 |
20 |
8 |
78171.25 |
6 |
4 |
6 |
6.1.14. ORDER BY Usage
The mandatory ORDER BY phrase gives a sequence to the ranking, and also tells Db2 when to start a new rank value. The following query illustrates both uses:
SELECT job AS job
, years AS yr
, id AS id
, name AS name
, RANK() OVER(ORDER BY job ASC ) AS a1
, RANK() OVER(ORDER BY job ASC, years ASC) AS a2
, RANK() OVER(ORDER BY job ASC, years ASC ,id ASC ) AS a3
, RANK() OVER(ORDER BY job DESC) AS d1
, RANK() OVER(ORDER BY job DESC, years DESC) AS d2
, RANK() OVER(ORDER BY job DESC, years DESC, id DESC) AS d3
, RANK() OVER(ORDER BY job ASC, years DESC, id ASC ) AS m1
, RANK() OVER(ORDER BY job DESC, years ASC, id DESC) AS m2
FROM staff
WHERE id < 150
AND years IN (6,7)
AND job > 'L'
ORDER BY job
, years
, id;
ANSWER
JOB | YR | ID | NAME | A1 | A2 | A3 | D1 | D2 | D3 | M1 | M2 |
---|---|---|---|---|---|---|---|---|---|---|---|
Mgr |
6 |
140 |
Fraye |
1 |
1 |
1 |
4 |
6 |
6 |
3 |
4 |
Mgr |
7 |
10 |
Sanders |
1 |
2 |
2 |
4 |
4 |
5 |
1 |
6 |
Mgr |
7 |
100 |
Plotz |
1 |
2 |
3 |
4 |
4 |
4 |
2 |
5 |
Sales |
6 |
40 |
O’Brien |
4 |
4 |
4 |
1 |
2 |
3 |
5 |
2 |
Sales |
6 |
90 |
Koonitz |
4 |
4 |
5 |
1 |
2 |
2 |
6 |
1 |
Sales |
7 |
70 |
Rothman |
4 |
6 |
6 |
1 |
1 |
1 |
4 |
3 |
Observe above that adding more fields to the ORDER BY phrase resulted in more ranking values being generated.
6.1.15. PARTITION Usage
The optional PARTITION phrase lets one rank the data by subsets of the rows returned. In the following example, the rows are ranked by salary within year:
SELECT id
, years AS yr
, salary
, RANK() OVER(PARTITION BY years
ORDER BY salary) AS r1
FROM staff
WHERE id < 80
AND years IS NOT NULL
ORDER BY years
, salary;
ANSWER
ID | YR | SALARY | R1 |
---|---|---|---|
30 |
5 |
77506.75 |
1 |
40 |
6 |
78006.00 |
1 |
70 |
7 |
76502.83 |
1 |
10 |
7 |
98357.50 |
2 |
20 |
8 |
78171.25 |
1 |
50 |
0 |
80659.80 |
1 |
6.1.16. Multiple Rankings
One can do multiple independent rankings in the same query:
SELECT id
, years
, salary
, SMALLINT(RANK() OVER(ORDER BY years ASC)) AS rank_a
, SMALLINT(RANK() OVER(ORDER BY years DESC)) AS rank_d
, SMALLINT(RANK() OVER(ORDER BY id, years)) AS rank_iy
FROM staff
WHERE id < 100
AND years IS NOT NULL
ORDER BY years;
6.1.17. Dumb Rankings
If one wants to, one can do some really dumb rankings. All of the examples below are fairly stupid, but arguably the dumbest of the lot is the last. In this case, the "ORDER BY 1" phrase ranks the rows returned by the constant "one", so every row gets the same rank. By contrast the "ORDER BY 1" phrase at the bottom of the query sequences the rows, and so has valid business meaning:
SELECT id
, years
, name
, salary
, SMALLINT(RANK() OVER(ORDER BY SUBSTR(name,3,2))) AS dumb1
, SMALLINT(RANK() OVER(ORDER BY salary / 1000)) AS dumb2
, SMALLINT(RANK() OVER(ORDER BY years * ID)) AS dumb3
, SMALLINT(RANK() OVER(ORDER BY 1)) AS dumb4
FROM staff
WHERE id < 40
AND years IS NOT NULL
ORDER BY 1;
ID | YEARS | NAME | SALARY | DUMB1 | DUMB2 | DUMB3 | DUMB4 |
---|---|---|---|---|---|---|---|
10 |
7 |
Sanders |
98357.50 |
1 |
3 |
1 |
1 |
20 |
8 |
Pernal |
78171.25 |
3 |
2 |
3 |
1 |
30 |
5 |
Marenghi |
77506.75 |
2 |
1 |
2 |
1 |
6.1.18. Subsequent Processing
The ranking function gets the rank of the value as of when the function was applied. Subsequent processing may mean that the rank no longer makes sense. To illustrate this point, the following query ranks the same field twice. Between the two ranking calls, some rows were removed from the answer set, which has caused the ranking results to differ:
SELECT xxx.*
, RANK()OVER(ORDER BY id) AS r2
FROM (SELECT id
, name
, RANK() OVER(ORDER BY id) AS r1
FROM staff
WHERE id < 100
AND years IS NOT NULL
) AS xxx
WHERE id > 30
ORDER BY id;
ANSWER
ID | NAME | R1 | R2 |
---|---|---|---|
40 |
O’Brien |
4 |
1 |
50 |
Hanes |
5 |
2 |
70 |
Rothman |
6 |
3 |
90 |
Koonitz |
7 |
4 |
6.1.19. Ordering Rows by Rank
One can order the rows based on the output of a ranking function. This can let one sequence the data in ways that might be quite difficult to do using ordinary SQL. For example, in the following query the matching rows are ordered so that all those staff with the highest salary in their respective department come first, followed by those with the second highest salary, and so on. Within each ranking value, the person with the highest overall salary is listed first:
SELECT id
, RANK() OVER(PARTITION BY dept
ORDER BY salary DESC) AS r1
, salary
, dept AS dp
FROM staff
WHERE id < 80
AND years IS NOT NULL
ORDER BY r1 ASC
, salary DESC;
ANSWER
ID | R1 | SALARY | DP |
---|---|---|---|
10 |
1 |
98357.50 |
20 |
50 |
1 |
80659.80 |
15 |
40 |
1 |
78006.00 |
38 |
20 |
2 |
78171.25 |
20 |
30 |
2 |
77506.75 |
38 |
70 |
2 |
76502.83 |
15 |
Here is the same query, written without the ranking function:
SELECT id
, (SELECT COUNT(*)
FROM staff s2
WHERE s2.id < 80
AND s2.years IS NOT NULL
AND s2.dept = s1.dept
AND s2.salary >= s1.salary
) AS R1
, salary
, dept AS dp
FROM staff s1
WHERE id < 80
AND years IS NOT NULL
ORDER BY r1 ASC
, salary DESC;
ANSWER
ID | R1 | SALARY | DP |
---|---|---|---|
10 |
1 |
98357.50 |
20 |
50 |
1 |
80659.80 |
15 |
40 |
1 |
78006.00 |
38 |
20 |
2 |
78171.25 |
20 |
30 |
2 |
77506.75 |
38 |
70 |
2 |
76502.83 |
15 |
The above query has all of the failings that were discussed at the beginning of this chapter:
-
The nested table expression has to repeat all of the predicates in the main query, and have predicates that define the ordering sequence. Thus it is hard to read.
-
The nested table expression will (inefficiently) join every matching row to all prior rows.
6.1.20. Selecting the Highest Value
The ranking functions can also be used to retrieve the row with the highest value in a set of rows. To do this, one must first generate the ranking in a nested table expression, and then query the derived field later in the query. The following statement illustrates this concept by getting the person, or persons, in each department with the highest salary:
SELECT id
, salary
, dept AS dp
FROM
(SELECT s1.*
, RANK() OVER(PARTITION BY dept
ORDER BY salary DESC) AS r1
FROM staff s1
WHERE id < 80
AND years IS NOT NULL
) AS xxx
WHERE r1 = 1
ORDER BY dp;
ANSWER
ID | SALARY | DP |
---|---|---|
50 |
80659.80 |
15 |
10 |
98357.50 |
20 |
40 |
78006.00 |
38 |
Here is the same query, written using a correlated sub-query:
SELECT id
, salary
, dept AS dp
FROM staff s1
WHERE id < 80
AND years IS NOT NULL
AND NOT EXISTS
(SELECT *
FROM staff s2
WHERE s2.id < 80
AND s2.years IS NOT NULL
AND s2.dept = s1.dept
AND s2.salary > s1.salary)
ORDER BY dp;
ID | SALARY | DP |
---|---|---|
50 |
80659.80 |
15 |
10 |
98357.50 |
20 |
40 |
78006.00 |
38 |
Here is the same query, written using an uncorrelated sub-query:
SELECT id
, salary
, dept AS dp
FROM staff
WHERE id < 80
AND years IS NOT NULL
AND (dept, salary) IN
(SELECT dept, MAX(salary)
FROM staff
WHERE id < 80
AND years IS NOT NULL
GROUP BY dept)
ORDER BY dp;
ANSWER
ID | SALARY | DP |
---|---|---|
50 |
80659.80 |
15 |
10 |
98357.50 |
20 |
40 |
78006.00 |
38 |
Arguably, the first query above (i.e. the one using the RANK function) is the most elegant of the series because it is the only statement where the basic predicates that define what rows match are written once. With the two sub-query examples, these predicates have to be repeated, which can often lead to errors.
6.1.21. ROW_NUMBER
The ROW_NUMBER function lets one number the rows being returned. The result is of type BIGINT. A syntax diagram follows. Observe that unlike with the ranking functions, the ORDER BY is not required.
ORDER BY Usage
You don’t have to provide an ORDER BY when using the ROW_NUMBER function, but not doing so can be considered to be either brave or foolish, depending on one’s outlook on life. To illustrate this issue, consider the following query:
SELECT id
, name
, ROW_NUMBER() OVER() AS r1
, ROW_NUMBER() OVER(ORDER BY id) AS r2
FROM staff
WHERE id < 50
AND years IS NOT NULL
ORDER BY id;
ANSWER
ID | NAME | R1 | R2 |
---|---|---|---|
10 |
Sanders |
1 |
1 |
20 |
Pernal |
2 |
2 |
30 |
Marenghi |
3 |
3 |
40 |
O’Brien |
4 |
4 |
In the above example, both ROW_NUMBER functions return the same set of values, which happen to correspond to the sequence in which the rows are returned. In the next query, the second ROW_NUMBER function purposely uses another sequence:
SELECT id
, name
, ROW_NUMBER() OVER() AS r1
, ROW_NUMBER() OVER(ORDER BY name) AS r2
FROM staff
WHERE id < 50
AND years IS NOT NULL
ORDER BY id;
ANSWER
ID | NAME | R1 | R2 |
---|---|---|---|
10 |
Sanders |
4 |
4 |
20 |
Pernal |
3 |
3 |
30 |
Marenghi |
1 |
1 |
40 |
O’Brien |
2 |
2 |
Observe that changing the second function has had an impact on the first. Now lets see what happens when we add another ROW_NUMBER function:
SELECT id
,name
,ROW_NUMBER() OVER() AS r1
,ROW_NUMBER() OVER(ORDER BY ID) AS r2
,ROW_NUMBER() OVER(ORDER BY NAME) AS r3
FROM staff
WHERE id < 50
AND years IS NOT NULL
ORDER BY id;
ANSWER
ID | NAME | R1 | R2 | R3 |
---|---|---|---|---|
10 |
Sanders |
1 |
1 |
4 |
20 |
Pernal |
2 |
2 |
3 |
30 |
Marenghi |
3 |
3 |
1 |
40 |
O’Brien |
4 |
4 |
2 |
Observe that now the first function has reverted back to the original sequence.
When not given an explicit ORDER BY, the ROW_NUMBER function, may create a value in any odd order. Usually, the sequence will reflect the order in which the rows are returned - but not always. |
6.1.22. PARTITION Usage
The PARTITION phrase lets one number the matching rows by subsets of the rows returned. In the following example, the rows are both ranked and numbered within each JOB:
SELECT job
, years
, id
, name
, ROW_NUMBER() OVER(PARTITION BY job ORDER BY years) AS row#
, RANK() OVER(PARTITION BY job ORDER BY years) AS rn1#
, DENSE_RANK() OVER(PARTITION BY job ORDER BY years) AS rn2#
FROM staff
WHERE id < 150
AND years IN (6,7)
AND job > 'L'
ORDER BY job,years;
ANSWER
JOB | YEARS | ID | NAME | ROW# | RN1# | RN2# |
---|---|---|---|---|---|---|
Mgr |
6 |
140 |
Fraye |
1 |
1 |
1 |
Mgr |
7 |
10 |
Sanders |
2 |
2 |
2 |
Mgr |
7 |
100 |
Plotz |
3 |
2 |
2 |
Sales |
6 |
40 |
O’Brien |
1 |
1 |
1 |
Sales |
6 |
90 |
Koonitz |
2 |
1 |
1 |
Sales |
7 |
70 |
Rothman |
3 |
3 |
2 |
One problem with the above query is that the final ORDER BY that sequences the rows does not identify a unique field (e.g. ID). Consequently, the rows can be returned in any sequence within a given JOB and YEAR. Because the ORDER BY in the ROW_NUMBER function also fails to identify a unique row, this means that there is no guarantee that a particular row will always give the same row number. For consistent results, ensure that both the ORDER BY phrase in the function call, and at the end of the query, identify a unique row. And to always get the rows returned in the desired row-number sequence, these phrases must be equal.
Selecting "n" Rows
To query the output of the ROW_NUMBER function, one has to make a nested temporary table that contains the function expression. In the following example, this technique is used to limit the query to the first three matching rows:
SELECT *
FROM
(SELECT id
, name
, ROW_NUMBER() OVER(ORDER BY id) AS r
FROM staff
WHERE id < 100
AND years IS NOT NULL
) AS xxx
WHERE r <= 3
ORDER BY id;
ANSWER
ID | NAME | R |
---|---|---|
10 |
Sanders |
1 |
20 |
Pernal |
2 |
30 |
Marenghi |
3 |
In the next query, the FETCH FIRST "n" ROWS notation is used to achieve the same result:
SELECT id
, name
, ROW_NUMBER() OVER(ORDER BY id) AS r
FROM staff
WHERE id < 100
AND years IS NOT NULL
ORDER BY id
FETCH FIRST 3 ROWS ONLY;
ANSWER
ID | NAME | R |
---|---|---|
10 |
Sanders |
1 |
20 |
Pernal |
2 |
30 |
Marenghi |
3 |
So far, the ROW_NUMBER and the FETCH FIRST notations seem to be about the same. But the former is much more flexible. To illustrate, the next query gets the 3rd through 6th rows:
SELECT *
FROM
(SELECT id
, name
, ROW_NUMBER() OVER(ORDER BY id) AS r
FROM staff
WHERE id < 200
AND years IS NOT NULL
) AS xxx
WHERE r BETWEEN 3 AND 6
ORDER BY id;
ANSWER
ID | NAME | R |
---|---|---|
30 |
Marenghi |
3 |
40 |
O’Brien |
4 |
50 |
Hanes |
5 |
70 |
Rothman |
6 |
In the next query we get every 5th matching row - starting with the first:
SELECT *
FROM
(SELECT id
, name
, ROW_NUMBER() OVER(ORDER BY id) AS r
FROM staff
WHERE id < 200
AND years IS NOT NULL
) AS xxx
WHERE (r - 1) = ((r - 1) / 5) * 5
ORDER BY id;
ANSWER
ID | NAME | R |
---|---|---|
10 |
Sanders |
1 |
70 |
Rothman |
6 |
140 |
Fraye |
11 |
190 |
Sneider |
16 |
In the next query we get the last two matching rows:
SELECT *
FROM
(SELECT id
, name
, ROW_NUMBER() OVER(ORDER BY id DESC) AS r
FROM staff
WHERE id < 200
AND years IS NOT NULL
) AS xxx
WHERE r <= 2
ORDER BY id;
ANSWER
ID | NAME | R |
---|---|---|
180 |
Abrahams |
2 |
190 |
Sneider |
1 |
6.1.23. Selecting "n" or more Rows
Imagine that one wants to fetch the first "n" rows in a query. This is easy to do, and has been illustrated above. But imagine that one also wants to keep on fetching if the following rows have the same value as the "nth". In the next example, we will get the first three matching rows in the STAFF table, ordered by years of service. However, if the 4th row, or any of the following rows, has the same YEAR as the 3rd row, then we also want to fetch them.
The query logic goes as follows:
-
Select every matching row in the STAFF table, and give them all both a row-number and a ranking value. Both values are assigned according to the order of the final output. Do all of this work in a nested table expression.
-
Select from the nested table expression where the rank is three or less.
The query relies on the fact that the RANK function (see RANK and DENSE_RANK) assigns the lowest common row number to each row with the same ranking:
SELECT *
FROM
(SELECT years
, id
, name
, RANK() OVER(ORDER BY years) AS rnk
, ROW_NUMBER() OVER(ORDER BY years, id) AS row
FROM staff
WHERE id < 200
AND years IS NOT NULL
) AS xxx
WHERE rnk <= 3
ORDER BY years
, id;
ANSWER
YEARS | ID | NAME | RNK | ROW |
---|---|---|---|---|
3 |
180 |
Abrahams |
1 |
1 |
4 |
170 |
Kermisch |
2 |
2 |
5 |
30 |
Marenghi |
3 |
3 |
5 |
110 |
Ngan |
3 |
4 |
The type of query illustrated above can be extremely useful in certain business situations. To illustrate, imagine that one wants to give a reward to the three employees that have worked for the company the longest. Stopping the query that lists the lucky winners after three rows are fetched can get one into a lot of trouble if it happens that there are more than three employees that have worked for the company for the same number of years.
6.1.24. Selecting "n" Rows - Efficiently
Sometimes, one only wants to fetch the first "n" rows, where "n" is small, but the number of matching rows is extremely large. In this section, we will discus how to obtain these "n" rows efficiently, which means that we will try to fetch just them without having to process any of the many other matching rows. Below is an invoice table. Observe that we have defined the INV# field as the primary key, which means that Db2 will build a unique index on this column:
CREATE TABLE invoice
( inv# INTEGER NOT NULL
, customer# INTEGER NOT NULL
, sale_date DATE NOT NULL
, sale_value DECIMAL(9,2) NOT NULL
, CONSTRAINT ctx1 PRIMARY KEY (inv#)
, CONSTRAINT ctx2 CHECK(inv# >= 0));
The next SQL statement will insert 1,000,000 rows into the above table. After the rows are inserted a REORG and RUNSTATS is run, so the optimizer can choose the best access path.
INSERT INTO invoice
WITH temp (n, m) AS
(VALUES
(INTEGER(0), RAND(1))
UNION ALL
SELECT n+1, RAND()
FROM temp
WHERE n+1 < 1000000
)
SELECT n AS inv#
, INT(m * 1000) AS customer#
, DATE('2000-11-01') + (m*40) DAYS AS sale_date
, DECIMAL((m * m * 100),8,2) AS sale_value
FROM temp;
Imagine we want to retrieve the first five rows (only) from the above table. Below are several queries that get this result. For each query, the elapsed time, as measured by Db2BATCH, is provided. Below we use the "FETCH FIRST n ROWS" notation to stop the query at the 5th row. The query scans the primary index to get first five matching rows, and thus is cheap:
SELECT s.*
FROM invoice s
ORDER BY inv#
FETCH FIRST 5 ROWS ONLY;
The next query is essentially the same as the prior, but this time we tell Db2 to optimize the query for fetching five rows. Nothing has changed, and all is good:
SELECT s.*
FROM invoice s
ORDER BY inv#
FETCH FIRST 5 ROWS ONLY
OPTIMIZE FOR 5 ROWS;
The next query is the same as the first, except that it invokes the ROW_NUMBER function to passively sequence the output. This query also uses the primary index to identify the first five matching rows, and so is cheap:
SELECT s.*
, ROW_NUMBER() OVER() AS row#
FROM invoice s
ORDER BY inv#
FETCH FIRST 5 ROWS ONLY;
The next query is the same as the previous. It uses a nested-table-expression, but no action is taken subsequently, so this code is ignored:
SELECT *
FROM
(SELECT s.*
, ROW_NUMBER() OVER() AS row#
FROM invoice s
) xxx
ORDER BY inv#
FETCH FIRST 5 ROWS ONLY;
All of the above queries processed only five matching rows. The next query will process all one million matching rows in order to calculate the ROW_NUMBER value, which is on no particular column. It will cost:
SELECT *
FROM
(SELECT s.*
, ROW_NUMBER() OVER() AS row#
FROM invoice s
) xxx
WHERE row# <= 5
ORDER BY inv#;
In the above query the "OVER()" phrase told Db2 to assign row numbers to each row. In the next query we explicitly provide the ROW_NUMBER with a target column, which happens to be the same at the ORDER BY sequence, and is also an indexed column. Db2 can use all this information to confine the query to the first "n" matching rows:
SELECT *
FROM
(SELECT s.*
, ROW_NUMBER() OVER(ORDER BY inv#) AS row#
FROM invoice s
) xxx
WHERE row# <= 5
ORDER BY inv#;
Changing the above predicate to: "WHERE row# BETWEEN 1 AND 5" will get the same answer, but use a much less efficient access path. |
One can also use recursion to get the first "n" rows. One begins by getting the first matching row, and then uses that row to get the next, and then the next, and so on (in a recursive join), until the required number of rows have been obtained. In the following example, we start by getting the row with the MIN invoice-number. This row is then joined to the row with the next to lowest invoice-number, which is then joined to the next, and so on. After five such joins, the cycle is stopped and the result is selected:
WITH temp (inv#, c#, sd, sv, n) AS
(SELECT inv.*
, 1
FROM invoice inv
WHERE inv# =
(SELECT MIN(inv#)
FROM invoice)
UNION ALL
SELECT new.*
, n + 1
FROM temp old
, invoice new
WHERE old.inv# < new.inv#
AND old.n < 5
AND new.inv# =
(SELECT MIN(xxx.inv#)
FROM invoice xxx
WHERE xxx.inv# > old.inv#)
)
SELECT *
FROM temp;
The above technique is nice to know, but it has several major disadvantages:
-
It is not exactly easy to understand.
-
It requires that all primary predicates (e.g. get only those rows where the sale-value is greater than $10,000) be repeated four times. In the above example there are none, which is unusual in the real world.
-
It quickly becomes both very complicated and quite inefficient when the sequencing value is made up of multiple fields. In the above example, we sequenced by the INV# column, but imagine if we had used the sale-date, sale-value, and customer-number.
-
It is extremely vulnerable to inefficient access paths. For example, if instead of joining from one (indexed) invoice-number to the next, we joined from one (non-indexed) customer-number to the next, the query would run forever.
In this section we have illustrated how minor changes to the SQL syntax can cause major changes in query performance. But to illustrate this phenomenon, we used a set of queries with 1,000,000 matching rows. In situations where there are far fewer matching rows, one can reasonably assume that this problem is not an issue.
6.1.25. FIRST_VALUE and LAST_VALUE
The FIRST_VALUE and LAST_VALUE functions get first or last value in the (moving) window of matching rows.
Usage Notes
-
An expression value must be provided in the first set of parenthesis. Usually this will be a column name, but any valid scalar expression is acceptable.
-
The PARTITION BY expression is optional.
-
The ORDER BY expression is optional.
-
See Window Definition for notes on how to define a moving-window of rows to process.
-
If no explicit moving-window definition is provided, the default window size is between UNBOUNDED PRECEDING (of the partition and/or range) and the CURRENT ROW. This can sometimes cause logic errors when using the LAST_VALUE function. The last value is often simply the current row. To get the last matching value within the partition and/or range, set the upper bound to UNBOUNDED FOLLOWING.
-
If IGNORE NULLS is specified, null values are ignored, unless all values are null, in which case the result is null. The default is RESPECT NULLS.
Examples
The following query illustrates the basics. The first matching name (in ID order) within each department is obtained:
SELECT dept
, id
, name
, FIRST_VALUE(name) OVER(PARTITION BY dept
ORDER BY id) AS frst
FROM staff
WHERE dept <= 15
AND id > 160
ORDER BY dept ,id;
ANSWER
DEPT | ID | NAME | FRST |
---|---|---|---|
10 |
210 |
Lu |
Lu |
10 |
240 |
Daniels |
Lu |
10 |
260 |
Jones |
Lu |
15 |
170 |
Kermisch |
Kermisch |
The next uses various ordering schemas and moving-window sizes the get a particular first or last value (within a department):
SELECT dept
, id
, comm
, FIRST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm) AS first1
, FIRST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm NULLS FIRST) AS first2
, FIRST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm NULLS LAST) AS first3
, FIRST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm NULLS LAST
ROWS BETWEEN 1 PRECEDING AND CURRENT ROW) AS first4
, LAST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm) AS last1
, LAST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm NULLS FIRST
ROWS UNBOUNDED FOLLOWING) AS last2
FROM
staff
WHERE id < 100
AND dept < 30
ORDER BY dept ,comm;
ANSWER
DEPT | ID | COMM | FIRST1 | FIRST2 | FIRST3 | FIRST4 | LAST1 | LAST2 |
---|---|---|---|---|---|---|---|---|
15 |
70 |
1152.00 |
1152.00 |
- |
1152.00 |
1152.00 |
1152.00 |
1152.00 |
15 |
50 |
- |
1152.00 |
- |
1152.00 |
1152.00 |
- |
1152.00 |
20 |
80 |
128.20 |
128.20 |
- |
128.20 |
128.20 |
128.20 |
612.45 |
20 |
20 |
612.45 |
128.20 |
- |
128.20 |
128.20 |
612.45 |
612.45 |
20 |
10 |
- |
128.20 |
- |
128.20 |
612.45 |
- |
612.45 |
The next query illustrates what happens when one, or all, of the matching values are null:
SELECT dept
, id
, comm
, FIRST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm) AS rn_lst
, FIRST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm NULLS LAST) AS rn_ls2
, FIRST_VALUE(comm) OVER(PARTITION BY dept
ORDER BY comm NULLS FIRST) AS rn_fst
, FIRST_VALUE(comm,'IGNORE NULLS') OVER(PARTITION BY dept
ORDER BY comm NULLS FIRST) AS in_fst
FROM staff
WHERE id BETWEEN 20 AND 160
AND dept <= 20
ORDER BY dept ,comm;
ANSWER
DEPT | ID | COMM | RN_LST | RN_LS2 | RN_FST | IN_FST |
---|---|---|---|---|---|---|
10 |
160 |
|||||
15 |
110 |
206.60 |
206.60 |
206.60 |
- |
206.60 |
15 |
70 |
1152.00 |
206.60 |
206.60 |
- |
206.60 |
15 |
50 |
- |
206.60 |
206.60 |
||
20 |
80 |
128.20 |
128.20 |
128.20 |
128.20 |
128.20 |
20 |
20 |
612.45 |
128.20 |
128.20 |
128.20 |
128.20 |
6.1.26. LAG and LEAD
The LAG, and LEAD functions get the previous or next value from the (moving) window of matching rows:
-
LAG: Get previous value. Return null if at first value.
-
LEAD: Get next value. Return null if at last value.
Usage Notes
-
An expression value must be provided in the first set of parenthesis. Usually this will be a column name, but any valid scalar expression is acceptable.
-
The PARTITION BY expression is optional.
-
The ORDER BY expression is mandatory.
-
The default OFFSET value is 1. A value of zero refers to the current row. An offset that is outside of the moving-window returns null.
-
If IGNORE NULLS is specified, a default (override) value must also be provided.
Examples
The next query uses the LAG function to illustrate what happens when one messes around with the ORDER BY expression:
SELECT dept
, id
, comm
, LAG(comm) OVER(PARTITION BY dept ORDER BY comm) AS lag1
, LAG(comm,0) OVER(PARTITION BY dept ORDER BY comm) AS lag2
, LAG(comm,2) OVER(PARTITION BY dept ORDER BY comm) AS lag3
, LAG(comm,1,-1,'IGNORE NULLS') OVER(PARTITION BY dept ORDER BY comm) AS lag4
, LEAD(comm) OVER(PARTITION BY dept ORDER BY comm) AS led1
FROM staff
WHERE id BETWEEN 20 AND 160
AND dept <= 20
ORDER BY dept ,comm;
ANSWER
DEPT | ID | COMM | LAG1 | LAG2 | LAG3 | LAG4 | LED1 |
---|---|---|---|---|---|---|---|
10 |
160 |
- |
- |
- |
- |
-1.00 |
- |
15 |
110 |
206.60 |
- |
206.60 |
- |
-1.00 |
1152.00 |
15 |
70 |
1152.00 |
206.60 |
1152.00 |
- |
206.60 |
- |
15 |
50 |
- |
1152.00 |
- |
206.60 |
1152.00 |
- |
20 |
80 |
128.20 |
- |
128.20 |
- |
-1.00 |
612.45 |
20 |
20 |
612.45 |
128.20 |
612.45 |
- |
128.20 |
- |
6.1.27. Aggregation
The various aggregation functions let one do cute things like get cumulative totals or running averages. In some ways, they can be considered to be extensions of the existing Db2 column functions. The output type is dependent upon the input type.
Syntax Notes
Guess what - this is a complicated function. Be aware of the following:
-
Any Db2 column function (e.g. AVG, SUM, COUNT), except ARRAY_AGG, can use the aggregation function.
-
The OVER() usage aggregates all of the matching rows. This is equivalent to getting the current row, and also applying a column function (e.g. MAX, SUM) against all of the matching rows.
-
The PARTITION BY expression is optional.
-
The ORDER BY expression is mandatory if the aggregation is confined to a set of rows or range of values. Otherwise it is optional. If a RANGE is specified (see ROWS vs. RANGE for definition), then the ORDER BY expression must be a single value that allows subtraction.
-
If an ORDER BY phrase is provided, but neither a RANGE nor ROWS is specified, then the aggregation is done from the first row to the current row.
-
See Window Definition for notes on how to define a moving-window of rows to process.
Basic Usage
In its simplest form, with just an "OVER()" phrase, an aggregation function works on all of the matching rows, running the column function specified. Thus, one gets both the detailed data, plus the SUM, or AVG, or whatever, of all the matching rows. In the following example, five rows are selected from the STAFF table. Along with various detailed fields, the query also gets sum summary data about the matching rows:
SELECT id
, name
, salary
, SUM(salary) OVER() AS sum_sal
, AVG(salary) OVER() AS avg_sal
, MIN(salary) OVER() AS min_sal
, MAX(salary) OVER() AS max_sal
, COUNT(*) OVER() AS #rows
FROM staff
WHERE id < 30
ORDER BY id;
ANSWER
ID | NAME | SALARY | SUM_SAL | AVG_SAL | MIN_SAL | MAX_SAL | #ROWS |
---|---|---|---|---|---|---|---|
10 |
Sanders |
98357.50 |
254035.50 |
84678.50 |
77506.75 |
98357.50 |
3 |
20 |
Pernal |
78171.25 |
254035.50 |
84678.50 |
77506.75 |
98357.50 |
3 |
30 |
Marenghi |
77506.75 |
254035.50 |
84678.50 |
77506.75 |
98357.50 |
3 |
An aggregation function with just an "OVER()" phrase is logically equivalent to one that has an ORDER BY on a field that has the same value for all matching rows. To illustrate, in the following query, the four aggregation functions are all logically equivalent:
SELECT id
, name
, salary
, SUM(salary) OVER() AS sum1
, SUM(salary) OVER(ORDER BY id * 0) AS sum2
, SUM(salary) OVER(ORDER BY 'ABC') AS sum3
, SUM(salary) OVER(ORDER BY 'ABC'
RANGE BETWEEN UNBOUNDED PRECEDING
AND UNBOUNDED FOLLOWING) AS sum4
FROM staff
WHERE id < 60
ORDER BY id;
ANSWER
ID | NAME | SALARY | SUM1 | SUM2 | SUM3 | SUM4 |
---|---|---|---|---|---|---|
10 |
Sanders |
98357.50 |
412701.30 |
412701.30 |
412701.30 |
412701.30 |
20 |
Pernal |
78171.25 |
412701.30 |
412701.30 |
412701.30 |
412701.30 |
30 |
Marenghi |
77506.75 |
412701.30 |
412701.30 |
412701.30 |
412701.30 |
40 |
O’Brien |
78006.00 |
412701.30 |
412701.30 |
412701.30 |
412701.30 |
50 |
Hanes |
80659.80 |
412701.30 |
412701.30 |
412701.30 |
412701.30 |
6.1.28. ORDER BY Usage
The ORDER BY phrase (see Order By, Group By, and Having for syntax) has two main purposes:
-
It provides a set of values to do aggregations on. Each distinct value gets a new result.
-
It gives a direction to the aggregation function processing (i.e. ASC or DESC).
In the next query, various aggregations are run on the DEPT field, which is not unique, and on the DEPT and NAME fields combined, which are unique (for these rows). Both ascending and descending aggregations are illustrated. Observe that the ascending fields sum or count up, while the descending fields sum down. Also observe that each aggregation field gets a separate result for each new set of rows, as defined in the ORDER BY phrase:
SELECT dept
, name
, salary
, SUM(salary) OVER(ORDER BY dept) AS sum1
, SUM(salary) OVER(ORDER BY dept DESC) AS sum2
, SUM(salary) OVER(ORDER BY dept, NAME) AS sum3
, SUM(salary) OVER(ORDER BY dept DESC, name DESC) AS sum4
, COUNT(*) OVER(ORDER BY dept) AS rw1
, COUNT(*) OVER(ORDER BY dept, NAME) AS rw2
FROM staff
WHERE id < 60
ORDER BY dept
, name;
ANSWER
DEPT | NAME | SALARY | SUM1 | SUM2 | SUM3 | SUM4 | RW1 | RW2 |
---|---|---|---|---|---|---|---|---|
15 |
Hanes |
80659.80 |
80659.80 |
412701.30 |
80659.80 |
412701.30 |
1 |
1 |
20 |
Pernal |
78171.25 |
257188.55 |
332041.50 |
158831.05 |
332041.50 |
3 |
2 |
20 |
Sanders |
98357.50 |
257188.55 |
332041.50 |
257188.55 |
253870.25 |
3 |
3 |
38 |
Marenghi |
77506.75 |
412701.30 |
155512.75 |
334695.30 |
155512.75 |
5 |
4 |
38 |
O’Brien |
78006.00 |
412701.30 |
155512.75 |
412701.30 |
78006.00 |
5 |
5 |
6.1.29. ROWS Usage
The ROWS phrase (see Window Definition for syntax) is used to limit the aggregation function to a subset of the matching rows. The set of rows to process are defined thus:
-
No ORDER BY: UNBOUNDED PRECEDING to UNBOUNDED FOLLOWING.
-
ORDER BY only: UNBOUNDED PRECEDING to CURRENT ROW.
-
No BETWEEN: CURRENT ROW to "n" preceding/following row.
-
BETWEEN stmt: From "n" to "n" preceding/following row. The end-point must be greater than or equal to the starting point.
The following query illustrates these concepts:
SELECT id
, years
, AVG(years) OVER() AS "p_f"
, AVG(years) OVER(ORDER BY id ROWS
BETWEEN UNBOUNDED PRECEDING AND UNBOUNDED FOLLOWING) AS "p_f"
, AVG(years) OVER(ORDER BY id) AS "p_c"
, AVG(years) OVER(ORDER BY id
ROWS BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW) AS "p_c"
, AVG(years) OVER(ORDER BY id
ROWS UNBOUNDED PRECEDING) AS "p_c"
, AVG(years) OVER(ORDER BY id
ROWS UNBOUNDED FOLLOWING) AS "c_f"
, AVG(years) OVER(ORDER BY id
ROWS 2 FOLLOWING) AS "c_2"
, AVG(years) OVER(ORDER BY id
ROWS 1 PRECEDING) AS "1_c"
, AVG(years) OVER(ORDER BY id
ROWS BETWEEN 1 FOLLOWING AND 2 FOLLOWING) AS "1_2"
FROM staff
WHERE dept IN (15,20)
AND id > 20
AND years > 1
ORDER BY id;
ID | YEARS | p_f | p_f | p_c | p_c | p_c | c_f | c_2 | 1_c | 1_2 |
---|---|---|---|---|---|---|---|---|---|---|
50 |
10 |
6 |
6 |
10 |
10 |
10 |
6 |
7 |
10 |
6 |
70 |
7 |
6 |
6 |
8 |
8 |
8 |
6 |
5 |
8 |
4 |
110 |
5 |
6 |
6 |
7 |
7 |
7 |
5 |
5 |
6 |
6 |
170 |
4 |
6 |
6 |
6 |
6 |
6 |
6 |
6 |
4 |
8 |
190 |
8 |
6 |
6 |
6 |
6 |
6 |
8 |
8 |
6 |
- |
6.1.30. RANGE Usage
The RANGE phrase limits the aggregation result to a range of numeric values - defined relative to the value of the current row being processed (see Window Definition). The range is obtained by taking the value in the current row (defined by the ORDER BY expression) and adding to and/or subtracting from it, then seeing what other matching rows are in the range.
When using a RANGE, only one expression can be specified in the ORDER BY, and that expression must be numeric. |
In the following example, the RANGE function adds to and/or subtracts from the DEPT field. For example, in the function that is used to populate the RG10 field, the current DEPT value is checked against the preceding DEPT values. If their value is within 10 digits of the current value, the related YEARS field is added to the SUM:
SELECT dept
, name
, years
, SMALLINT(SUM(years) OVER(ORDER BY dept
ROWS BETWEEN 1 PRECEDING AND CURRENT ROW)) AS row1
, SMALLINT(SUM(years) OVER(ORDER BY dept
ROWS BETWEEN 2 PRECEDING AND CURRENT ROW)) AS row2
, SMALLINT(SUM(years) OVER(ORDER BY dept
RANGE BETWEEN 1 PRECEDING AND CURRENT ROW)) AS rg01
, SMALLINT(SUM(years) OVER(ORDER BY dept
RANGE BETWEEN 10 PRECEDING AND CURRENT ROW)) AS rg10
, SMALLINT(SUM(years) OVER(ORDER BY dept
RANGE BETWEEN 20 PRECEDING AND CURRENT ROW)) AS rg20
, SMALLINT(SUM(years) OVER(ORDER BY dept
RANGE BETWEEN 10 PRECEDING AND 20 FOLLOWING)) AS rg11
, SMALLINT(SUM(years) OVER(ORDER BY dept
RANGE BETWEEN CURRENT ROW AND 20 FOLLOWING)) AS rg99
FROM staff
WHERE id < 100
AND years IS NOT NULL
ORDER BY dept
, name;
ANSWER
DEPT | NAME | YEARS | ROW1 | ROW2 | RG01 | RG10 | RG20 | RG11 | RG99 |
---|---|---|---|---|---|---|---|---|---|
15 |
Hanes |
10 |
10 |
10 |
17 |
17 |
17 |
32 |
32 |
15 |
Rothman |
7 |
17 |
17 |
17 |
17 |
17 |
32 |
32 |
20 |
Pernal |
8 |
15 |
25 |
15 |
32 |
32 |
43 |
26 |
20 |
Sanders |
7 |
15 |
22 |
15 |
32 |
32 |
43 |
26 |
38 |
Marengh |
5 |
12 |
20 |
11 |
11 |
26 |
17 |
17 |
38 |
O’Brien |
6 |
11 |
18 |
11 |
11 |
26 |
17 |
17 |
42 |
Koonitz |
6 |
12 |
17 |
6 |
17 |
17 |
17 |
6 |
Note the difference between the ROWS as RANGE expressions:
-
The ROWS expression refers to the "n" rows before and/or after (within the partition), as defined by the ORDER BY.
-
The RANGE expression refers to those before and/or after rows (within the partition) that are within an arithmetic range of the current row.
6.1.31. BETWEEN vs. ORDER BY
The BETWEEN predicate in an ordinary SQL statement is used to get those rows that have a value between the specified low-value (given first) and the high value (given last). Thus the predicate "BETWEEN 5 AND 10" may find rows, but the predicate "BETWEEN 10 AND 5" will never find any. The BETWEEN phrase in an aggregation function has a similar usage in that it defines the set of rows to be aggregated. But it differs in that the answer depends upon the related ORDER BY sequence, and a non-match returns a null value, not no-rows. Below is some sample SQL. Observe that the first two aggregations are ascending, while the last two are descending:
SELECT id
, name
, SMALLINT(SUM(id) OVER(ORDER BY id ASC
ROWS BETWEEN 1 PRECEDING AND CURRENT ROW)) AS apc
, SMALLINT(SUM(id) OVER(ORDER BY id ASC
ROWS BETWEEN CURRENT ROW AND 1 FOLLOWING)) AS acf
, SMALLINT(SUM(id) OVER(ORDER BY id DESC
ROWS BETWEEN 1 PRECEDING AND CURRENT ROW)) AS dpc
, SMALLINT(SUM(id) OVER(ORDER BY id DESC
ROWS BETWEEN CURRENT ROW AND 1 FOLLOWING)) AS dcf
FROM staff
WHERE id < 50
AND years IS NOT NULL
ORDER BY id;
ANSWER
ID | NAME | APC | ACF | DPC | DCF |
---|---|---|---|---|---|
10 |
Sanders |
10 |
30 |
30 |
10 |
20 |
Pernal |
30 |
50 |
50 |
30 |
30 |
Marenghi |
50 |
70 |
70 |
50 |
40 |
O’Brien |
70 |
40 |
40 |
70 |
The following table illustrates the processing sequence in the above query. Each BETWEEN is applied from left to right, while the rows are read either from left to right (ORDER BY ID ASC) or right to left (ORDER BY ID DESC):
ASC id (10,20,30,40)
READ ROWS, LEFT to RIGHT | 1ST-ROW | 2ND-ROW | 3RD-ROW | 4TH-ROW |
---|---|---|---|---|
1 PRECEDING to CURRENT ROW |
10=10 |
10+20=30 |
20+30=40 |
30+40=70 |
CURRENT ROW to 1 FOLLOWING |
10+20=30 |
20+30=50 |
30+40=70 |
40 =40 |
DESC id (40,30,20,10)
READ ROWS, RIGHT to LEFT | 1ST-ROW | 2ND-ROW | 3RD-ROW | 4TH-ROW |
---|---|---|---|---|
1 PRECEDING to CURRENT ROW |
20+10=30 |
30+20=50 |
40+30=70 |
40 =40 |
CURRENT ROW to 1 FOLLOWING |
10 =10 |
20+10=30 |
30+20=50 |
40+30=70 |
Preceding row is always on LEFT of current row. Following row_ is always on RIGHT of current row. |
The BETWEEN predicate, when used in an ordinary SQL statement, is not affected by the sequence of the input rows. But the BETWEEN phrase, when used in an aggregation function, is affected by the input sequence. |
7. Scalar Functions
=== Introduction
Scalar functions act on a single row at a time. In this section we shall list all of the ones that come with Db2 and look in detail at some of the more interesting ones. Refer to the SQL Reference for information on those functions not fully described here.
7.1. Sample Data
The following self-defined view will be used throughout this section to illustrate how some of the following functions work. Observe that the view has a VALUES expression that defines the contents- three rows and nine columns.
CREATE VIEW scalar
(d1, f1, s1, c1, v1, ts1, dt1, tm1, tc1) AS
WITH temp1 (n1, c1, t1) AS
( VALUES ( -2.4
, 'ABCDEF'
, '1996-04-22-23.58.58.123456')
, ( +0.0
, 'ABCD '
, '1996-08-15-15.15.15.151515')
, ( +1.8
, 'AB'
, '0001-01-01-00.00.00.000000')
)
SELECT DECIMAL(n1,3,1)
, DOUBLE(n1)
, SMALLINT(n1)
, CHAR(c1,6)
, VARCHAR(RTRIM(c1),6)
, TIMESTAMP(t1)
, DATE(t1)
, TIME(t1)
, CHAR(t1)
FROM temp1;
Below are the view contents:
D1 | F1 | S1 | C1 | V1 | TS1 | DT1 | TM1 | TC1 |
---|---|---|---|---|---|---|---|---|
-2.4 |
-2.4e+000 |
-2 |
ABCDEF |
ABCDEF |
1996-04-22-23.58.58.123456 |
1996-04-22 |
23:58:58 |
1996-04-22-23.58.58.123456 |
0.0 |
0.0e+000 |
0 |
ABCD |
ABCD |
1996-08-15-15.15.15.151515 |
1996-08-15 |
15:15:15 |
1996-08-15-15.15.15.151515 |
1.8 |
1.8e+000 |
1 |
AB |
AB |
0001-01-01-00.00.00.000000 |
0001-01-01 |
00:00:00 |
0001-01-01-00.00.00.000000 |
7.2. Scalar Functions, Definitions
7.2.1. ABS or ABSVAL
Returns the absolute value of a number (e.g. -0.4 returns + 0.4). The output field type will equal the input field type (i.e. double input returns double output).
SELECT d1 AS d1
, ABS(D1) AS d2
, f1 AS f1
, ABS(f1) AS F2
FROM scalar;
ANSWER (float output shortened)
D1 | D2 | F1 | F2 |
---|---|---|---|
-2.4 |
2.4 |
-2.400e+0 |
2.400e+00 |
0.0 |
0.0 |
0.000e+0 |
0.000e+00 |
1.8 |
1.8 |
1.800e+0 |
1.800e+00 |
7.2.2. ACOS
Returns the arccosine of the argument as an angle expressed in radians. The output format is double.
7.2.3. ADD_DAYS
Returns a datetime value that represents the first argument plus a specified number of days.
7.2.4. ADD_HOURS
Returns a timestamp value that represents the first argument plus a specified number of hours.
7.2.5. ADD_MINUTES
Returns a timestamp value that represents the first argument plus a specified number of minutes.
7.2.6. ADD_MONTHS
Returns a datetime value that represents expression plus a specified number of months.
7.2.7. ADD_SECONDS
Returns a timestamp value that represents the first argument plus a specified number of seconds and fractional seconds.
7.2.8. ADD_YEARS
Returns a datetime value that represents the first argument plus a specified number of years.
7.2.9. AGE
Returns a numeric value that represents the number of full years, full months, and full days between the current timestamp and the argument.
7.2.10. ARRAY_DELETE
Deletes elements from an array.
7.2.11. ARRAY_FIRST
Returns the minimum array index value of the array.
7.2.12. ARRAY_LAST
Returns the maximum array index value of the array.
7.2.13. ARRAY_NEXT
Returns the next larger array index value for an array relative to the specified array index argument.
7.2.14. ARRAY_PRIOR
Returns the next smaller array index value for an array relative to the specified array index argument.
7.2.15. ASCII
Returns the ASCII code value of the leftmost input character. Valid input types are any valid character type up to 1 MEG. The output type is integer.
SELECT c1
, ASCII(c1) AS ac1
, ASCII(SUBSTR(c1,2)) AS ac2
FROM scalar
WHERE c1 = 'ABCDEF';
ANSWER
C1 | AC1 | AC2 |
---|---|---|
ABCDEF |
65 |
66 |
The CHR function is the inverse of the ASCII function.
7.2.16. ASIN
Returns the arcsine of the argument as an angle expressed in radians. The output format is double.
7.2.17. ATAN
Returns the arctangent of the argument as an angle expressed in radians. The output format is double.
7.2.18. ATAN2
Returns the arctangent of x and y coordinates, specified by the first and second arguments, as an angle, expressed in radians. The output format is double.
7.2.19. ATANH
Returns the hyperbolic acrctangent of the argument, where the argument is and an angle expressed in radians. The output format is double.
7.2.20. BIGINT
Converts the input value to bigint (big integer) format. The input can be either numeric or character. If character, it must be a valid representation of a number.
WITH temp (big) AS
(VALUES BIGINT(1)
UNION ALL
SELECT big * 256
FROM temp
WHERE big < 1E16)
SELECT big
FROM temp;
ANSWER
BIG |
---|
1 |
256 |
65536 |
16777216 |
4294967296 |
1099511627776 |
281474976710656 |
72057594037927936 |
Converting certain float values to both BIGINT and decimal will result in different values being returned (see below). Both results are arguably correct, it is simply that the two functions use different rounding methods:
WITH temp (f1) AS
(VALUES FLOAT(1.23456789)
UNION ALL
SELECT f1 * 100
FROM temp
WHERE f1 < 1E18
)
SELECT f1 AS float1
, DEC(f1,19) AS decimal1
, BIGINT(f1) AS bigint1
FROM temp;
FLOAT1 | DECIMAL1 | BIGINT1 |
---|---|---|
+1.23456789000000E+000 |
1. |
1 |
+1.23456789000000E+002 |
123. |
123 |
+1.23456789000000E+004 |
12345. |
12345 |
+1.23456789000000E+006 |
1234567. |
1234567 |
+1.23456789000000E+008 |
123456789. |
123456788 |
+1.23456789000000E+010 |
12345678900. |
12345678899 |
+1.23456789000000E+012 |
1234567890000. |
1234567889999 |
+1.23456789000000E+014 |
123456789000000. |
123456788999999 |
+1.23456789000000E+016 |
12345678900000000. |
12345678899999996 |
+1.23456789000000E+018 |
1234567890000000000. |
1234567889999999488 |
See Floating Point Numbers for a discussion on floating-point number manipulation.
7.2.21. BINARY
Returns a fixed-length binary string representation of a string of any data type.
7.2.22. BIT Functions
There are five BIT functions:
-
BITAND 1 if both arguments are 1.
-
BITANDNOT Zero if bit in second argument is 1, otherwise bit in first argument.
-
BITOR 1 if either argument is 1.
-
BITXOR 1 if both arguments differ.
-
BITNOT Returns opposite of the single argument.
The arguments can be SMALLINT (16 bits), INTEGER (32 bits), BIGINT (64 bits), or DECFLOAT (113 bits). The result is the same as the argument with the largest data type.
Negative numbers can be used in bit manipulation. For example the SMALLINT value -1 will have all 16 bits set to "1" (see example on Updating BIT Values). As their name implies, the BIT functions can be used to do bit-by-bit comparisons between two numbers:
WITH temp1 (b1, b2) AS
(VALUES ( 1, 0) ,( 0, 1)
, ( 0, 0) ,( 1, 1)
, ( 2, 1) ,(15,-7)
, (15, 7) ,(-1, 1)
, (15,63) ,(63,31)
, (99,64) ,( 0,-2)
)
, temp2 (b1, b2) AS
(SELECT SMALLINT(b1)
, SMALLINT(b2)
FROM temp1)
SELECT b1
, b2
, HEX(b1) AS "hex1"
, HEX(b2) AS "hex2"
, BITAND(b1,b2) AS "and"
, BITANDNOT(b1,b2) AS "ano"
, BITOR(b1,b2) AS "or"
, BITXOR(b1,b2) AS "xor"
FROM temp2;
ANSWER
B1 | B2 | hex1 | hex2 | and | ano | or | xor |
---|---|---|---|---|---|---|---|
1 |
0 |
0100 |
0000 |
0 |
1 |
1 |
1 |
0 |
1 |
0000 |
0100 |
0 |
0 |
1 |
1 |
0 |
0 |
0000 |
0000 |
0 |
0 |
0 |
0 |
1 |
1 |
0100 |
0100 |
1 |
0 |
1 |
0 |
2 |
1 |
0200 |
0100 |
0 |
2 |
3 |
3 |
15 |
-7 |
0F00 |
F9FF |
9 |
6 |
-1 |
-10 |
15 |
7 |
0F00 |
0700 |
7 |
8 |
15 |
8 |
-1 |
1 |
FFFF |
0100 |
1 |
-2 |
-1 |
-2 |
15 |
63 |
0F00 |
3F00 |
15 |
0 |
63 |
48 |
63 |
31 |
3F00 |
1F00 |
31 |
32 |
63 |
32 |
99 |
64 |
6300 |
4000 |
64 |
35 |
99 |
35 |
0 |
-2 |
0000 |
FEFF |
0 |
0 |
-2 |
-2 |
Displaying BIT Values
It can sometimes be hard to comprehend what a given BASE 10 value is in BIT format. To help, the following user-defined-function converts SMALLINT numbers to BIT values:
CREATE FUNCTION bitdisplay(inparm SMALLINT)
RETURNS CHAR(16)
BEGIN ATOMIC
DECLARE outstr VARCHAR(16);
DECLARE inval INT;
IF inparm >= 0 THEN
SET inval = inparm;
ELSE
SET inval = INT(65536) + inparm;
END IF;
SET outstr = '';
WHILE inval > 0 DO
SET outstr = STRIP(CHAR(MOD(inval,2))) || outstr;
SET inval = inval / 2;
END WHILE;
RETURN RIGHT(REPEAT('0',16) || outstr,16);
END!
Below is an example of the above function in use:
WITH temp1 (b1) AS
(VALUES (32767)
, (16383)
, ( 4096)
, ( 118)
, ( 63)
, ( 16)
, ( 2)
, ( 1)
, ( 0)
, ( -1)
, ( -2)
, ( -3)
, ( -64)
, (-32768)
)
, temp2 (b1) AS
(SELECT SMALLINT(b1)
FROM temp1
)
SELECT b1
, HEX(b1) AS "hex1"
, BITDISPLAY(b1) AS "bit_display"
FROM temp2;
ANSWER
B1 | hex1 | bit_display |
---|---|---|
32767 |
FF7F |
0111111111111111 |
16383 |
FF3F |
0011111111111111 |
4096 |
0010 |
0001000000000000 |
118 |
7600 |
0000000001110110 |
63 |
3F00 |
0000000000111111 |
16 |
1000 |
0000000000010000 |
2 |
0200 |
0000000000000010 |
1 |
0100 |
0000000000000001 |
0 |
0000 |
0000000000000000 |
-1 |
FFFF |
1111111111111111 |
-2 |
FEFF |
1111111111111110 |
-3 |
FDFF |
1111111111111101 |
-64 |
C0FF |
1111111111000000 |
-32768 |
0080 |
1000000000000000 |
Updating BIT Values
Use the BITXOR function to toggle targeted bits in a value. Use the BITANDNOT function to clear the same targeted bits. To illustrate, the next query uses these two functions to toggle and clear the last four bits, because the second parameter is 15, which is b"1111":
WITH temp1 (b1) AS
( VALUES (32767)
, (21845)
, (4096)
, (0)
, (-1)
, (-64)
)
, temp2 (b1, s15) AS
(SELECT SMALLINT(b1)
, SMALLINT(15)
FROM temp1
)
SELECT b1
, BITDISPLAY(b1) AS "b1_display"
, BITXOR(b1,s15) AS "xor"
, BITDISPLAY(BITXOR(b1,s15)) AS "xor_display"
, BITANDNOT(b1,s15) AS "andnot"
, BITDISPLAY(BITANDNOT(b1,s15)) AS "andnot_display"
FROM temp2;
Below is the answer:
B1 | b1_display | xor | xor_display | andnot | andnot_display |
---|---|---|---|---|---|
32767 |
0111111111111111 |
32752 |
0111111111110000 |
32752 |
0111111111110000 |
21845 |
0101010101010101 |
21850 |
0101010101011010 |
21840 |
0101010101010000 |
4096 |
0001000000000000 |
4111 |
0001000000001111 |
4096 |
0001000000000000 |
0 |
0000000000000000 |
15 |
0000000000001111 |
0 |
0000000000000000 |
-1 |
1111111111111111 |
-16 |
1111111111110000 |
-16 |
1111111111110000 |
-64 |
1111111111000000 |
-49 |
1111111111001111 |
-64 |
1111111111000000 |
The next query illustrate the use of the BITAND function to return those bits that match both parameters, and the BITNOT function to toggle all bits:
WITH temp1 (b1) AS
(VALUES (32767)
, (21845)
, (4096)
, (0)
, (-1)
, (-64)
)
, temp2 (b1, s15) AS
(SELECT SMALLINT(b1)
, SMALLINT(15)
FROM temp1
)
SELECT b1
, BITDISPLAY(b1) AS "b1_display"
, BITAND(b1,s15) AS "and"
, BITDISPLAY(BITAND(b1,s15)) AS "and_display"
, BITNOT(b1) AS "not"
, BITDISPLAY(BITNOT(b1)) AS "not_display"
FROM temp2;
Below is the answer:
B1 | b1_display | and | and_display | not | not_display |
---|---|---|---|---|---|
32767 |
0111111111111111 |
15 |
0000000000001111 |
-32768 |
1000000000000000 |
21845 |
0101010101010101 |
5 |
0000000000000101 |
-21846 |
1010101010101010 |
4096 |
0001000000000000 |
0 |
0000000000000000 |
-4097 |
1110111111111111 |
0 |
0000000000000000 |
0 |
0000000000000000 |
-1 |
1111111111111111 |
-1 |
1111111111111111 |
15 |
0000000000001111 |
0 |
0000000000000000 |
-64 |
1111111111000000 |
0 |
0000000000000000 |
63 |
0000000000111111 |
7.2.23. BLOB
Converts the input (1st argument) to a blob. The output length (2nd argument) is optional.
7.2.24. BOOLEAN
Returns the actual Boolean value that corresponds to a non-Boolean representation of a Boolean value.
7.2.25. BTRIM
Removes the characters that are specified in a trim string from the beginning and end of a source string.
7.2.26. CARDINALITY
Returns a value of type BIGINT that is the number of elements in an array.
7.2.27. CEIL or CEILING
Returns the next smallest integer value that is greater than or equal to the input (e.g. 5.045 returns 6.000). The output field type will equal the input field type.
SELECT d1
, CEIL(d1) AS d2
, f1
, CEIL(f1) AS f2
FROM scalar;
ANSWER (float output shortened)
D1 | D2 | F1 | F2 |
---|---|---|---|
-2.4 |
-2. |
-2.400E+0 |
-2.000E+0 |
0.0 |
0. |
+0.000E+0 |
+0.000E+0 |
1.8 |
2. |
+1.800E+0 |
+2.000E+0 |
Usually, when Db2 converts a number from one format to another, any extra digits on the right are truncated, not rounded. For example, the output of INTEGER(123.9) is 123. Use the CEIL or ROUND functions to avoid truncation. |
7.2.28. CHAR
The CHAR function has a multiplicity of uses. The result is always a fixed-length character value, but what happens to the input along the way depends upon the input type:
-
For character input, the CHAR function acts a bit like the SUBSTR function, except that it can only truncate starting from the left-most character. The optional length parameter, if provided, must be a constant or keyword.
-
Date-time input is converted into an equivalent character string. Optionally, the external format can be explicitly specified (i.e. ISO, USA, EUR, JIS, or LOCAL).
-
Integer and double input is converted into a left-justified character string.
-
Decimal input is converted into a right-justified character string with leading zeros. The format of the decimal point can optionally be provided. The default decimal point is a dot. The '+' and '-' symbols are not allowed as they are used as sign indicators.
Below are some examples of the CHAR function in action:
SELECT name
, CHAR(name, 3)
, comm
, CHAR(comm)
, CHAR(comm,'@')
FROM staff
WHERE id BETWEEN 80 AND 100
ORDER BY id;
ANSWER
NAME | 2 | COMM | 4 | 5 |
---|---|---|---|---|
James |
Jam |
128.20 |
00128.20 |
00128@20 |
Koonitz |
Koo |
1386.70 |
01386.70 |
01386@70 |
Plotz |
Plo |
- |
- |
- |
The CHAR function treats decimal numbers quite differently from integer and real numbers. In particular, it right-justifies the former (with leading zeros), while it left-justifies the latter (with trailing blanks). The next example illustrates this point:
WITH temp1 (n) AS
(VALUES (3)
UNION ALL
SELECT n * n
FROM temp1
WHERE n < 9000
)
SELECT n AS int
, CHAR(INT(n)) AS char_int
, CHAR(FLOAT(n)) AS char_flt
, CHAR(DEC(n)) AS char_dec
FROM temp1;
ANSWER
INT | CHAR_INT | CHAR_FLT | CHAR_DEC |
---|---|---|---|
3 |
3 |
3.0E0 |
00000000003. |
9 |
9 |
9.0E0 |
00000000009. |
81 |
81 |
8.1E1 |
00000000081. |
6561 |
6561 |
6.561E3 |
00000006561. |
43046721 |
43046721 |
4.3046721E7 |
00043046721. |
Negative numeric input is given a leading minus sign. This messes up the alignment of digits in the column (relative to any positive values). In the following query, a leading blank is put in front of all positive numbers in order to realign everything:
WITH temp1 (n1, n2) AS
(VALUES (SMALLINT(+3)
, SMALLINT(-7))
UNION ALL
SELECT n1 * n2
, n2
FROM temp1
WHERE n1 < 300
)
SELECT n1
, CHAR(n1) AS i1
, CASE
WHEN n1 < 0 THEN CHAR(n1)
ELSE '+' CONCAT CHAR(n1)
END AS i2
, CHAR(DEC(n1)) AS d1
, CASE
WHEN n1 < 0 THEN CHAR(DEC(n1))
ELSE '+' CONCAT CHAR(DEC(n1))
END AS d2
FROM temp1;
ANSWER
N1 | I1 | I2 | D1 | D2 |
---|---|---|---|---|
3 |
3 |
+3 |
00003. |
+00003. |
-21 |
-21 |
-21 |
-00021. |
-00021. |
147 |
147 |
+147 |
00147. |
+00147. |
-1029 |
-1029 |
-1029 |
-01029. |
-01029. |
7203 |
7203 |
+7203 |
07203. |
+07203. |
Both the I2 and D2 fields above will have a trailing blank on all negative values - that was added during the concatenation operation. The RTRIM function can be used to remove it.
DATE-TIME Conversion
The CHAR function can be used to convert a date-time value to character. If the input is not a timestamp, the output layout can be controlled using the format option:
-
ISO: International Standards Organization.
-
USA: American.
-
EUR: European, which is usually the same as ISO.
-
JIS: Japanese Industrial Standard, which is usually the same as ISO.
-
LOCAL: Whatever your computer is set to.
Below are some DATE examples:
SELECT CHAR(CURRENT DATE,ISO) AS iso
, CHAR(CURRENT DATE,EUR) AS eur
, CHAR(CURRENT DATE,JIS) AS jis
, CHAR(CURRENT DATE,USA) AS usa
FROM sysibm.sysdummy1;
ANSWER
ISO | EUR | JIS | USA |
---|---|---|---|
2005-11-30 |
30.11.2005 |
2005-11-30 |
11/30/2005 |
Below are some TIME examples: .CHAR function examples - time value
SELECT CHAR(CURRENT TIME,ISO) AS iso
, CHAR(CURRENT TIME,EUR) AS eur
, CHAR(CURRENT TIME,JIS) AS jis
, CHAR(CURRENT TIME,USA) AS usa
FROM sysibm.sysdummy1;
ANSWER
ISO | EUR | JIS | USA |
---|---|---|---|
19.42.21 |
19.42.21 |
19:42:21 |
07:42 PM |
A timestamp cannot be formatted to anything other than ISO output:
SELECT CHAR(CURRENT TIMESTAMP) AS TS
FROM sysibm.sysdummy1;
ANSWER
TS |
---|
2005-11-30-19.42.21.873002 |
Converting a date or time value to character, and then ordering the set of matching rows can result in unexpected orders. See CASE Checks in Wrong Sequence for details. |
CHAR vs. DIGITS - A Comparison
Numeric input can be converted to character using either the DIGITS or the CHAR function, though the former does not support float. Both functions work differently, and neither gives perfect output. The CHAR function doesn’t properly align up positive and negative numbers, while the DIGITS function looses both the decimal point and sign indicator:
SELECT d2
, CHAR(d2) AS cd2
, DIGITS(d2) AS dd2
FROM
(SELECT DEC(d1, 4, 1) AS d2
FROM scalar
) AS xxx
ORDER BY 1;
ANSWER
D2 | CD2 | DD2 |
---|---|---|
-2.4 |
-002.4 |
0024 |
0.0 |
000.0 |
0000 |
1.8 |
001.8 |
0018 |
Neither the DIGITS nor the CHAR function do a great job of converting numbers to characters. See Convert Number to Character for some user-defined functions that can be used instead. |
7.2.29. CHARACTER_LENGTH
This function is similar to the LENGTH function, except that it works with different encoding schemas. The result is an integer value that is the length of the input string.
WITH temp1 (c1) AS
(VALUES (CAST('ÁÉÌ' AS VARCHAR(10)))
)
SELECT c1 AS C1
, LENGTH(c1) AS LEN
, OCTET_LENGTH(c1) AS OCT
, CHAR_LENGTH(c1,OCTETS) AS L08
, CHAR_LENGTH(c1,CODEUNITS16) AS L16
, CHAR_LENGTH(c1,CODEUNITS32) AS L32
FROM temp1;
ANSWER
C1 | LEN | OCT | L08 | L16 | L32 |
---|---|---|---|---|---|
ÁÉÌ |
6 |
6 |
6 |
3 |
3 |
7.2.30. CHR
Converts integer input in the range 0 through 255 to the equivalent ASCII character value. An input value above 255 returns 255. The ASCII function (see above) is the inverse of the CHR function.
SELECT 'A' AS "c"
, ASCII('A') AS "c>n"
, CHR(ASCII('A')) AS "c>n>c"
, CHR(333) AS "nl"
FROM staff
WHERE id = 10;
ANSWER
C | C>N | C>N>C | NL |
---|---|---|---|
A |
65 |
A |
ÿ |
At present, the CHR function has a bug that results in it not returning a null value when the input value is greater than 255. |
7.2.31. CLOB
Converts the input (1st argument) to a CLOB. The output length (2nd argument) is optional. If the input is truncated during conversion, a warning message is issued. For example, in the following example the second CLOB statement will induce a warning for the first two lines of input because they have non-blank data after the third byte:
SELECT c1
, CLOB(c1) AS cc1
, CLOB(c1,3) AS cc2
FROM scalar;
ANSWER
C1 | CC1 | CC2 |
---|---|---|
ABCDEF |
ABCDEF |
ABC |
ABCD |
ABCD |
ABC |
AB |
AB |
AB |
The Db2BATCH command processor dies a nasty death whenever it encounters a CLOB field in the output. If possible, convert to VARCHAR first to avoid this problem. |
7.2.32. COALESCE
Returns the first non-null value in a list of input expressions (reading from left to right). Each expression is separated from the prior by a comma. All input expressions must be compatible. VALUE is a synonym for COALESCE.
SELECT id
, comm
, COALESCE(comm, 0)
FROM staff
WHERE id < 30
ORDER BY id;
ANSWER
ID | COMM | 3 |
---|---|---|
10 |
- |
0.00 |
20 |
612.45 |
612.45 |
A CASE expression can be written to do exactly the same thing as the COALESCE function. The following SQL statement shows two logically equivalent ways to replace nulls:
WITH temp1(c1,c2,c3) AS
(VALUES (CAST(NULL AS SMALLINT)
, CAST(NULL AS SMALLINT)
, CAST(10 AS SMALLINT))
)
SELECT COALESCE(c1, c2, c3) AS cc1
, CASE
WHEN c1 IS NOT NULL THEN c1
WHEN c2 IS NOT NULL THEN c2
WHEN c3 IS NOT NULL THEN c3
END AS cc2
FROM temp1;
ANSWER
CC1 | CC2 |
---|---|
10 |
10 |
Be aware that a field can return a null value, even when it is defined as not null. This occurs if a column function is applied against the field, and no row is returned:
SELECT COUNT(*) AS #rows
, MIN(id) AS min_id
, COALESCE(MIN(id), -1) AS ccc_id
FROM staff
WHERE id < 5;
ANSWER
#ROWS | MIN_ID | CCC_ID |
---|---|---|
0 |
- |
-1 |
7.2.33. COLLATION_KEY
Returns a VARBINARY string that represents the collation key of the expression argument, in the specified collation.
7.2.34. COLLATION_KEY_BIT
Returns a VARCHAR FOR BIT DATA string that is the collation sequence of the first argument in the function. There three parameters:
-
String to be evaluated.
-
Collation sequence to use (must be valid).
-
Length of output (optional).
The following query displays three collation sequences:
-
All flavors of a given character as the same (i.e. "a" = "A" = "Ä").
-
Upper and lower case characters are equal, but sort lower than accented characters.
-
All variations of a character have a different collation value.
Now for the query:
WITH temp1 (c1) As
(VALUES ('a'), ('A'), ('Á'), ('Ä'), ('b'))
SELECT c1
, COLLATION_KEY_BIT(c1, 'UCA400R1_S1', 9) AS "a=A=Á=Ä"
, COLLATION_KEY_BIT(c1, 'UCA400R1_S2' ,9) AS "a=A<Á<Ä"
, COLLATION_KEY_BIT(c1, 'UCA400R1_S3' ,9) AS "a<A<Á<Ä"
FROM temp1
ORDER BY COLLATION_KEY_BIT(c1, 'UCA400R1_S3');
Below is the answer:
C1 | a=A=Á=Ä | a=A<Á<Ä | a<A<Á<Ä |
---|---|---|---|
a |
x'2600' |
x'26010500' |
x'260105010500' |
A |
x'2600' |
x'26010500' |
x'260105018F00' |
Á |
x'2600' |
x'2601868D00' |
x'2601868D018F0500' |
Ä |
x'2600' |
x'2601869D00' |
x'2601869D018F0500' |
b |
x'2800' |
x'28010500' |
x'280105010500' |
7.2.35. COMPARE_DECFLOAT
Compares two DECFLOAT expressions and returns a SMALLINT number:
-
0 if both values exactly equal (i.e. no trailing-zero differences)
-
1 if the first value is less than the second value.
-
2 if the first value is greater than the second value.
-
3 if the result is unordered (i.e. either argument is NaN or sNaN).
Query
-COMPARE_DECFLOAT function example
WITH temp1 (d1, d2) AS
(VALUES (DECFLOAT(+1.0), DECFLOAT(+1.0))
, (DECFLOAT(+1.0), DECFLOAT(+1.00))
, (DECFLOAT(-1.0), DECFLOAT(-1.00))
, (DECFLOAT(+0.0), DECFLOAT(+0.00))
, (DECFLOAT(-0.0), DECFLOAT(-0.00))
, (DECFLOAT(1234), +infinity)
, (+infinity, +infinity)
, (+infinity, -infinity)
, (DECFLOAT(1234), -NaN)
)
SELECT COMPARE_DECFLOAT(d1, d2) AS Result
FROM temp1;
ANSWER
Result |
---|
0 |
2 |
1 |
2 |
2 |
1 |
0 |
2 |
3 |
Several values that compare as "less than" or "greater than" above are equal in the usual sense. See the section on DECFLOAT arithmetic for details (see DECFLOAT Arithmetic). |
7.2.36. CONCAT
Joins two strings together. The CONCAT function has both "infix" and "prefix" notations. In the former case, the verb is placed between the two strings to be acted upon. In the latter case, the two strings come after the verb. Both syntax flavours are illustrated below:
SELECT 'A' || 'B'
, 'A' CONCAT 'B'
, CONCAT('A','B')
, 'A' || 'B' || 'C'
, CONCAT(CONCAT('A','B'),'C')
FROM staff
WHERE id = 10;
ANSWER
1 | 2 | 3 | 4 | 5 |
---|---|---|---|---|
AB |
AB |
AB |
ABC |
ABC |
Note that the "||" keyword can not be used with the prefix notation. This means that "||('a','b')" is not valid while "CONCAT('a','b')" is.
Using CONCAT with ORDER BY
When ordinary character fields are concatenated, any blanks at the end of the first field are left in place. By contrast, concatenating varchar fields removes any (implied) trailing blanks. If the result of the second type of concatenation is then used in an ORDER BY, the resulting row sequence will probably be not what the user intended. To illustrate:
WITH temp1 (col1, col2) AS
(VALUES
('A' , 'YYY')
, ('AE', 'OOO')
, ('AE', 'YYY')
)
SELECT col1
, col2
, col1 CONCAT col2 AS col3
FROM temp1
ORDER BY col3;
ANSWER
COL1 | COL2 | COL3 |
---|---|---|
AE |
OOO |
AEOOO |
AE |
YYY |
AEYYY |
A |
YYY |
AYYY |
Converting the fields being concatenated to character gets around this problem:
WITH temp1 (col1, col2) AS
(VALUES
('A' , 'YYY')
, ('AE', 'OOO')
, ('AE', 'YYY')
)
SELECT col1
, col2
, CHAR(col1, 2) CONCAT CHAR(col2, 3) AS col3
FROM temp1
ORDER BY col3;
ANSWER
COL1 | COL2 | COL3 |
---|---|---|
A |
YYY |
A YYY |
AE |
OOO |
AEOOO |
AE |
YYY |
AEYYY |
Never do an ORDER BY on a concatenated set of variable length fields. The resulting row sequence is probably not what the user intended (see above). |
7.2.37. COS
Returns the cosine of the argument where the argument is an angle expressed in radians. The output format is double.
WITH temp1(n1) AS
(VALUES (0)
UNION ALL
SELECT n1 + 10
FROM temp1
WHERE n1 < 90
)
SELECT n1
, DEC(RADIANS(n1),4,3) AS ran
, DEC(COS(RADIANS(n1)),4,3) AS cos
, DEC(SIN(RADIANS(n1)),4,3) AS sin
FROM temp1;
ANSWER
N1 | RAN | COS | SIN |
---|---|---|---|
0 |
0.000 |
1.000 |
0.000 |
10 |
0.174 |
0.984 |
0.173 |
20 |
0.349 |
0.939 |
0.342 |
30 |
0.523 |
0.866 |
0.500 |
40 |
0.698 |
0.766 |
0.642 |
50 |
0.872 |
0.642 |
0.766 |
60 |
1.047 |
0.500 |
0.866 |
70 |
1.221 |
0.342 |
0.939 |
80 |
1.396 |
0.173 |
0.984 |
90 |
1.570 |
0.000 |
1.000 |
7.2.38. COSH
Returns the hyperbolic cosine for the argument, where the argument is an angle expressed in radians. The output format is double.
7.2.39. COT
Returns the cotangent of the argument where the argument is an angle expressed in radians. The output format is double.
7.2.40. CURSOR_ROWCOUNT
Returns the cumulative count of all rows fetched by the specified cursor since the cursor was opened.
7.2.41. DATAPARTITIONNUM
Returns the sequence number of the partition in which the row resides.
7.2.42. DATE
Converts the input into a date value. The nature of the conversion process depends upon the input type and length:
-
Timestamp and date input have the date part extracted.
-
Char or varchar input that is a valid string representation of a date or a timestamp (e.g. "1997-12-23") is converted as is.
-
Char or varchar input that is seven bytes long is assumed to be a Julian date value in the format yyyynnn where yyyy is the year and nnn is the number of days since the start of the year (in the range 001 to 366).
-
Numeric input is assumed to have a value which represents the number of days since the date "0001-01-01" inclusive. All numeric types are supported, but the fractional part of a value is ignored (e.g. 12.55 becomes 12 which converts to "0001-01-12").
If the input can be null, the output will also support null. Null values convert to null output.
SELECT ts1
, DATE(ts1) AS dt1
FROM scalar;
ANSWER
TS1 | DT1 |
---|---|
1996-04-22-23.58.58.123456 |
1996-04-22 |
1996-08-15-15.15.15.151515 |
1996-08-15 |
0001-01-01-00.00.00.000000 |
0001-01-01 |
WITH temp1(n1) AS
(VALUES
(000001)
, (728000)
, (730120)
)
SELECT n1
, DATE(n1) AS d1
FROM temp1;
ANSWER
N1 | D1 |
---|---|
1 |
0001-01-01 |
728000 |
1994-03-13 |
730120 |
2000-01-01 |
7.2.43. DATE_PART
Returns a portion of a datetime based on its arguments. It extracts the subfield that is specified from the date, time, timestamp, and duration values.
7.2.44. DATE_TRUNC
Truncates a date, time, or timestamp value to the specified time unit.
7.2.45. DAY
Returns the day (as in day of the month) part of a date (or equivalent) value. The output format is integer.
SELECT dt1
, DAY(dt1) AS day1
FROM scalar
WHERE DAY(dt1) > 10;
ANSWER
DT1 | DAY1 |
---|---|
1996-04-22 |
22 |
1996-08-15 |
15 |
If the input is a date or timestamp, the day value must be between 1 and 31. If the input is a date or timestamp duration, the day value can ran from -99 to +99, though only -31 to +31 actually make any sense:
SELECT dt1
, DAY(dt1) AS day1
, dt1 -'1996-04-30' AS dur2
, DAY(dt1 -'1996-04-30') AS day2
FROM scalar
WHERE DAY(dt1) > 10
ORDER BY dt1;
ANSWER
DT1 | DAY1 | DUR2 | DAY2 |
---|---|---|---|
1996-04-22 |
22 |
-8. |
-8 |
1996-08-15 |
15 |
315. |
15 |
A date-duration is what one gets when one subtracts one date from another. The field is of type decimal(8), but the value is not really a number. It has digits in the format: YYYYMMDD, so in the above query the value "315" represents 3 months, 15 days. |
7.2.46. DAYNAME
Returns the name of the day (e.g. Friday) as contained in a date (or equivalent) value. The output format is varchar(100).
SELECT dt1
, DAYNAME(dt1) AS dy1
, LENGTH(DAYNAME(dt1)) AS dy2
FROM scalar
WHERE DAYNAME(dt1) LIKE '%a%y'
ORDER BY dt1;
ANSWER
DT1 | DY1 | DY2 |
---|---|---|
0001-01-01 |
Monday |
6 |
1996-04-22 |
Monday |
6 |
1996-08-15 |
Thursday |
8 |
7.2.47. DAYOFMONTH
Returns an integer between 1 and 31 that represents the day of the month.
7.2.48. DAYOFWEEK
Returns a number that represents the day of the week (where Sunday is 1 and Saturday is 7) from a date (or equivalent) value. The output format is integer.
SELECT dt1
, DAYOFWEEK(dt1) AS dwk
, DAYNAME(dt1) AS dnm
FROM scalar
ORDER BY dwk
, dnm;
ANSWER
DT1 | DWK | DNM |
---|---|---|
1996-04-22 |
2 |
Monday |
0001-01-01 |
2 |
Saturday |
1996-08-15 |
5 |
Thursday |
7.2.49. DAYOFWEEK_ISO
Returns an integer value that represents the day of the "ISO" week. An ISO week differs from an ordinary week in that it begins on a Monday (i.e. day-number = 1) and it neither ends nor begins at the exact end of the year. Instead, the final ISO week of the prior year will continue into the new year. This often means that the first days of the year have an ISO week number of 52, and that one gets more than seven days in a week for ISO week 52.
WITH temp1 (n) AS
(VALUES (0)
UNION ALL
SELECT n+1
FROM temp1
WHERE n < 9
),
temp2 (dt1) AS
(VALUES(DATE('1999-12-25'))
, (DATE('2000-12-24'))
),
temp3 (dt2) AS
(SELECT dt1 + n DAYS
FROM temp1
, temp2)
SELECT CHAR(dt2,ISO) AS date
, SUBSTR(DAYNAME(dt2),1,3) AS day
, WEEK(dt2) AS w
, DAYOFWEEK(dt2) AS d
, WEEK_ISO(dt2) AS wi
, DAYOFWEEK_ISO(dt2) AS di
FROM
temp3
ORDER BY 1;
ANSWER
DATE | DAY | W | D | WI | DI |
---|---|---|---|---|---|
1999-12-25 |
Sat |
52 |
7 |
51 |
6 |
1999-12-26 |
Sun |
53 |
1 |
51 |
7 |
1999-12-27 |
Mon |
53 |
2 |
52 |
1 |
1999-12-28 |
Tue |
53 |
3 |
52 |
2 |
1999-12-29 |
Wed |
53 |
4 |
52 |
3 |
1999-12-30 |
Thu |
53 |
5 |
52 |
4 |
1999-12-31 |
Fri |
53 |
6 |
52 |
5 |
2000-01-01 |
Sat |
1 |
7 |
52 |
6 |
2000-01-02 |
Sun |
2 |
1 |
52 |
7 |
2000-01-03 |
Mon |
2 |
2 |
1 |
1 |
2000-12-24 |
Sun |
53 |
1 |
51 |
7 |
2000-12-25 |
Mon |
53 |
2 |
52 |
1 |
2000-12-26 |
Tue |
53 |
3 |
52 |
2 |
2000-12-27 |
Wed |
53 |
4 |
52 |
3 |
2000-12-28 |
Thu |
53 |
5 |
52 |
4 |
2000-12-29 |
Fri |
53 |
6 |
52 |
5 |
2000-12-30 |
Sat |
53 |
7 |
52 |
6 |
2000-12-31 |
Sun |
54 |
1 |
52 |
7 |
2001-01-01 |
Mon |
1 |
2 |
1 |
1 |
2001-01-02 |
Tue |
1 |
3 |
1 |
2 |
7.2.50. DAYOFYEAR
Returns a number that is the day of the year (from 1 to 366) from a date (or equivalent) value. The output format is integer.
SELECT dt1
, DAYOFYEAR(dt1) AS dyr
FROM scalar
ORDER BY dyr;
ANSWER
DT1 | DYR |
---|---|
0001-01-01 |
1 |
1996-04-22 |
113 |
1996-08-15 |
228 |
7.2.51. DAYS
Converts a date (or equivalent) value into a number that represents the number of days since the date "0001-01-01" inclusive. The output format is INTEGER.
SELECT dt1
, DAYS(dt1) AS dy1
FROM scalar
ORDER BY dy1
, dt1;
ANSWER
DT1 | DY1 |
---|---|
0001-01-01 |
1 |
1996-04-22 |
728771 |
1996-08-15 |
728886 |
The DATE function can act as the inverse of the DAYS function. It can convert the DAYS output back into a valid date.
7.2.52. DAYS_BETWEEN
Returns the number of full days between the specified arguments.
7.2.53. DAYS_TO_END_OF_MONTH
Returns the number of days to the end of the month.
7.2.54. DBCLOB
Converts the input (1st argument) to a dbclob. The output length (2nd argument) is optional.
7.2.55. DBPARTITIONNUM
Returns the partition number of the row. The result is zero if the table is not partitioned. The output is of type integer, and is never null.
SELECT DBPARTITIONNUM(id) AS dbnum
FROM staff
WHERE id = 10;
ANSWER
DBNUM |
---|
0 |
The DBPARTITIONNUM function will generate a SQL error if the column/row used can not be related directly back to specific row in a real table. Therefore, one can not use this function on fields in GROUP BY statements, nor in some views. It can also cause an error when used in an outer join, and the target row failed to match in the join.
7.2.56. DECFLOAT
Converts a character or numeric expression to DECFLOAT.
The first parameter is the input expression. The second is the number of digits of precision (default = 34). And the third is the decimal character value (default = '.').
SELECT DECFLOAT(+123.4)
, DECFLOAT(1.0, 16)
, DECFLOAT(1.0000, 16)
, DECFLOAT(1.2e-3, 34)
, DECFLOAT('1.2e-3', 34)
, DECFLOAT(-1E3, 34)
, DECFLOAT('-1E3', 34)
, DECFLOAT('12.5', 16)
, DECFLOAT('12#5', 16, '#')
FROM sysibm.sysdummy1;
ANSWER
1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 |
---|---|---|---|---|---|---|---|---|
123.4 |
1.0 |
1.0000 |
0.0011999999999999999 |
0.0012 |
-1000 |
-1E+3 |
12.5 |
12.5 |
The function does not always precisely convert floating-point numeric values to their DECFLOAT equivalent (see example above). Use character conversion instead. |
7.2.57. DECFLOAT_FORMAT
Returns a DECFLOAT(34) value that is based on the interpretation of the input string using the specified format.
7.2.58. DEC or DECIMAL
Converts either character or numeric input to decimal. When the input is of type character, the decimal point format can be specified.
WITH temp1(n1, n2, c1, c2) AS
(VALUES
(123, 1E2, '123.4', '567$8')
)
SELECT DEC(n1, 3) AS dec1
, DEC(n2, 4, 1) AS dec2
, DEC(c1, 4, 1) AS dec3
, DEC(c2, 4, 1, '$') AS dec4
FROM temp1;
ANSWER
DEC1 | DEC2 | DEC3 | DEC4 |
---|---|---|---|
123. |
100.0 |
123.4 |
567.8 |
Converting a floating-point number to decimal may get different results from converting the same number to integer. See Floating Point Numbers for a discussion of this issue. |
7.2.59. DECODE
The DECODE function is a simplified form of the CASE expression. The first parameter is the expression to be evaluated. This is followed by pairs of "before" and "after" expressions. At the end is the "else" result:
SELECT firstnme
, sex
, CASE sex
WHEN 'F' THEN 'FEMALE'
WHEN 'M' THEN 'MALE'
ELSE '?'
END AS sex2
, DECODE(sex, 'F', 'FEMALE', 'M', 'MALE', '?') AS sex3
FROM employee
WHERE firstnme < 'D'
ORDER BY firstnme;
ANSWER
FIRSTNME | SEX | SEX2 | SEX3 |
---|---|---|---|
BRUCE |
M |
MALE |
MALE |
CHRISTINE |
F |
FEMALE |
FEMALE |
7.2.60. DECRYPT_BIN and DECRYPT_CHAR
Decrypts data that has been encrypted using the ENCRYPT function. Use the BIN function to decrypt binary data (e.g. BLOBS, CLOBS) and the CHAR function to do character data. Numeric data cannot be encrypted.
If the password is null or not supplied, the value of the encryption password special register will be used. If it is incorrect, a SQL error will be generated.
SELECT id
, name
, DECRYPT_CHAR(name2,'CLUELESS') AS name3
, GETHINT(name2) AS hint
, name2
FROM
(SELECT id
, name
, ENCRYPT(name, 'CLUELESS', 'MY BOSS') AS name2
FROM staff
WHERE id < 30
) AS xxx
ORDER BY id;
7.2.61. DEGREES
Returns the number of degrees converted from the argument as expressed in radians. The output format is double.
7.2.62. DEREF
Returns an instance of the target type of the argument.
7.2.63. DIFFERENCE
Returns the difference between the sounds of two strings as determined using the SOUNDEX function. The output (of type integer) ranges from 4 (good match) to zero (poor match).
SELECT a.name AS n1
, SOUNDEX(a.name) AS s1
, b.name AS n2
, SOUNDEX(b.name) AS s2
, DIFFERENCE (a.name,b.name) AS df
FROM staff a
, staff b
WHERE a.id = 10
AND b.id > 150
AND b.id < 250
ORDER BY df DESC
, n2 ASC;
ANSWER
N1 | S1 | N2 | S2 | DF |
---|---|---|---|---|
Sanders |
S536 |
Sneider |
S536 |
4 |
Sanders |
S536 |
Smith |
S530 |
3 |
Sanders |
S536 |
Lundquist |
L532 |
2 |
Sanders |
S536 |
Daniels |
D542 |
1 |
Sanders |
S536 |
Molinare |
M456 |
1 |
Sanders |
S536 |
Scoutten |
S350 |
1 |
Sanders |
S536 |
Abrahams |
A165 |
0 |
Sanders |
S536 |
Kermisch |
K652 |
0 |
Sanders |
S536 |
Lu |
L000 |
0 |
The difference function returns one of five possible values. In many situations, it would be imprudent to use a value with such low granularity to rank values. |
7.2.64. DIGITS
Converts an integer or decimal value into a character string with leading zeros. Both the sign indicator and the decimal point are lost in the translation.
SELECT s1
, DIGITS(s1) AS ds1
, d1
, DIGITS(d1) AS dd1
FROM scalar;
ANSWER
S1 | DS1 | D1 | DD1 |
---|---|---|---|
2 |
00002 |
-2.4 |
024 |
0 |
00000 |
0.0 |
000 |
1 |
00001 |
1.8 |
018 |
The CHAR function can sometimes be used as alternative to the DIGITS function. Their output differs slightly - see Convert Number to Character for a comparison.
Neither the DIGITS nor the CHAR function do a great job of converting numbers to characters. See Convert Number to Character for some user-defined functions that can be used instead. |
7.2.65. DOUBLE or DOUBLE_PRECISION
Converts numeric or valid character input to type double. This function is actually two with the same name. The one that converts numeric input is a SYSIBM function, while the other that handles character input is a SYSFUN function. The keyword DOUBLE_PRECISION has not been defined for the latter.
WITH temp1(c1,d1) AS
(VALUES ('12345',12.4)
, ('-23.5',1234)
, ('1E+45',-234)
, ('-2e05',+2.4)
)
SELECT DOUBLE(c1) AS c1d
, DOUBLE(d1) AS d1d
FROM temp1;
ANSWER (output shortened)
C1D | D1D |
---|---|
+1.23450000E+004 |
+1.24000000E+001 |
-2.35000000E+001 |
+1.23400000E+003 |
+1.00000000E+045 |
-2.34000000E+002 |
-2.00000000E+005 |
+2.40000000E+000 |
See Floating Point Numbers for a discussion on floating-point number manipulation.
7.2.66. EMPTY_BLOB, EMPTY_CLOB, EMPTY_DBCLOB, and EMPTY_NCLOB
These functions return a zero-length value with a data type of BLOB, CLOB, or DBCLOB.
ENCRYPT
Returns a encrypted rendition of the input string. The input must be char or varchar. The output is varchar for bit data.
The input values are defined as follows:
-
ENCRYPTED DATA: A char or varchar string 32633 bytes that is to be encrypted. Numeric data must be converted to character before encryption.
-
PASSWORD: A char or varchar string of at least six bytes and no more than 127 bytes. If the value is null or not provided, the current value of the encryption password special register will be used. Be aware that a password that is padded with blanks is not the same as one that lacks the blanks.
-
HINT: A char or varchar string of up to 32 bytes that can be referred to if one forgets what the password is. It is included with the encrypted string and can be retrieved using the GETHINT function.
The length of the output string can be calculated thus:
-
When the hint is provided, the length of the input data, plus eight bytes, plus the distance to the next eight-byte boundary, plus thirty-two bytes for the hint.
-
When the hint is not provided, the length of the input data, plus eight bytes, plus the distance to the next eight-byte boundary.
SELECT id
, name
, ENCRYPT(name,'THAT IDIOT','MY BROTHER') AS name2
FROM staff
WHERE ID < 30
ORDER BY id;
7.2.67. EVENT_MON_STATE
Returns an operational state of a particular event monitor.
7.2.68. EXP
Returns the exponential function of the argument. The output format is double.
WITH temp1(n1) AS
(VALUES (0)
UNION ALL
SELECT n1 + 1
FROM temp1
WHERE n1 < 10
)
SELECT n1
, EXP(n1) AS e1
, SMALLINT(EXP(n1)) AS e2
FROM temp1;
ANSWER
N1 | E1 | E2 |
---|---|---|
0 |
+1.00000000000000E+0 |
1 |
1 |
+2.71828182845904E+0 |
2 |
2 |
+7.38905609893065E+0 |
7 |
3 |
+2.00855369231876E+1 |
20 |
4 |
+5.45981500331442E+1 |
54 |
5 |
+1.48413159102576E+2 |
148 |
6 |
+4.03428793492735E+2 |
403 |
7 |
+1.09663315842845E+3 |
1096 |
8 |
+2.98095798704172E+3 |
2980 |
9 |
+8.10308392757538E+3 |
8103 |
10 |
+2.20264657948067E+4 |
22026 |
7.2.69. EXTRACT
Returns a portion of a datetime based on its arguments.
7.2.70. FIRST_DAY
Returns a date or timestamp that represents the first day of the month of the argument.
7.2.71. FLOAT
Same as DOUBLE or DOUBLE_PRECISION.
7.2.72. FLOOR
Returns the next largest integer value that is smaller than or equal to the input (e.g. 5.945 returns 5.000). The output field type will equal the input field type.
SELECT d1
, FLOOR(d1) AS d2
, f1
, FLOOR(f1) AS f2
FROM scalar;
ANSWER (float output shortened)
D1 | D2 | F1 | F2 |
---|---|---|---|
2.4 |
-3. |
-2.400E+0 |
-3.000E+0 |
0.0 |
+0. |
+0.000E+0 |
+0.000E+0 |
1.8 |
+1. |
+1.800E+0 |
+1.000E+0 |
7.2.73. FROM_UTC_TIMESTAMP
Returns a TIMESTAMP that is converted from Coordinated Universal Time to the time zone specified by the time zone string. FROM_UTC_TIMESTAMP is a statement deterministic function.
7.2.74. GENERATE_UNIQUE
Uses the system clock and node number to generate a value that is guaranteed unique (as long as one does not reset the clock). The output is of type CHAR(13) FOR BIT DATA. There are no arguments. The result is essentially a timestamp (set to universal time, not local time), with the node number appended to the back.
SELECT id
, GENERATE_UNIQUE() AS unique_val#1
, DEC(HEX(GENERATE_UNIQUE()),26) AS unique_val#2
FROM staff
WHERE id < 50
ORDER BY id;
ANSWER
ID | UNIQUE_VAL#1 | UNIQUE_VAL#2 |
---|---|---|
10 |
20011017191648990521000000. |
|
20 |
20011017191648990615000000. |
|
30 |
20011017191648990642000000. |
|
40 |
20011017191648990669000000. |
Observe that in the above example, each row gets a higher value. This is to be expected, and is in contrast to a CURRENT TIMESTAMP call, where every row returned by the cursor will have the same timestamp value. Also notice that the second invocation of the function on the same row got a lower value (than the first). In the prior query, the HEX and DEC functions were used to convert the output value into a number. Alternatively, the TIMESTAMP function can be used to convert the date component of the data into a valid timestamp. In a system with multiple nodes, there is no guarantee that this timestamp (alone) is unique.
7.2.75. Generate Unique Timestamps
The GENERATE_UNIQUE output can be processed using the TIMESTAMP function to obtain a unique timestamp value. Adding the CURRENT TIMEZONE special register to the TIMESTAMP output will convert it to local time:
SELECT CURRENT TIMESTAMP AS ts1
, TIMESTAMP(GENERATE_UNIQUE()) AS ts2
, TIMESTAMP(GENERATE_UNIQUE()) + CURRENT TIMEZONE AS ts3
FROM sysibm.sysdummy1;
ANSWER
TS1 | TS2 | TS3 |
---|---|---|
2007-01-19-18.12.33.587000 |
2007-01-19-22.12.28.434960 |
2007-01-19-18.12.28.434953 |
This code can be useful if one is doing a multi-row insert, and one wants each row inserted to have a distinct timestamp value. However, there are a few qualifications:
-
The timestamp values generated will be unique in themselves. But concurrent users may also generate the same values. There is no guarantee of absolute uniqueness.
-
Converting the universal-time value to local-time does not always return a value is equal to the CURRENT TIMESTAMP special register. As is illustrated above, the result can differ by a few seconds. This may cause business problems if one is relying on the value to be the "true time" when something happened.
Making Random
One thing that Db2 lacks is a random number generator that makes unique values. However, if we flip the characters returned in the GENERATE_UNIQUE output, we have something fairly close to what is needed. Unfortunately, Db2 also lacks a REVERSE function, so the data flipping has to be done the hard way.
SELECT u1
, SUBSTR(u1,20,1) CONCAT SUBSTR(u1,19,1) CONCAT
SUBSTR(u1,18,1) CONCAT SUBSTR(u1,17,1) CONCAT
SUBSTR(u1,16,1) CONCAT SUBSTR(u1,15,1) CONCAT
SUBSTR(u1,14,1) CONCAT SUBSTR(u1,13,1) CONCAT
SUBSTR(u1,12,1) CONCAT SUBSTR(u1,11,1) CONCAT
SUBSTR(u1,10,1) CONCAT SUBSTR(u1,09,1) CONCAT
SUBSTR(u1,08,1) CONCAT SUBSTR(u1,07,1) CONCAT
SUBSTR(u1,06,1) CONCAT SUBSTR(u1,05,1) CONCAT
SUBSTR(u1,04,1) CONCAT SUBSTR(u1,03,1) CONCAT
SUBSTR(u1,02,1) CONCAT SUBSTR(u1,01,1) AS U2
FROM (SELECT HEX(GENERATE_UNIQUE()) AS u1
FROM staff
WHERE id < 50) AS xxx
ORDER BY u2;
ANSWER
U1 | U2 |
---|---|
20000901131649119940000000 |
04991194613110900002 |
20000901131649119793000000 |
39791194613110900002 |
20000901131649119907000000 |
70991194613110900002 |
20000901131649119969000000 |
96991194613110900002 |
Pseudo-random
Observe above that we used a nested table expression to temporarily store the results of the GENERATE_UNIQUE calls. Alternatively, we could have put a GENERATE_UNIQUE call inside each SUBSTR, but these would have amounted to separate function calls, and there is a very small chance that the net result would not always be unique.
Using REVERSE Function
One can refer to a user-defined reverse function (see Reversing Field Contents for the definition code) to flip the U1 value, and thus greatly simplify the query:
SELECT u1
, SUBSTR(reverse(CHAR(u1)),7,20) AS u2
FROM
(SELECT HEX(GENERATE_UNIQUE()) AS u1
FROM staff
WHERE ID < 50) AS xxx
ORDER BY U2;
7.2.76. GETHINT
Returns the password hint, if one is found in the encrypted data.
-GETHINT function example
SELECT id
, name
, GETHINT(name2) AS hint
FROM
(SELECT id
, name
, ENCRYPT(name,'THAT IDIOT','MY BROTHER') AS name2
FROM staff
WHERE id < 30
) AS xxx
ORDER BY id;
ANSWER
ID | NAME | HINT |
---|---|---|
10 |
Sanders |
MY BROTHER |
20 |
Pernal |
MY BROTHER |
7.2.77. GRAPHIC
Converts the input (1st argument) to a graphic data type. The output length (2nd argument) is optional.
7.2.78. GREATEST
See MAX scalar function.
7.2.79. HASH
Returns a 128-bit, 160-bit, 256-bit or 512-bit hash of the input data, depending on the algorithm selected, and is intended for cryptographic purposes.
7.2.80. HASH4
Returns the 32-bit checksum hash of the input data. The function provides 232 distinct return values and is intended for data retrieval (lookups).
7.2.81. HASH8
Returns the 64-bit hash of an input string. The function provides 264 distinct return values and is intended for data retrieval (that is, lookups). The result for a particular input string differs depending on the endianness (big-endian or little-endian) of your system.
7.2.82. HASHEDVALUE
Returns the partition number of the row. The result is zero if the table is not partitioned. The output is of type integer, and is never null.
SELECT HASHEDVALUE(id) AS hvalue
FROM staff
WHERE id = 10;
ANSWER
HVALUE |
---|
0 |
The DBPARTITIONNUM function will generate a SQL error if the column/row used can not be related directly back to specific row in a real table. Therefore, one can not use this function on fields in GROUP BY statements, nor in some views. It can also cause an error when used in an outer join, and the target row failed to match in the join.
7.2.83. HEX
Returns the hexadecimal representation of a value. All input types are supported.
WITH temp1(n1) AS
(VALUES (-3)
UNION ALL
SELECT n1 + 1
FROM temp1
WHERE n1 < 3)
SELECT SMALLINT(n1) AS s
, HEX(SMALLINT(n1)) AS shx
, HEX(DEC(n1,4,0)) AS dhx
, HEX(DOUBLE(n1)) AS fhx
FROM temp1;
ANSWER
S | SHX | DHX | FHX |
---|---|---|---|
3 |
FDFF |
00003D |
00000000000008C0 |
-2 |
FEFF |
00002D |
00000000000000C0 |
-1 |
FFFF |
00001D |
000000000000F0BF |
0 |
0000 |
00000C |
0000000000000000 |
1 |
0100 |
00001C |
000000000000F03F |
2 |
0200 |
00002C |
0000000000000040 |
3 |
0300 |
00003C |
0000000000000840 |
SELECT c1
, HEX(c1) AS chx
, v1
, HEX(v1) AS vhx
FROM scalar;
ANSWER
C1 | CHX | V1 | VHX |
---|---|---|---|
ABCDEF |
414243444546 |
ABCDEF |
414243444546 |
ABCD |
414243442020 |
ABCD |
41424344 |
AB |
414220202020 |
AB |
4142 |
SELECT dt1
, HEX(dt1) AS dthx
, tm1
, HEX(tm1) AS tmhx
FROM scalar;
ANSWER
DT1 | DTHX | TM1 | TMHX |
---|---|---|---|
1996-04-22 |
19960422 |
23:58:58 |
235858 |
1996-08-15 |
19960815 |
15:15:15 |
151515 |
0001-01-01 |
00010101 |
00:00:00 |
000000 |
7.2.84. HEXTORAW
Returns a bit string representation of a hexadecimal character string.
7.2.85. HOUR
Returns the hour (as in hour of day) part of a time value. The output format is integer.
SELECT tm1
, HOUR(tm1) AS hr
FROM scalar
ORDER BY tm1;
ANSWER
TM1 | HR |
---|---|
00:00:00 |
0 |
15:15:15 |
15 |
23:58:58 |
23 |
7.2.86. HOURS_BETWEEN
Returns the number of full hours between the specified arguments.
7.2.87. IDENTITY_VAL_LOCAL
Returns the most recently assigned value (by the current user) to an identity column. The result type is decimal (31,0), regardless of the field type of the identity column. See Find Gaps in Values for detailed notes on using this function.
CREATE TABLE seq#
( ident_val INTEGER NOT NULL GENERATED ALWAYS AS IDENTITY
, cur_ts TIMESTAMP NOT NULL
, PRIMARY KEY (ident_val));
COMMIT;
INSERT INTO seq# VALUES(DEFAULT,CURRENT TIMESTAMP);
WITH temp (idval) AS
(VALUES (IDENTITY_VAL_LOCAL())
)
SELECT *
FROM temp;
ANSWER
IDVAL |
---|
1. |
7.2.88. INITCAP
Returns a string with the first character of each word converted to uppercase, using the UPPER function semantics, and the other characters converted to lowercase, using the LOWER function semantics.
7.2.89. INSERT
Insert one string in the middle of another, replacing a portion of what was already there. If the value to be inserted is either longer or shorter than the piece being replaced, the remainder of the data (on the right) is shifted either left or right accordingly in order to make a good fit.
Usage Notes
*Acceptable input types are varchar, clob(1M), and blob(1M). The first and last parameters must always have matching field types. To insert a new value in the middle of another without removing any of what is already there, set the third parameter to zero. The varchar output is always of length 4K.
SELECT name
, INSERT(name,3,2,'A')
, INSERT(name,3,2,'AB')
, INSERT(name,3,2,'ABC')
FROM staff
WHERE id < 40;
ANSWER (4K output fields shortened)
NAME | 2 | 3 | 4 |
---|---|---|---|
Sanders |
SaAers |
SaABers |
SaABCers |
Pernal |
PeAal |
PeABal |
PeABCal |
Marenghi |
MaAnghi |
MaABnghi |
MaABCnghi |
7.2.90. INSTR
Returns the starting position of a string (the search string) within another string (the source string). The INSTR scalar function is a synonym for the LOCATE_IN_STRING scalar function.
7.2.91. INSTR2
Returns the starting position, in 16-bit UTF-16 string units (CODEUNITS16), of a string within another string.
7.2.92. INSTR4
Returns the starting position, in 32-bit UTF-32 string units (CODEUNITS32), of a string within another string.
7.2.93. INSTRB
Returns the starting position, in bytes, of a string within another string.
7.2.94. INT or INTEGER
The INTEGER or INT function converts either a number or a valid character value into an integer. The character input can have leading and/or trailing blanks, and a sign indictor, but it can not contain a decimal point. Numeric decimal input works just fine.
SELECT d1
, INTEGER(d1)
, INT('+123')
, INT('-123')
, INT(' 123 ')
FROM scalar;
ANSWER
D1 | 2 | 3 | 4 | 5 |
---|---|---|---|---|
2.4 |
-2 |
123 |
-123 |
123 |
0.0 |
0 |
123 |
-123 |
123 |
1.8 |
1 |
123 |
-123 |
123 |
7.2.95. INTNAND , INTNOR, INTNXOR, and INTNNOT
These bitwise functions operate on the "two’s complement" representation of the integer value of the input arguments and return the result as a corresponding base 10 integer value.
7.2.96. JULIAN_DAY
Converts a date value into a number that represents the number of days since January the 1st, 4,713 BC. The output format is integer.
WITH temp1(dt1) AS
(VALUES ('0001-01-01-00.00.00')
, ('1752-09-10-00.00.00')
, ('2007-06-03-00.00.00')
, ('2007-06-03-23.59.59')
)
SELECT DATE(dt1) AS dt
, DAYS(dt1) AS dy
, JULIAN_DAY(dt1) AS dj
FROM temp1;
ANSWER
DT | DY | DJ |
---|---|---|
0001-01-01 |
1 |
1721426 |
1752-09-10 |
639793 |
2361218 |
2007-06-03 |
732830 |
2454255 |
2007-06-03 |
732830 |
2454255 |
Julian Days, A History
I happen to be a bit of an Astronomy nut, so what follows is a rather extended description of Julian Days - their purpose, and history (taken from the web). The Julian Day calendar is used in Astronomy to relate ancient and modern astronomical observations. The Babylonians, Egyptians, Greeks (in Alexandria), and others, kept very detailed
records of astronomical events, but they all used different calendars.
By converting all such observations to Julian Days, we can compare and correlate them. For example, a solar eclipse is said to have been seen at Ninevah on Julian day 1,442,454 and a lunar eclipse is said to have been observed at Babylon on Julian day number 1,566,839. These numbers correspond to the Julian Calendar dates -763-03-23 and -423-10-09 respectively). Thus the lunar eclipse occurred 124,384 days after the solar eclipse. The Julian Day number system was invented by Joseph Justus Scaliger (born 1540-08-05 J in Agen, France, died 1609-01-21 J in Leiden, Holland) in 1583. Although the term Julian Calendar derives from the name of Julius Caesar, the term Julian day number probably does not.
Evidently, this system was named, not after Julius Caesar, but after its inventor’s father, Julius Caesar Scaliger (1484-1558). The younger Scaliger combined three traditionally recognized temporal cycles of 28, 19 and 15 years to obtain a great cycle, the Scaliger cycle, or Julian period, of 7980 years (7980 is the least common multiple of 28, 19 and 15). The length of 7,980 years was chosen as the product of 28 times 19 times 15; these, respectively, are:
-
The number of years when dates recur on the same days of the week.
-
The lunar or Metonic cycle, after which the phases of the Moon recur on a particular day in the solar year, or year of the seasons.
-
The cycle of indiction, originally a schedule of periodic taxes or government requisitions in ancient Rome.
The first Scaliger cycle began with Year 1 on -4712-01-01 (Julian) and will end after 7980 years on 3267-12-31 (Julian), which is 3268-01-22 (Gregorian). 3268-01-01 (Julian) is the first day of Year 1 of the next Scaliger cycle. Astronomers adopted this system and adapted it to their own purposes, and they took noon GMT -4712-01-01 as their zero point. For astronomers a day begins at noon and runs until the next noon (so that the nighttime falls conveniently within one "day"). Thus they defined the Julian day number of a day as the number of days (or part of a day) elapsed since noon GMT on January 1st, 4713 B.C.E. This was not to the liking of all scholars using the Julian day number system, in particular, historians. For chronologists who start "days" at midnight, the zero point for the Julian day number system is 00:00 at the start of -4712-01-01 J, and this is day 0. This means that 200001-01 G is 2,451,545 JD. Since most days within about 150 years of the present have Julian day numbers beginning with "24", Julian day numbers within this 300-odd-year period can be abbreviated. In 1975 the convention of the modified Julian day number was adopted: Given a Julian day number JD, the modified Julian day number MJD is defined as MJD = JD - 2,400,000.5. This has two purposes:
-
Days begin at midnight rather than noon.
-
For dates in the period from 1859 to about 2130 only five digits need to be used to specify the date rather than seven.
MJD 0 thus corresponds to JD 2,400,000.5, which is twelve hours after noon on JD 2,400,000 = 1858-11-16. Thus MJD 0 designates the midnight of November 16th/17th, 1858, so day 0 in the system of modified Julian day numbers is the day 1858-11-17. The following SQL statement uses the JULIAN_DAY function to get the Julian Date for certain days. The same calculation is also done using hand-coded SQL.
SELECT bd
, JULIAN_DAY(bd)
, (1461 * (YEAR(bd) + 4800 +
(MONTH(bd)-14)/12))/4 +
( 367 * (MONTH(bd)- 2 -
12*((MONTH(bd)-14)/12)))/12 -
(3 * ((YEAR(bd) + 4900 +
(MONTH(bd)-14)/12)/100))/4 +
DAY(bd) - 32075
FROM
(SELECT birthdate AS bd
FROM employee
WHERE midinit = 'R'
) AS xxx
ORDER BY bd;
ANSWER
BD | 2 | 3 |
---|---|---|
1926-05-17 |
2424653 |
2424653 |
1936-03-28 |
2428256 |
2428256 |
1946-07-09 |
2432011 |
2432011 |
1955-04-12 |
2435210 |
2435210 |
Julian Dates
Many computer users think of the "Julian Date" as a date format that has a layout of "yynnn" or "yyyynnn" where "yy" is the year and "nnn" is the number of days since the start of the same. A more correct use of the term "Julian Date" refers to the current date according to the calendar as originally defined by Julius Caesar - which has a leap year on every fourth year. In the US/UK, this calendar was in effect until "1752-09-14". The days between the 3rd and 13th of September in 1752 were not used in order to put everything back in sync. In the 20th and 21st centuries, to derive the Julian date one must subtract 13 days from the relevant Gregorian date (e.g.1994-01-22 becomes 1994-01-07). The following SQL illustrates how to convert a standard Db2 Gregorian Date to an equivalent Julian Date (calendar) and a Julian Date (output format):
WITH temp1(dt1) AS
(VALUES ('2007-01-01')
, ('2007-01-02')
, ('2007-12-31')
)
SELECT DATE(dt1) AS dt
, DATE(dt1) - 13 DAYS AS dj1
, YEAR(dt1) * 1000 + DAYOFYEAR(dt1) AS dj2
FROM temp1;
ANSWER
DT | DJ1 | DJ2 |
---|---|---|
2007-01-01 |
2006-12-19 |
2007001 |
2007-01-02 |
2006-12-20 |
2007002 |
2007-12-31 |
2007-12-18 |
2007365 |
Db2 does not make allowances for the days that were not used when Englishspeaking countries converted from the Julian to the Gregorian calendar in 1752. |
7.2.97. LAST_DAY
Feturns a date or timestamp value that represents the last day of the month of the argument.
7.2.98. LCASE or LOWER
Converts a mixed or upper-case string to lower case. The output is the same data type and length as the input.
SELECT name
, LCASE(name) AS lname
, UCASE(name) AS uname
FROM staff
WHERE id < 30;
ANSWER
NAME | LNAME | UNAME |
---|---|---|
Sanders |
sanders |
SANDERS |
Pernal |
pernal |
PERNAL |
7.2.99. LEAST
See MIN scalar function.
7.2.100. LEFT
The LEFT function has two arguments: The first is an input string of type char, varchar, clob, or blob. The second is a positive integer value. The output is the left most characters in the string. Trailing blanks are not removed.
WITH temp1(c1) AS
(VALUES (' ABC')
, (' ABC ')
, ('ABC ')
)
SELECT c1
, LEFT(c1,4) AS c2
, LENGTH(LEFT(c1,4)) AS l2
FROM temp1;
ANSWER
C1 | C2 | L2 |
---|---|---|
ABC |
AB |
4 |
ABC |
ABC |
4 |
ABC |
ABC |
4 |
If the input is either char or varchar, the output is varchar(4000). A column this long is a nuisance to work with. Where possible, use the SUBSTR function to get around this problem.
7.2.101. LENGTH
Returns an integer value with the internal length of the expression (except for double-byte string types, which return the length in characters). The value will be the same for all fields in a column, except for columns containing varying-length strings.
SELECT LENGTH(d1)
, LENGTH(f1)
, LENGTH(s1)
, LENGTH(c1)
, LENGTH(RTRIM(c1))
FROM scalar;
ANSWER
1 | 2 | 3 | 4 | 5 |
---|---|---|---|---|
2 |
8 |
2 |
6 |
6 |
2 |
8 |
2 |
6 |
4 |
2 |
8 |
2 |
6 |
2 |
7.2.102. LENGTH2
Returns the length of expression in 16-bit UTF-16 string units (CODEUNITS16).
7.2.103. LENGTH4
Returns the length of expression in 32-bit UTF-32 string units (CODEUNITS32).
7.2.104. LENGTHB
Returns the length of expression in bytes.
7.2.105. LN or LOG
Returns the natural logarithm of the argument (same as LOG). The output format is double.
WITH temp1(n1) AS
(VALUES (1)
, (123)
, (1234)
, (12345)
, (123456)
)
SELECT n1
, LOG(n1) AS l1
FROM temp1;
ANSWER
N1 | L1 |
---|---|
1 |
+0.00000000000000E+000 |
123 |
+4.81218435537241E+000 |
1234 |
+7.11801620446533E+000 |
12345 |
+9.42100640177928E+000 |
123456 |
+1.17236400962654E+001 |
7.2.106. LOCATE
Returns an integer value with the absolute starting position of the first occurrence of the first string within the second string. If there is no match, the result is zero. The optional third parameter indicates where to start the search.
The result, if there is a match, is always the absolute position (i.e. from the start of the string), not the relative position (i.e. from the starting position).
WITH temp1 (c1) As
(VALUES ('abcdÄ')
, ('Äbcd')
, ('AÄ')
, ('ÁÄ')
)
SELECT c1
, LOCATE('Ä',c1) AS "l1"
, LOCATE('Ä',c1,2) AS "l2"
, LOCATE('Ä',c1,OCTETS) AS "l3"
, LOCATE('Ä',c1,CODEUNITS16) AS "l4"
, LOCATE('Ä',c1,2,CODEUNITS16) AS "l5"
FROM temp1;
ANSWER
C1 | l1 | l2 | l3 | l4 | l5 |
---|---|---|---|---|---|
abcdÄ |
5 |
5 |
5 |
5 |
5 |
Äbcd |
1 |
0 |
1 |
1 |
0 |
AÄ |
2 |
2 |
2 |
2 |
2 |
ÁÄ |
3 |
3 |
3 |
2 |
2 |
When a special character like "Á" is encountered before the find-string (see last line) the plain LOCATE returns the number of bytes searched, not the number of characters.
7.2.107. LOCATE_IN_STRING
Returns the starting position of a string (called the search-string ) within another string (called the source-string).
7.2.108. LOG10
Returns the common logarithm (base 10) of a number.
WITH temp1(n1) AS
(VALUES (1)
, (123)
, (1234)
, (12345)
, (123456)
)
SELECT n1
, LOG10(n1) AS l1
FROM temp1;
ANSWER
N1 | L1 |
---|---|
1 |
+0.00000000000000E+000 |
123 |
+2.08990511143939E+000 |
1234 |
+3.09131515969722E+000 |
12345 |
+4.09149109426795E+000 |
123456 |
+5.09151220162777E+000 |
7.2.109. LONG_VARCHAR
Converts the input (1st argument) to a long_varchar data type. The output length (2nd argument) is optional.
7.2.110. LONG_VARGRAPHIC
Converts the input (1st argument) to a long_vargraphic data type. The output length (2nd argument) is optional.
7.2.111. LOWER
See the description for the LCASE function.
7.2.112. LPAD
Pads a string on the left with a specified character string or with blanks.
7.2.113. LTRIM
Remove leading blanks, but not trailing blanks, from the argument.
WITH temp1(c1) AS
(VALUES (' ABC')
, (' ABC ')
, ('ABC ')
)
SELECT c1
, LTRIM(c1) AS c2
, LENGTH(LTRIM(c1)) AS l2
FROM temp1;
ANSWER
C1 | C2 | L2 |
---|---|---|
ABC |
ABC |
3 |
ABC |
ABC |
4 |
ABC |
ABC |
5 |
7.2.114. MAX
Returns the largest item from a list that must be at least two items long:
VALUES MAX(5, 8, 4)
ANSWER ⇒ 8
One can combine the MAX scalar and column functions to get the combined MAX value of a set of rows and columns:
SELECT MAX(MAX(salary, years, comm))
FROM staff;
ANSWER ⇒ 87654.50
Db2 knows which function is which because the MAX scalar value must have at least two input values, while the column function can only have one.
Null Processing
The MAX and MIN scalar functions return null if any one of the input list items is null. The MAX and MIN column functions ignore null values. They do however return null when no rows match.
7.2.115. MAX_CARDINALITY
Returns a BIGINT value that is the maximum number of values that an array can contain.
7.2.116. MICROSECOND
Returns the microsecond part of a timestamp (or equivalent) value. The output is integer.
SELECT ts1
, MICROSECOND(ts1)
FROM scalar
ORDER BY ts1;
ANSWER
TS1 | 2 |
---|---|
0001-01-01-00.00.00.000000 |
0 |
1996-04-22-23.58.58.123456 |
123456 |
1996-08-15-15.15.15.151515 |
151515 |
7.2.117. MIDNIGHT_SECONDS
Returns the number of seconds since midnight from a timestamp, time or equivalent value. The output format is integer.
SELECT ts1
, MIDNIGHT_SECONDS(ts1)
, HOUR(ts1)*3600 + MINUTE(ts1)*60 + SECOND(ts1)
FROM scalar
ORDER BY ts1;
ANSWER
TS1 | 2 | 3 |
---|---|---|
0001-01-01-00.00.00.000000 |
0 |
0 |
1996-04-22-23.58.58.123456 |
86338 |
86338 |
1996-08-15-15.15.15.151515 |
54915 |
54915 |
There is no single function that will convert the MIDNIGHT_SECONDS output back into a valid time value. However, it can be done using the following SQL:
WITH temp1 (ms) AS
(SELECT MIDNIGHT_SECONDS(ts1)
FROM scalar
)
SELECT ms
, SUBSTR(DIGITS(ms/3600), 9)
|| ':' ||
SUBSTR(DIGITS((ms-((MS/3600) * 3600))/60 ), 9)
|| ':' ||
SUBSTR(DIGITS(ms-((MS/60)*60)), 9) AS tm
FROM temp1
ORDER BY 1;
ANSWER
MS | TM |
---|---|
0 |
00:00:00 |
54915 |
15:15:15 |
86338 |
23:58:58 |
The following two identical timestamp values: "2005-07-15.24.00.00" and "2005-07-16.00.00.00" will return different MIDNIGHT_SECONDS results. See the chapter titled Quirks in SQL for a detailed discussion of this issue. |
7.2.118. MIN
Returns the smallest item from a list that must be at least two items long:
VALUES MIN(5, 8, 4)
ANSWER ⇒ 4
Null is returned if any one of the list items is null.
7.2.119. MINUTE
Returns the minute part of a time or timestamp (or equivalent) value. The output is integer.
SELECT ts1
, MINUTE(ts1)
FROM scalar
ORDER BY ts1;
ANSWER
TS1 | 2 |
---|---|
0001-01-01-00.00.00.000000 |
0 |
1996-04-22-23.58.58.123456 |
58 |
1996-08-15-15.15.15.151515 |
15 |
7.2.120. MINUTES_BETWEEN
Returns the number of full minutes between the specified arguments.
7.2.121. MOD
Returns the remainder (modulus) for the first argument divided by the second. In the following example the last column uses the MOD function to get the modulus, while the second to last column obtains the same result using simple arithmetic.
WITH temp1(n1,n2) AS
(VALUES (-31,+11)
UNION ALL
SELECT n1 + 13
, n2 - 4
FROM temp1
WHERE n1 < 60
)
SELECT n1
, n2
, n1/n2 AS div
, n1-((n1/n2)*n2) AS md1
, MOD(n1,n2) AS md2
FROM temp1
ORDER BY 1;
ANSWER
N1 | N2 | DIV | MD1 | MD2 |
---|---|---|---|---|
31 |
11 |
-2 |
-9 |
-9 |
-18 |
7 |
-2 |
-4 |
-4 |
-5 |
3 |
-1 |
-2 |
-2 |
8 |
-1 |
-8 |
0 |
0 |
21 |
-5 |
-4 |
1 |
1 |
34 |
-9 |
-3 |
7 |
7 |
47 |
-13 |
-3 |
8 |
8 |
60 |
-17 |
-3 |
9 |
9 |
7.2.122. MONTH
Returns an integer value in the range 1 to 12 that represents the month part of a date or timestamp (or equivalent) value.
MONTHNAME
Returns the name of the month (e.g. October) as contained in a date (or equivalent) value. The output format is varchar(100).
SELECT dt1
, MONTH(dt1)
, MONTHNAME(dt1)
FROM scalar
ORDER BY dt1;
ANSWER
DT1 | 2 | 3 |
---|---|---|
0001-01-01 |
1 |
January |
1996-04-22 |
4 |
April |
1996-08-15 |
8 |
August |
7.2.123. MONTHS_BETWEEN
Returns an estimate of the number of months between expression1 and expression2.
7.2.124. MULTIPLY_ALT
Returns the product of two arguments as a decimal value. Use this function instead of the multiplication operator when you need to avoid an overflow error because Db2 is putting aside too much space for the scale (i.e. fractional part of number) Valid input is any exact numeric type: decimal, integer, bigint, or smallint (but not float).
WITH temp1 (n1,n2) AS
(VALUES (DECIMAL(1234,10)
, DECIMAL(1234,10))
)
SELECT n1
, n2
, n1 * n2 AS p1
, "*"(n1,n2) AS p2
, MULTIPLY_ALT(n1,n2) AS p3
FROM temp1;
ANSWER
N1 | N2 | P1 | P2 | P3 |
---|---|---|---|---|
1234. |
1234. |
1522756. |
1522756. |
1522756. |
When doing ordinary multiplication of decimal values, the output precision and the scale is the sum of the two input precisions and scales - with both having an upper limit of 31. Thus, multiplying a DEC(10,5) number and a DEC(4,2) number returns a DEC(14,7) number. Db2 always tries to avoid losing (truncating) fractional digits, so multiplying a DEC(20,15) number with a DEC(20,13) number returns a DEC(31,28) number, which is probably going to be too small. The MULTIPLY_ALT function addresses the multiplication overflow problem by, if need be, truncating the output scale. If it is used to multiply a DEC(20,15) number and a DEC(20,13) number, the result is a DEC(31,19) number. The scale has been reduced to accommodate the required precision. Be aware that when there is a need for a scale in the output, and it is more than three digits, the function will leave at least three digits. Below are some examples of the output precisions and scales generated by this function:
INPUT#1 | INPUT#2 | RESULT "*" OPERATOR | RESULT MULTIPLY_ALT | M_A→SCALE TRUNCATD | M_A→PRECISION TRUNCATD |
---|---|---|---|---|---|
DEC(05,00) |
DEC(05,00) |
DEC(10,00) |
DEC(10,00) |
NO |
NO |
DEC(10,05) |
DEC(11,03) |
DEC(21,08) |
DEC(21,08) |
NO |
NO |
DEC(20,15) |
DEC(21,13) |
DEC(31,28) |
DEC(31,18) |
YES |
NO |
DEC(26,23) |
DEC(10,01) |
DEC(31,24) |
DEC(31,19) |
YES |
NO |
DEC(31,03) |
DEC(15,08) |
DEC(31,11) |
DEC(31,03) |
YES |
YES |
7.2.125. NCHAR
Returns a fixed-length national character string representation of a variety of data types.
7.2.126. NCLOB
Returns a NCLOB representation of any type of national character string.
7.2.127. NVARCHAR
Returns a varying-length national character string representation of a variety of data types.
7.2.128. NEXT_DAY
Returns a datetime value that represents the first weekday, named by string-expression, that is later than the date in expression
7.2.129. NEXT_MONTH
Returns the first day of the next month after the specified date.
7.2.130. NEXT_QUARTER
Returns the first day of the next quarter after the date specified by the input.
7.2.131. NEXT_WEEK
Returns the first day of the next week after the specified date. Sunday is considered the first day of that new week.
7.2.132. NEXT_YEAR
Returns the first day of the year follows the year containing the date specified by the input.
7.2.133. NORMALIZE_DECFLOAT
Removes any trailing zeros from a DECFLOAT value.
WITH temp1 (d1) AS
(VALUES (DECFLOAT(1))
, (DECFLOAT(1.0))
, (DECFLOAT(1.00))
, (DECFLOAT(1.000))
, (DECFLOAT('12.3'))
, (DECFLOAT('12.30'))
, (DECFLOAT(1.2e4))
, (DECFLOAT('1.2e4'))
, (DECFLOAT(1.2e-3))
, (DECFLOAT('1.2e-3'))
)
SELECT d1
, NORMALIZE_DECFLOAT(d1) AS d2
FROM temp1;
ANSWER
D1 | D2 |
---|---|
1 |
1 |
1.0 |
1 |
1.00 |
1 |
1.000 |
1 |
12.3 |
12.3 |
12.30 |
12.3 |
12000 |
1.2E+4 |
1.2E+4 |
1.2E+4 |
0.001200000000000000 |
0.0012 |
0.0012 |
0.0012 |
7.2.134. NOW
Returns a timestamp based on a reading of the time-of-day clock when the SQL statement is executed at the current server.
7.2.135. NULLIF
Returns null if the two values being compared are equal, otherwise returns the first value.
SELECT s1
, NULLIF(s1, 0)
, c1
, NULLIF(c1, 'AB')
FROM scalar
WHERE NULLIF(0, 0) IS NULL;
ANSWER
S1 | 2 | C1 | 4 |
---|---|---|---|
2 |
-2 |
ABCDEF |
ABCDEF |
0 |
- |
ABCD |
ABCD |
1 |
1 |
AB |
- |
7.2.136. NVL
Same as COALESCE.
7.2.137. NVL2
Returns the second argument when the first argument is not NULL. If the first argument is NULL, the third argument is returned.
7.2.138. OCTET_LENGTH
Returns the length of the input expression in octets (bytes).
WITH temp1 (c1) AS
(VALUES (CAST('ÁÉÌ' AS VARCHAR(10)))
)
SELECT c1 AS C1
, LENGTH(c1) AS LEN
, OCTET_LENGTH(c1) AS OCT
, CHAR_LENGTH(c1,OCTETS) AS L08
, CHAR_LENGTH(c1,CODEUNITS16) AS L16
, CHAR_LENGTH(c1,CODEUNITS32) AS L32
FROM temp1;
ANSWER
C1 | LEN | OCT | L08 | L16 | L32 |
---|---|---|---|---|---|
ÁÉÌ |
6 |
6 |
6 |
3 |
3 |
7.2.139. OVERLAY
Overlay (i.e. replace) some part of a string with another string. There are five parameters:
-
The source string to be edited.
-
The new string to be inserted. This value can be zero length, but must be provided.
-
Start position for new string, and also to where start deleting. This value must be between one and the string length.
-
Number of bytes in the source to be overlaid. This value is optional.
-
The code unit to use.
There are two function notations. One uses keywords to separate each parameter. The other uses commas.
WITH temp1 (txt) AS
(VALUES('abcded')
, ('addd')
, ('adq')
)
SELECT txt
, OVERLAY(txt,'XX',3,1,OCTETS) AS "s3f1"
, OVERLAY(txt,'XX',2, OCTETS) AS "s2f0"
, OVERLAY(txt,'XX',1,1,OCTETS) AS "s1f1"
, OVERLAY(txt,'XX',2,2,OCTETS) AS "s2f2"
FROM temp1;
ANSWER
TXT | s3f1 | s2f0 | s1f1 | s2f2 |
---|---|---|---|---|
abcded |
abXXded |
aXXcded |
XXbcded |
aXXded |
addd |
adXXd |
aXXdd |
XXddd |
aXXd |
adq |
adXX |
aXXq |
XXdq |
aXX |
7.2.140. PARAMETER
The PARAMETER function represents a position in an SQL statement where the value is provided dynamically by XQuery as part of the invocation of the db2-fn:sqlquery function.
7.2.141. POSITION
Returns an integer value with the absolute starting position of the first occurrence of the first string within the second string. If there is no match, the result is zero. The third parameter indicates what code-unit to use.
When a special character like "Á" is encountered before the find-string (see last two lines in next example) the plain OCTETS search returns the number of bytes searched, not the number of characters:
WITH temp1 (c1) As
(VALUES ('Ä')
, ('aÄ')
, ('ÁÄ')
, ('ÁÁÄ')
)
SELECT c1
, POSITION('Ä',c1,OCTETS) AS "p1"
, POSITION('Ä',c1,CODEUNITS16) AS "p2"
, POSITION('Ä',c1,CODEUNITS32) AS "p3"
, POSITION('Ä' IN c1 USING OCTETS) AS "p4"
FROM temp1;
ANSWER
C1 | p1 | p2 | p3 | p4 |
---|---|---|---|---|
Ä |
1 |
1 |
1 |
1 |
aÄ |
2 |
2 |
2 |
2 |
ÁÄ |
3 |
2 |
2 |
3 |
ÁÁÄ |
5 |
3 |
3 |
5 |
The LOCATE function (see LOCATE) is very similar to the POSITION function. It has the additional capability of being able to start the search at any position in the search string.
7.2.142. POSSTR
Returns the position at which the second string is contained in the first string. If there is no match the value is zero. The test is case sensitive. The output format is integer.
POSSTR function example
SELECT c1
, POSSTR(c1,' ') AS p1
, POSSTR(c1,'CD') AS p2
, POSSTR(c1,'cd') AS p3
FROM scalar
ORDER BY 1;
ANSWER
C1 | P1 | P2 | P3 |
---|---|---|---|
AB |
3 |
0 |
0 |
ABCD |
5 |
3 |
0 |
ABCDEF |
0 |
3 |
0 |
POSSTR vs. LOCATE
The LOCATE and POSSTR functions are very similar. Both look for matching strings searching from the left. The only functional differences are that the input parameters are reversed and the LOCATE function enables one to begin the search at somewhere other than the start. When either is suitable for the task at hand, it is probably better to use the POSSTR function because it is a SYSIBM function and so should be faster.
SELECT c1
, POSSTR(c1, ' ') AS p1
, LOCATE(' ', c1) AS l1
, POSSTR(c1, 'CD') AS p2
, LOCATE('CD', c1) AS l2
, POSSTR(c1, 'cd') AS p3
, LOCATE('cd', c1) AS l3
, LOCATE('D', c1, 2) AS l4
FROM scalar
ORDER BY 1;
ANSWER
C1 | P1 | L1 | P2 | L2 | P3 | L3 | L4 |
---|---|---|---|---|---|---|---|
AB |
3 |
3 |
0 |
0 |
0 |
0 |
0 |
ABCD |
5 |
5 |
3 |
3 |
0 |
0 |
4 |
ABCDEF |
0 |
0 |
3 |
3 |
0 |
0 |
4 |
7.2.143. POWER or POW
Returns the value of the first argument to the power of the second argument
WITH temp1(n1) AS
(VALUES (1)
, (10)
, (100)
)
SELECT n1
, POWER(n1, 1) AS p1
, POWER(n1, 2) AS p2
, POWER(n1, 3) AS p3
FROM temp1;
ANSWER
N1 | P1 | P2 | P3 |
---|---|---|---|
1 |
1 |
1 |
1 |
10 |
10 |
100 |
1000 |
100 |
100 |
10000 |
1000000 |
7.2.144. QUANTIZE
Convert the first input parameter to DECFLOAT, using the second parameter as a mask. The specific value of the second parameter is irrelevant. But the precision (i.e. number of digits after the decimal point) defines the precision of the DECFLOAT result:
WITH temp1 (d1, d2) AS
(VALUES (+1.23, DECFLOAT(1.0))
, (+1.23, DECFLOAT(1.00))
, (-1.23, DECFLOAT(1.000))
, (+123, DECFLOAT(9.8765))
, (+123, DECFLOAT(1E-3))
, (+123, DECFLOAT(1E+3))
, (SQRT(2), DECFLOAT(0.0))
, (SQRT(2), DECFLOAT('1E-5'))
, (SQRT(2), DECFLOAT( 1E-5 ))
)
SELECT QUANTIZE(d1,d2) AS q
FROM temp1;
ANSWER
Q |
---|
1.2 |
1.23 |
-1.230 |
123.0000 |
123.000 |
123 |
1.4 |
1.41421 |
1.414213562373095100000 |
Observe that the function returns a very different result when the second parameter is '1E-5' vs. 1E-5 (i.e. with no quotes). This is because the number 1E-5 is not precisely converted to the DECFLOAT value 0.00001, as the following query illustrates:
-DECFLOAT conversion example
WITH temp1 (d1) AS
(VALUES (DECFLOAT('1E-5'))
, (DECFLOAT( 1E-5 ))
)
SELECT d1
FROM temp1;
ANSWER
D1 |
---|
0.00001 |
0.000010000000000000001 |
7.2.145. QUARTER
Returns an integer value in the range 1 to 4 that represents the quarter of the year from a date or timestamp (or equivalent) value.
7.2.146. RADIANS
Returns the number of radians converted from the input, which is expressed in degrees. The output format is double.
7.2.147. RAISE_ERROR
Causes the SQL statement to stop and return a user-defined error message when invoked. There are a lot of usage restrictions involving this function, see the SQL Reference for details.
SELECT s1
, CASE
WHEN s1 < 1 THEN s1
ELSE RAISE_ERROR('80001', c1)
END AS s2
FROM scalar;
ANSWER
S1 | S2 |
---|---|
-2 |
-2 |
0 |
0 |
SQLSTATE=80001 |
The SIGNAL statement (see SIGNAL Statement) is the statement equivalent of this function.
7.2.148. RAND
WARNING: Using the RAND function in a predicate can result in unpredictable results. See RAND in Predicate for a detailed description of this issue. To randomly sample the rows in a table reliably and efficiently, use the TABLESAMPLE feature. See Randomly Sample Data for details.
Returns a pseudo-random floating-point value in the range of zero to one inclusive. An optional seed value can be provided to get reproducible random results. This function is especially useful when one is trying to create somewhat realistic sample data.
Usage Notes
-
The RAND function returns any one of 32K distinct floating-point values in the range of zero to one inclusive. Note that many equivalent functions in other languages (e.g. SAS) return many more distinct values over the same range.
-
The values generated by the RAND function are evenly distributed over the range of zero to one inclusive.
-
A seed can be provided to get reproducible results. The seed can be any valid number of type integer. Note that the use of a seed alone does not give consistent results. Two different SQL statements using the same seed may return different (but internally consistent) sets of pseudo-random numbers.
-
If the seed value is zero, the initial result will also be zero. All other seed values return initial values that are not the same as the seed. Subsequent calls of the RAND function in the same statement are not affected.
-
If there are multiple references to the RAND function in the same SQL statement, the seed of the first RAND invocation is the one used for all.
-
If the seed value is not provided, the pseudo-random numbers generated will usually be unpredictable. However, if some prior SQL statement in the same thread has already invoked the RAND function, the newly generated pseudo-random numbers "may" continue where the prior ones left off.
Typical Output Values
The following recursive SQL generates 100,000 random numbers using two as the seed value. The generated data is then summarized using various Db2 column functions:
WITH temp (num, ran) AS
(VALUES (INT(1)
, RAND(2))
UNION ALL
SELECT num + 1
, RAND()
FROM temp
WHERE num < 100000
)
SELECT COUNT(*) AS #rows
, COUNT(DISTINCT ran) AS #values
, DEC(AVG(ran), 7, 6) AS avg_ran
, DEC(STDDEV(ran), 7, 6) AS std_devm
, DEC(MIN(ran), 7, 6) AS min_ran
, DEC(MAX(ran), 7, 6) AS max_ran
, DEC(MAX(ran), 7, 6) - DEC(MIN(ran), 7, 6) AS range
, DEC(VAR(ran), 7, 6) AS variance
FROM temp;
ANSWER
#ROWS | #VALUES | AVG_RAN | STD_DEVM | MIN_RAN | MAX_RAN | RANGE | VARIANCE |
---|---|---|---|---|---|---|---|
100000 |
31242 |
0.499838 |
0.288706 |
0.000000 |
1.000000 |
1.000000 |
0.083351 |
Observe that less than 32K distinct numbers were generated. Presumably, this is because the RAND function uses a 2-byte carry. Also observe that the values range from a minimum of zero to a maximum of one.
Unlike most, if not all, other numeric functions in Db2, the RAND function returns different results in different flavors of Db2. |
Reproducible Random Numbers
The RAND function creates pseudo-random numbers. This means that the output looks random, but it is actually made using a very specific formula. If the first invocation of the function uses a seed value, all subsequent invocations will return a result that is explicitly derived from the initial seed. To illustrate this concept, the following statement selects six random numbers. Because of the use of the seed, the same six values will always be returned when this SQL statement is invoked (when invoked on my machine):
SELECT deptno AS dno
, RAND(0) AS ran
FROM department
WHERE deptno < 'E'
ORDER BY 1;
ANSWER
DNO | RAN |
---|---|
A00 |
+1.15970336008789E-003 |
B01 |
+2.35572374645222E-001 |
C01 |
+6.48152104251228E-001 |
D01 |
+7.43736075930052E-002 |
D11 |
+2.70241401409955E-001 |
D21 |
+3.60026856288339E-001 |
To get random numbers that are not reproducible, simply leave the seed out of the first invocation of the RAND function. To illustrate, the following statement will give differing results with each invocation:
SELECT deptno AS dno
, RAND() AS ran
FROM department
WHERE deptno < 'D'
ORDER BY 1;
ANSWER
DNO | RAN |
---|---|
A00 |
+2.55287331766717E-001 |
B01 |
+9.85290078432569E-001 |
C01 |
+3.18918424024171E-001 |
Use of the seed value in the RAND function has an impact across multiple SQL statements. For example, if the above two statements were always run as a pair (with nothing else run in between), the result from the second would always be the same. |
Generating Random Values
Imagine that we need to generate a set of reproducible random numbers that are within a certain range (e.g. 5 to 15). Recursive SQL can be used to make the rows, and various scalar functions can be used to get the right range of data. In the following example we shall make a list of three columns and ten rows. The first field is a simple ascending sequence. The second is a set of random numbers of type smallint in the range zero to 350 (by increments of ten). The last is a set of random decimal numbers in the range of zero to 10,000.
WITH Temp1 (col1, col2, col3) AS
(VALUES (0
, SMALLINT(RAND(2)*35)*10
, DECIMAL(RAND()*10000, 7, 2)
)
UNION ALL
SELECT col1 + 1
, SMALLINT(RAND()*35)*10
, DECIMAL(RAND()*10000, 7, 2)
FROM temp1
WHERE col1 + 1 < 10
)
SELECT *
FROM temp1;
ANSWER
COL1 | COL2 | COL3 |
---|---|---|
0 |
0 |
9342.32 |
1 |
250 |
8916.28 |
2 |
310 |
5430.76 |
3 |
150 |
5996.88 |
4 |
110 |
8066.34 |
5 |
50 |
5589.77 |
6 |
130 |
8602.86 |
7 |
340 |
184.94 |
8 |
310 |
5441.14 |
9 |
70 |
9267.55 |
See the section titled "Making Sample Data" for more detailed examples of using the RAND function and recursion to make test data. |
Making Many Distinct Random Values
The RAND function generates 32K distinct random values. To get a larger set of (evenly distributed) random values, combine the result of two RAND calls in the manner shown below for the RAN2 column:
WITH temp1 (col1,ran1,ran2) AS
(VALUES (0,
RAND(2),
RAND()+(RAND()/1E5)
)
UNION ALL
SELECT col1 + 1
, RAND()
, RAND() +(RAND()/1E5)
FROM temp1
WHERE col1 + 1 < 30000
)
SELECT COUNT(*) AS col#1
, COUNT(DISTINCT ran1) AS ran#1
, COUNT(DISTINCT ran2) AS ran#2
FROM temp1;
ANSWER
COL#1 | RAN#1 | RAN#2 |
---|---|---|
30000 |
19698 |
29998 |
Observe that we do not multiply the two values that make up the RAN2 column above. If we did this, it would skew the average (from 0.5 to 0.25), and we would always get a zero whenever either one of the two RAND functions returned a zero.
The GENERATE_UNIQUE function can also be used to get a list of distinct values, and actually does a better job that the RAND function. With a bit of simple data manipulation (see GENERATE_UNIQUE), these values can also be made random. |
Selecting Random Rows, Percentage
Using the RAND function in a predicate can result in unpredictable results. See RAND in Predicate for a detailed description of this issue. |
Imagine that you want to select approximately 10% of the matching rows from some table. The predicate in the following query will do the job:
SELECT id
, name
FROM staff
WHERE RAND() < 0.1
ORDER BY id;
ANSWER
ID | NAME |
---|---|
140 |
Fraye |
190 |
Sneider |
290 |
Quill |
The RAND function randomly generates values in the range of zero through one, so the above query should return approximately 10% the matching rows. But it may return anywhere from zero to all of the matching rows - depending on the specific values that the RAND function generates. If the number of rows to be processed is large, then the fraction (of rows) that you get will be pretty close to what you asked for. But for small sets of matching rows, the result set size is quite often anything but what you wanted.
Selecting Random Rows, Number
The following query will select five random rows from the set of matching rows. It begins (in the inner-most nested table expression) by using the RAND function to assign random values to each matching row. Subsequently, the ROW_NUMBER function is used to sequence each random value. Finally, those rows with the five lowest row numbers are selected:
SELECT id
, name
FROM
(SELECT s2.*
, ROW_NUMBER() OVER(ORDER BY r1) AS r2
FROM
(SELECT s1.*
, RAND() AS r1
FROM staff s1
WHERE id <= 100) AS s2
) as s3
WHERE r2 <= 5
ORDER BY id;
ANSWER
ID | NAME |
---|---|
10 |
Sanders |
30 |
Marenghi |
40 |
O’Brien |
70 |
Rothman |
100 |
Plotz |
Use in DML
Imagine that in act of inspired unfairness, we decided to update a selected set of employee’s salary to a random number in the range of zero to $10,000. This too is easy:
UPDATE staff
SET salary = RAND()*10000
WHERE id < 50;
7.2.149. RAWTOHEX
Returns a hexadecimal representation of a value as a character string.
7.2.150. REAL
Returns a single-precision floating-point representation of a number.
SELECT n1 AS dec
, DOUBLE(n1) AS dbl
, REAL(n1) AS rel
, INTEGER(n1) AS int
, BIGINT(n1) AS big
FROM
(SELECT 1234567890.123456789012345678901 AS n1
FROM staff
WHERE id = 10) AS xxx;
ANSWER
DEC | DBL | REL | INT | BIG |
---|---|---|---|---|
1234567890.123456789012345678901 |
1.23456789012346e+009 |
1.234568e+009 |
1234567890 |
1234567890 |
7.2.151. REC2XML
Returns a string formatted with XML tags, containing column names and column data.
7.2.152. REGEXP_COUNT
Returns a count of the number of times that a regular expression pattern is matched in a string.
7.2.153. REGEXP_EXTRACT
Returns one occurrence of a substring of a string that matches the regular expression pattern.
7.2.154. REGEXP_INSTR
Returns the starting or ending position of the matched substring, depending on the value of the return_option argument.
7.2.155. REGEXP_LIKE
Returns a boolean value indicating if the regular expression pattern is found in a string. The function can be used only where a predicate is supported.
7.2.156. REGEXP_MATCH_COUNT
Returns a count of the number of times that a regular expression pattern is matched in a string.
7.2.157. REGEXP_REPLACE
Returns a modified version of the source string where occurrences of the regular expression pattern found in the source string are replaced with the specified replacement string.
7.2.158. REGEXP_SUBSTR
Returns one occurrence of a substring of a string that matches the regular expression pattern.
7.2.159. REPEAT
Repeats a character string "n" times.
SELECT id
, CHAR(REPEAT(name, 3), 40)
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | 2 |
---|---|
10 |
SandersSandersSanders |
20 |
PernalPernalPernal |
30 |
MarenghiMarenghiMarenghi |
7.2.160. REPLACE
Replaces all occurrences of one string with another. The output is of type varchar(4000).
SELECT c1
, REPLACE(c1,'AB','XY') AS r1
, REPLACE(c1,'BA','XY') AS r2
FROM scalar;
ANSWER
C1 | R1 | R2 |
---|---|---|
ABCDEF |
XYCDEF |
ABCDEF |
ABCD |
XYCD |
ABCD |
AB |
XY |
AB |
The REPLACE function is case sensitive. To replace an input value, regardless of the case, one can nest the REPLACE function calls. Unfortunately, this technique gets to be a little tedious when the number of characters to replace is large.
SELECT c1
, REPLACE(
REPLACE(
REPLACE(
REPLACE(c1, 'AB', 'XY')
, 'ab', 'XY')
, 'Ab', 'XY')
, 'aB', 'XY')
FROM scalar;
ANSWER
C1 | R1 |
---|---|
ABCDEF |
XYCDEF |
ABCD |
XYCD |
AB |
XY |
7.2.161. RID
Returns the RID (i.e. row identifier - of type BIGINT) for the matching row. The row identifier contains the page number, and the row number within the page. A unique table identifier must be provided.
SELECT id
, salary
, RID(staff) AS staff_rid
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SALARY | STAFF_RID |
---|---|---|
10 |
98357.50 |
100663300 |
20 |
78171.25 |
100663301 |
30 |
77506.75 |
100663302 |
The RID function is similar to the RID_BIT function, but less useful (e.g. does not work in a DPF environment). All subsequent examples will refer to the RID_BIT function.
7.2.162. RID_BIT
Returns the row identifier, of type VARCHAR(16) FOR BIT DATA, for the row. The row identifier contains the page number, and the row number within the page. The only input value, which must be provided, is the (unique) table identifier. The table must be listed in the subsequent FROM statement.
SELECT id
, RID_BIT(staff) AS rid_bit
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | RID_BIT |
---|---|
10 |
x'04000006000000000000000000FCE14D' |
20 |
x'05000006000000000000000000FCE14D' |
30 |
x'06000006000000000000000000FCE14D' |
When the same table is referenced twice in the FROM, the correlation name must be used:
SELECT s1.id
, RID_BIT(s1) AS rid_bit
FROM staff s1
, staff s2
WHERE s1.id < 40
AND s1.id = s2.id - 10
ORDER BY s1.id;
ANSWER
ID | RID_BIT |
---|---|
10 |
x'04000006000000000000000000FCE14D' |
20 |
x'05000006000000000000000000FCE14D' |
30 |
x'06000006000000000000000000FCE14D' |
The RID function can be used in a predicate to uniquely identify a row. To illustrate, the following query gets the RID and ROW CHANGE TOKEN for a particular row:
SELECT id
, salary
, RID_BIT(staff)
, ROW CHANGE TOKEN FOR staff
FROM staff
WHERE id = 20;
ANSWER
ID | SALARY | 3 | 4 |
---|---|---|---|
20 |
78171.25 |
x'05000006000000000000000000FCE14D' |
3999250443959009280 |
If at some subsequent point in time we want to update this row, we can use the RID value to locate it directly, and the ROW CHANGE TOKEN to confirm that it has not been changed:
UPDATE staff
SET salary = salary * 1.1
WHERE RID_BIT(staff) = x'05000006000000000000000000FCE14D'
AND ROW CHANGE TOKEN FOR staff = 3999250443959009280;
Usage Notes
-
The table name provided to the RID_BIT function must uniquely identify the table being processed. If a view is referenced, the view must be deletable.
-
The RID_BIT function will return a different value for a particular row a REORG is run.
-
The ROW CHANGE TOKEN changes every time a row is updated, including when an update is rolled back. So after a rollback the value will be different from what it was at the beginning of the unit of work.
-
The ROW CHANGE TOKEN is unique per page, not per row. So if any other row in the same page is changed, the prior update will not match. This is called a "false negative". To avoid, define a ROW CHANGE TIMESTAMP column for the table, as the value in this field is unique per row.
7.2.163. RIGHT
Has two arguments: The first is an input string of type char, varchar, clob, or blob. The second is a positive integer value. The output, of type varchar(4000), is the right most characters in the string.
WITH temp1(c1) AS
(VALUES (' ABC')
, (' ABC ')
, ('ABC ')
)
SELECT c1
, RIGHT(c1, 4) AS c2
, LENGTH(RIGHT(c1,4)) AS l2
FROM temp1;
ANSWER
C1 | C2 | L2 |
---|---|---|
ABC |
ABC |
4 |
ABC |
ABC |
4 |
ABC |
BC |
4 |
7.2.164. ROUND
Rounds the rightmost digits of number (1st argument). If the second argument is positive, it rounds to the right of the decimal place. If the second argument is negative, it rounds to the left. A second argument of zero results rounds to integer. The input and output types are the same, except for decimal where the precision will be increased by one - if possible. Therefore, a DEC(5,2)field will be returned as DEC(6,2), and a DEC(31,2) field as DEC(31,2). To truncate instead of round, use the TRUNCATE function.
WITH temp1(d1) AS
(VALUES (123.400)
, ( 23.450)
, ( 3.456)
, ( .056))
SELECT d1
, DEC(ROUND(d1,+2),6,3) AS p2
, DEC(ROUND(d1,+1),6,3) AS p1
, DEC(ROUND(d1,+0),6,3) AS p0
, DEC(ROUND(d1,-1),6,3) AS n1
, DEC(ROUND(d1,-2),6,3) AS v
FROM temp1;
ANSWER
D1 | P2 | P1 | P0 | N1 | N2 |
---|---|---|---|---|---|
123.400 |
123.400 |
123.400 |
123.000 |
120.000 |
100.000 |
23.450 |
23.450 |
23.400 |
23.000 |
20.000 |
0.000 |
3.456 |
3.460 |
3.500 |
3.000 |
0.000 |
0.000 |
0.056 |
0.060 |
0.100 |
0.000 |
0.000 |
0.000 |
7.2.165. ROUND_TIMESTAMP
Returns a TIMESTAMP based on a provided argument (expression), rounded to the unit specified in another argument (format-string).
7.2.166. RPAD
Returns a string composed of string-expression padded on the right, with pad or blanks.
7.2.167. RTRIM
Trims the right-most blanks of a character string.
SELECT c1
, RTRIM(c1) AS r1
, LENGTH(c1) AS r2
, LENGTH(RTRIM(c1)) AS r3
FROM scalar;
ANSWER
C1 | R1 | R2 | R3 |
---|---|---|---|
ABCDEF |
ABCDEF |
6 |
6 |
ABCD |
ABCD |
6 |
4 |
AB |
AB |
6 |
2 |
7.2.168. SECLABEL Functions
The SECLABEL, SECLABEL_BY_NAME, and SECLABEL_BY_CHAR functions are used to process security labels. See the SQL Reference for more details.
7.2.169. SECLABEL_BY_NAME
Returns the specified security label. The security label returned has a data type of DB2SECURITYLABEL. Use this function to insert a named security label.
7.2.170. SECLABEL_TO_CHAR
Accepts a security label and returns a string that contains all elements in the security label. The string is in the security label string format.
7.2.171. SECOND
Returns the second (of minute) part of a time or timestamp (or equivalent) value.
7.2.172. SECONDS_BETWEEN
Returns the number of full seconds between the specified arguments.
7.2.173. SIGN
Returns -1 if the input number is less than zero, 0 if it equals zero, and +1 if it is greater than zero. The input and output types will equal, except for decimal which returns double.
SELECT d1
, SIGN(d1)
, f1
, SIGN(f1)
FROM scalar;
ANSWER (float output shortened)
D1 | 2 | F1 | 4 |
---|---|---|---|
-2.4 |
-1.000E+0 |
-2.400E+0 |
-1.000E+0 |
0.0 |
+0.000E+0 |
+0.000E+0 |
+0.000E+0 |
1.8 |
+1.000E+0 |
+1.800E+0 |
+1.000E+0 |
7.2.174. SIN
Returns the SIN of the argument where the argument is an angle expressed in radians. The output format is double.
WITH temp1(n1) AS
(VALUES (0)
UNION ALL
SELECT n1 + 10
FROM temp1
WHERE n1 < 80
)
SELECT n1
, DEC(RADIANS(n1),4,3) AS ran
, DEC(SIN(RADIANS(n1)),4,3) AS sin
, DEC(TAN(RADIANS(n1)),4,3) AS tan
FROM temp1;
ANSWER
N1 | RAN | SIN | TAN |
---|---|---|---|
0 |
0.000 |
0.000 |
0.000 |
10 |
0.174 |
0.173 |
0.176 |
20 |
0.349 |
0.342 |
0.363 |
30 |
0.523 |
0.500 |
0.577 |
40 |
0.698 |
0.642 |
0.839 |
50 |
0.872 |
0.766 |
1.191 |
60 |
1.047 |
0.866 |
1.732 |
70 |
1.221 |
0.939 |
2.747 |
80 |
1.396 |
0.984 |
5.671 |
7.2.175. SINH
Returns the hyperbolic sin for the argument, where the argument is an angle expressed in radians. The output format is double.
7.2.176. SMALLINT
Converts either a number or a valid character value into a smallint value.
SELECT d1
, SMALLINT(d1)
, SMALLINT('+123')
, SMALLINT('-123')
, SMALLINT(' 123 ')
FROM scalar;
ANSWER
D1 | 2 | 3 | 4 | 5 |
---|---|---|---|---|
-2.4 |
-2 |
123 |
-123 |
123 |
0.0 |
0 |
123 |
-123 |
123 |
1.8 |
1 |
123 |
-123 |
123 |
7.2.177. SNAPSHOT Functions
The various SNAPSHOT functions can be used to analyze the system. They are beyond the scope of this book. Refer instead to the Db2 System Monitor Guide and Reference.
7.2.178. SOUNDEX
Returns a 4-character code representing the sound of the words in the argument. Use the DIFFERENCE function to convert words to soundex values and then compare.
SELECT a.name AS n1
, SOUNDEX(a.name) AS s1
, b.name AS n2
, SOUNDEX(b.name) AS s2
, DIFFERENCE(a.name,b.name) AS df
FROM staff a
, staff b
WHERE a.id = 10
AND b.id > 150 AND b.id < 250
ORDER BY df DESC
, n2 ASC;
ANSWER
N1 | S1 | N2 | S2 | DF |
---|---|---|---|---|
Sanders |
S536 |
Sneider |
S536 |
4 |
Sanders |
S536 |
Smith |
S530 |
3 |
Sanders |
S536 |
Lundquist |
L532 |
2 |
Sanders |
S536 |
Daniels |
D542 |
1 |
Sanders |
S536 |
Molinare |
M456 |
1 |
Sanders |
S536 |
Scoutten |
S350 |
1 |
Sanders |
S536 |
Abrahams |
A165 |
0 |
Sanders |
S536 |
Kermisch |
K652 |
0 |
Sanders |
S536 |
Lu |
L000 |
0 |
SOUNDEX Formula
There are several minor variations on the SOUNDEX algorithm. Below is one example:
-
The first letter of the name is left unchanged.
-
The letters W and H are ignored.
-
The vowels, A, E, I, O, U, and Y are not coded, but are used as separators (see last item).
-
The remaining letters are coded as:
Letters |
Value |
B, P, F, V |
1 |
C, G, J, K, Q, S, X, Z 2 |
2 |
D, T |
3 |
L |
4 |
M, N |
5 |
R |
6 |
-
Letters that follow letters with same code are ignored unless a separator (see the third item above) precedes them.
The result of the above calculation is a four byte value. The first byte is a character as defined in step one. The remaining three bytes are digits as defined in steps two through four.
Output longer than four bytes is truncated If the output is not long enough, it is padded on the right with zeros. The maximum number of distinct values is 8,918.
The SOUNDEX function is something of an industry standard that was developed several decades ago. Since that time, several other similar functions have been developed. You may want to investigate writing your own Db2 function to search for similarsounding names. |
7.2.179. SPACE
Returns a string consisting of "n" blanks. The output format is varchar(4000).
WITH temp1(n1) AS
(VALUES (1)
, (2)
, (3)
)
SELECT n1
, SPACE(n1) AS s1
, LENGTH(SPACE(n1)) AS s2
, SPACE(n1) || 'X' AS s3
FROM temp1;
ANSWER
N1 | S1 | S2 | S3 |
---|---|---|---|
1 |
1 |
"X" |
|
2 |
2 |
" X" |
|
3 |
3 |
" X" |
7.2.180. SQRT
Returns the square root of the input value, which can be any positive number. The output format is double.
WITH temp1(n1) AS
(VALUES (0.5)
, (0.0)
, (1.0)
, (2.0)
)
SELECT DEC(n1,4,3) AS n1
, DEC(SQRT(n1),4,3) AS s1
FROM temp1;
ANSWER
N1 | S1 |
---|---|
0.500 |
0.707 |
0.000 |
0.000 |
1.000 |
1.000 |
2.000 |
1.414 |
7.2.181. STRIP
Removes leading, trailing, or both (the default), characters from a string. If no strip character is provided, leading and/or trailing blank characters are removed.
Observe in the following query that the last example removes leading "A" characters:
WITH temp1(c1) AS
(VALUES (' ABC')
, (' ABC ')
, ('ABC ')
)
SELECT c1 AS c1
, STRIP(c1) AS c2
, LENGTH(STRIP(c1)) AS l2
, STRIP(c1,LEADING) AS c3
, LENGTH(STRIP(c1,LEADING)) AS l3
, STRIP(c1,LEADING,'A') AS C4
FROM temp1;
ANSWER
C1 | C2 | L2 | C3 | L3 | C4 |
---|---|---|---|---|---|
ABC |
ABC |
3 |
ABC |
3 |
ABC |
ABC |
ABC |
3 |
ABC |
4 |
ABC |
ABC |
ABC |
3 |
ABC |
5 |
BC |
The TRIM function works the same way.
7.2.182. STRLEFT
Returns the leftmost string of string-expression of length length, expressed in the specified string unit.
7.2.183. STRPOS
Is a synonym for POSSTR.
7.2.184. STRRIGHT
Returns the right most string of string-expression of length length, expressed in the specified string unit.
7.2.185. SUBSTR or SUBSTRING
Returns part of a string. If the length is not provided, the output is from the start value to the end of the string.
If the length is provided, and it is longer than the field length, a SQL error results. The following statement illustrates this. Note that in this example the DAT1 field has a "field length" of 9 (i.e. the length of the longest input string).
WITH temp1 (len, dat1) AS
(VALUES ( 6,'123456789')
, ( 4,'12345')
, ( 16,'123')
)
SELECT len
, dat1
, LENGTH(dat1) AS ldat
, SUBSTR(dat1, 1, len) AS subdat
FROM temp1;
ANSWER
LEN | DAT1 | LDAT | SUBDAT |
---|---|---|---|
6 |
123456789 |
9 |
123456 |
4 |
12345 |
5 |
1234 |
The best way to avoid the above problem is to simply write good code. If that sounds too much like hard work, try the following SQL:
WITH temp1 (len, dat1) AS
(VALUES ( 6,'123456789')
, ( 4,'12345')
, ( 16,'123'
)
SELECT len
, dat1
, LENGTH(dat1) AS ldat
, SUBSTR(dat1, 1,
CASE
WHEN len < LENGTH(dat1) THEN len
ELSE LENGTH(dat1)
END ) AS subdat
FROM temp1;
ANSWER
LEN | DAT1 | LDAT | SUBDAT |
---|---|---|---|
6 |
123456789 |
9 |
123456 |
4 |
12345 |
5 |
1234 |
16 |
123 |
3 |
123 |
In the above SQL a CASE statement is used to compare the LEN value against the length of the DAT1 field. If the former is larger, it is replaced by the length of the latter. If the input is varchar, and no length value is provided, the output is varchar. However, if the length is provided, the output is of type char - with padded blanks (if needed):
SELECT name
, LENGTH(name) AS len
, SUBSTR(name,5) AS s1
, LENGTH(SUBSTR(name,5)) AS l1
, SUBSTR(name,5,3) AS s2
, LENGTH(SUBSTR(name,5,3)) AS l2
FROM staff
WHERE id < 60;
ANSWER
NAME | LEN | S1 | L1 | S2 | L2 |
---|---|---|---|---|---|
Sanders |
7 |
ers |
3 |
ers |
3 |
Pernal |
6 |
al |
2 |
al |
3 |
Marenghi |
8 |
nghi |
4 |
ngh |
3 |
O’Brien |
7 |
ien |
3 |
ien |
3 |
Hanes |
5 |
s |
1 |
s |
3 |
7.2.186. SUBSTR2
Returns a substring from a string. The resulting substring starts at a specified position in the string and continues for a specified or default length. The start and length arguments are expressed in 16-bit UTF-16 string units (CODEUNITS16).
7.2.187. SUBSTR4
Returns a substring from a string. The resulting substring starts at a specified position in the string and continues for a specified or default length. The start and length arguments are expressed in 32-bit UTF-32 string units (CODEUNITS32).
7.2.188. SUBSTRB
Returns a substring of a string, beginning at a specified position in the string. Lengths are calculated in bytes.
7.2.189. TABLE
There isn’t really a TABLE function, but there is a TABLE phrase that returns a result, one row at a time, from either an external (e.g. user written) function, or from a nested table expression. The TABLE phrase (function) has to be used in the latter case whenever there is a reference in the nested table expression to a row that exists outside of the expression. An example follows:
SELECT a.id
, a.dept
, a.salary
, b.deptsal
FROM staff a
, TABLE
(SELECT b.dept
, SUM(b.salary) AS deptsal
FROM staff b
WHERE b.dept = a.dept
GROUP BY b.dept
) AS b
WHERE a.id < 40
ORDER BY a.id;
ANSWER
ID | DEPT | SALARY | DEPTSAL |
---|---|---|---|
10 |
20 |
98357.50 |
254286.10 |
20 |
20 |
78171.25 |
254286.10 |
30 |
38 |
77506.75 |
302285.55 |
See Table Function Usage for more details on using of the TABLE phrase in a nested table expression.
7.2.190. TABLE_NAME
Returns the base view or table name for a particular alias after all alias chains have been resolved. The output type is varchar(18). If the alias name is not found, the result is the input values. There are two input parameters. The first, which is required, is the alias name. The second, which is optional, is the alias schema. If the second parameter is not provided, the default schema is used for the qualifier.
TABLE_NAME function example
CREATE ALIAS emp1 FOR employee;
CREATE ALIAS emp2 FOR emp1;
SELECT tabschema
, tabname
, card
FROM syscat.tables
WHERE tabname = TABLE_NAME('emp2', 'graeme');
ANSWER
TABSCHEMA | TABNAME | CARD |
---|---|---|
graeme |
employee |
-1 |
7.2.191. TABLE_SCHEMA
Returns the base view or table schema for a particular alias after all alias chains have been resolved. The output type is char(8). If the alias name is not found, the result is the input values. There are two input parameters. The first, which is required, is the alias name. The second, which is optional, is the alias schema. If the second parameter is not provided, the default schema is used for the qualifier.
Resolving non-existent Objects
Dependent aliases are not dropped when a base table or view is removed. After the base table or view drop, the TABLE_SCHEMA and TABLE_NAME functions continue to work fine (see the 1st output line below). However, when the alias being checked does not exist, the original input values (explicit or implied) are returned (see the 2nd output line below).
CREATE VIEW fred1 (c1, c2, c3)
AS VALUES (11, 'AAA', 'BBB');
CREATE ALIAS fred2 FOR fred1;
CREATE ALIAS fred3 FOR fred2;
DROP VIEW fred1;
WITH temp1 (tab_sch, tab_nme) AS
(VALUES (TABLE_SCHEMA('fred3', 'graeme')
, TABLE_NAME('fred3'))
, (TABLE_SCHEMA('xxxxx')
, TABLE_NAME('xxxxx', 'xxx')))
SELECT *
FROM temp1;
ANSWER
TAB_SCH | TAB_NME |
---|---|
graeme |
fred1 |
graeme |
xxxxx |
7.2.192. TAN
Returns the tangent of the argument where the argument is an angle expressed in radians.
7.2.193. TANH
Returns the hyperbolic tan for the argument, where the argument is an angle expressed in radians. The output format is double.
7.2.194. THIS_MONTH
Returns the first day of the month in the specified date.
7.2.195. THIS_QUARTER
Returns the first day of the quarter that contains the specified date.
7.2.196. THIS_WEEK
Returns the first day of the week that contains the specified date. Sunday is considered the first day of that week.
7.2.197. THIS_YEAR
Returns the first day of the year in the specified date.
7.2.198. TIME
Converts the input into a time value.
7.2.199. TIMESTAMP
Converts the input(s) into a timestamp value.
Argument Options
If only one argument is provided, it must be (one of):
-
A timestamp value.
-
A character representation of a timestamp (the microseconds are optional).
-
A 14 byte string in the form: YYYYMMDDHHMMSS.
If both arguments are provided:
-
The first must be a date, or a character representation of a date.
-
The second must be a time, or a character representation of a time.
SELECT TIMESTAMP('1997-01-11-22.44.55.000000')
, TIMESTAMP('1997-01-11-22.44.55.000')
, TIMESTAMP('1997-01-11-22.44.55')
, TIMESTAMP('19970111224455')
, TIMESTAMP('1997-01-11','22.44.55')
FROM staff WHERE id = 10;
7.2.200. TIMESTAMP_FORMAT
Takes an input string with the format: "YYYY-MM-DD HH:MM:SS" and converts it into a valid timestamp value. The VARCHAR_FORMAT function does the inverse.
WITH temp1 (ts1) AS
(VALUES ('1999-12-31 23:59:59')
, ('2002-10-30 11:22:33')
)
SELECT ts1
, TIMESTAMP_FORMAT(ts1, 'YYYY-MM-DD HH24:MI:SS') AS ts2
FROM temp1
ORDER BY ts1;
ANSWER
TS1 | TS2 |
---|---|
1999-12-31 23:59:59 |
1999-12-31-23.59.59.000000 |
2002-10-30 11:22:33 |
2002-10-30-11.22.33.000000 |
Note that the only allowed formatting mask is the one shown.
7.2.201. TIMESTAMP_ISO
Returns a timestamp in the ISO format (yyyy-mm-dd hh:mm:ss.nnnnnn) converted from the IBM internal format (yyyy-mm-dd-hh.mm.ss.nnnnnn). If the input is a date, zeros are inserted in the time part. If the input is a time, the current date is inserted in the date part and zeros in the microsecond section.
SELECT tm1
, TIMESTAMP_ISO(tm1)
FROM scalar;
ANSWER
TM1 | 2 |
---|---|
23:58:58 |
2000-09-01-23.58.58.000000 |
15:15:15 |
2000-09-01-15.15.15.000000 |
00:00:00 |
2000-09-01-00.00.00.000000 |
7.2.202. TIMESTAMPDIFF
Returns an integer value that is an estimate of the difference between two timestamp values. Unfortunately, the estimate can sometimes be seriously out (see the example below), so this function should be used with extreme care.
Arguments
There are two arguments. The first argument indicates what interval kind is to be returned. Valid options are:
-
1 = Microseconds.
-
2 = Seconds.
-
4 = Minutes.
-
8 = Hours.
-
16 = Days.
-
32 = Weeks.
-
64 = Months.
-
128 = Quarters.
-
256 = Years.
The second argument is the result of one timestamp subtracted from another and then converted to character.
WITH
temp1 (ts1,ts2) AS
(VALUES ('1996-03-01-00.00.01','1995-03-01-00.00.00')
, ('1996-03-01-00.00.00','1995-03-01-00.00.01')
),
temp2 (ts1,ts2) AS
(SELECT TIMESTAMP(ts1)
, TIMESTAMP(ts2)
FROM temp1
),
temp3 (ts1,ts2,df) AS
(SELECT ts1
, ts2
, CHAR(TS1 - TS2) AS df
FROM temp2)
SELECT df
, TIMESTAMPDIFF(16,df) AS dif
, DAYS(ts1) - DAYS(ts2) AS dys
FROM temp3;
ANSWER
DF | DIF | DYS |
---|---|---|
00010000000001.000000 |
365 |
366 |
00001130235959.000000 |
360 |
366 |
Some the interval types return estimates, not definitive differences, so should be used with care. For example, to get the difference between two timestamps in days, use the DAYS function as shown above. It is always correct. |
Roll Your Own
The following user-defined function will get the difference, in microseconds, between two timestamp values. It can be used as an alternative to the above:
CREATE FUNCTION ts_diff_works(in_hi TIMESTAMP, in_lo TIMESTAMP)
RETURNS BIGINT
RETURN (BIGINT(DAYS(in_hi)) * 86400000000
+ BIGINT(MIDNIGHT_SECONDS(in_hi)) * 1000000
+ BIGINT(MICROSECOND(in_hi)))
-(BIGINT(DAYS(in_lo)) * 86400000000
+ BIGINT(MIDNIGHT_SECONDS(in_lo)) * 1000000
+ + BIGINT(MICROSECOND(in_lo)));
7.2.203. TO_CHAR
This function is a synonym for VARCHAR_FORMAT (see VARCHAR_FORMAT). It converts a timestamp value into a string using a template to define the output layout.
7.2.204. TO_CLOB
Returns a CLOB representation of a character string type.
7.2.205. TO_DATE
This function is a synonym for TIMESTAMP_FORMAT (see TIMESTAMP_FORMAT). It converts a character string value into a timestamp using a template to define the input layout.
7.2.206. TO_HEX
Converts a numeric expression into the hexadecimal representation.
7.2.207. TO_NCHAR
Returns a national character representation of an input expression that has been formatted using a character template.
7.2.208. TO_NCLOB
Returns any type of national character string.
7.2.209. TO_NUMBER
Returns a DECFLOAT(34) value that is based on the interpretation of the input string using the specified format.
7.2.210. TO_SINGLE_BYTE
Returns a string in which multi-byte characters are converted to the equivalent single-byte character where an equivalent character exists.
7.2.211. TO_TIMESTAMP
Returns a timestamp that is based on the interpretation of the input string using the specified format
7.2.212. TO_UTC_TIMESTAMP
Returns a TIMESTAMP that is converted to Coordinated Universal Time from the timezone that is specified by the timezone string. TO_UTC_TIMESTAMP is a a statement deterministic function.
7.2.213. TOTALORDER
Compares two DECFLOAT expressions and returns a SMALLINT number:
-
-1 if the first value is less than the second value.
-
0 if both values exactly equal (i.e. no trailing-zero differences)
-
+1 if the first value is greater than the second value.
Several values that compare as "less than" or "greater than" in the example below are equal in the usual sense. See the section on DECFLOAT arithmetic for details (see DECFLOAT Arithmetic).
,TOTALORDER function example
WITH temp1 (d1, d2) AS
(VALUES (DECFLOAT(+1.0), DECFLOAT(+1.0))
, (DECFLOAT(+1.0), DECFLOAT(+1.00))
, (DECFLOAT(-1.0), DECFLOAT(-1.00))
, (DECFLOAT(+0.0), DECFLOAT(+0.00))
, (DECFLOAT(-0.0), DECFLOAT(-0.00))
, (DECFLOAT(1234), +infinity)
, (+infinity, +infinity)
, (+infinity, -infinity)
, (DECFLOAT(1234), -NaN)
)
SELECT TOTALORDER(d1,d2)
FROM temp1;
ANSWER
1 |
---|
0 |
1 |
-1 |
1 |
1 |
-1 |
0 |
1 |
1 |
7.2.214. TRANSLATE
Converts individual characters in either a character or graphic input string from one value to another. It can also convert lower case data to upper case.
Usage Notes
-
The use of the input string alone generates upper case output.
-
When "from" and "to" values are provided, each individual "from" character in the input string is replaced by the corresponding "to" character (if there is one).
-
If there is no "to" character for a particular "from" character, those characters in the input string that match the "from" are set to blank (if there is no substitute value).
-
A fourth, optional, single-character parameter can be provided that is the substitute character to be used for those "from" values having no "to" value.
-
If there are more "to" characters than "from" characters, the additional "to" characters are ignored.
SELECT 'abcd'
, TRANSLATE('abcd')
, TRANSLATE('abcd', '', 'a')
, TRANSLATE('abcd', 'A', 'A')
, TRANSLATE('abcd', 'A', 'a')
, TRANSLATE('abcd', 'A', 'ab')
, TRANSLATE('abcd', 'A', 'ab', ' ')
, TRANSLATE('abcd', 'A', 'ab', 'z')
, TRANSLATE('abcd', 'AB', 'a')
FROM staff WHERE id = 10;
ANS. | NOTES |
---|---|
abcd |
No change |
ABCD |
Make upper case |
bcd |
'a'⇒' ' |
abcd |
'A'⇒'A' |
Abcd |
'a'⇒'A' |
A cd |
'a'⇒'A','b'⇒' ' |
A cd |
'a'⇒'A','b'⇒' ' |
Azcd |
'a'⇒'A','b'⇒'z' |
Abcd |
'a'⇒'A' |
REPLACE vs. TRANSLATE - A Comparison
Both the REPLACE and the TRANSLATE functions alter the contents of input strings. They differ in that the REPLACE converts whole strings while the TRANSLATE converts multiple sets of individual characters. Also, the "to" and "from" strings are back to front.
SELECT c1
, REPLACE(c1, 'AB', 'XY')
, REPLACE(c1, 'BA', 'XY')
, TRANSLATE(c1, 'XY', 'AB')
, TRANSLATE(c1, 'XY', 'BA')
FROM scalar
WHERE c1 = 'ABCD';
ANSWER
1 | 2 | 3 | 4 | 5 |
---|---|---|---|---|
ABCD |
XYCD |
ABCD |
XYCD |
YXCD |
7.2.215. TRIM
See STRIP function.
7.2.216. TRIM_ARRAY
Deletes elements from the end of an array.
7.2.217. TRUNC_TIMESTAMP
Returns a TIMESTAMP that is an argument (expression) truncated to the unit specified by another argument (format-string).
7.2.218. TRUNC or TRUNCATE
Truncates (not rounds) the rightmost digits of an input number (1st argument). If the second argument is positive, it truncates to the right of the decimal place. If the second value is negative, it truncates to the left. A second value of zero truncates to integer. The input and output types will equal. To round instead of truncate, use the ROUND function.
WITH temp1(d1) AS
(VALUES (123.400)
, ( 23.450)
, ( 3.456)
, ( .056)
)
SELECT d1
, DEC(TRUNC(d1,+2), 6, 3) AS pos2
, DEC(TRUNC(d1,+1), 6, 3) AS pos1
, DEC(TRUNC(d1,+0), 6, 3) AS zero
, DEC(TRUNC(d1,-1), 6, 3) AS neg1
, DEC(TRUNC(d1,-2), 6, 3) AS neg2
FROM temp1
ORDER BY 1 DESC;
ANSWER
D1 | POS2 | POS1 | ZERO | NEG1 | NEG2 |
---|---|---|---|---|---|
123.400 |
123.400 |
123.400 |
123.000 |
120.000 |
100.000 |
23.450 |
23.440 |
23.400 |
23.000 |
20.000 |
0.000 |
3.456 |
3.450 |
3.400 |
3.000 |
0.000 |
0.000 |
0.056 |
0.050 |
0.000 |
0.000 |
0.000 |
0.000 |
7.2.219. TYPE_ID
Returns the internal type identifier of the dynamic data type of the expression.
7.2.220. TYPE_NAME
Returns the unqualified name of the dynamic data type of the expression.
7.2.221. TYPE_SCHEMA
Returns the schema name of the dynamic data type of the expression.
7.2.222. UCASE or UPPER
Converts a mixed or lower-case string to upper case. The output is the same data type and length as the input.
SELECT name
, LCASE(name) AS lname
, UCASE(name) AS uname
FROM staff
WHERE id < 30;
ANSWER
NAME | LNAME | UNAME |
---|---|---|
Sanders |
sanders |
SANDERS |
Pernal |
pernal |
PERNAL |
7.2.223. VALUE
Same as COALESCE.
7.2.224. VARBINARY
Returns a VARBINARY (varying-length binary string) representation of a string of any data type.
7.2.225. VARCHAR
Converts the input (1st argument) to a varchar data type. The output length (2nd argument) is optional. Trailing blanks are not removed.
,VARCHAR function examples
SELECT c1
, LENGTH(c1) AS l1
, VARCHAR(c1) AS v2
, LENGTH(VARCHAR(c1)) AS l2
, VARCHAR(c1,4) AS v3
FROM scalar;
ANSWER
C1 | L1 | V2 | L2 | V3 |
---|---|---|---|---|
ABCDEF |
6 |
ABCDEF |
6 |
ABCD |
ABCD |
6 |
ABCD |
6 |
ABCD |
AB |
6 |
AB |
6 |
AB |
7.2.226. VARCHAR_BIT_FORMAT
Returns a VARCHAR bit-data representation of a character string. See the SQL Reference for more details.
7.2.227. VARCHAR_FORMAT
Returns a character representation of an input expression using the format specified by the format string.
WITH temp1 (ts1) AS
(VALUES (TIMESTAMP('1999-12-31-23.59.59'))
, (TIMESTAMP('2002-10-30-11.22.33'))
)
SELECT ts1
, VARCHAR_FORMAT(ts1,'YYYY-MM-DD HH24:MI:SS') AS vc1
FROM temp1
ORDER BY ts1;
ANSWER
TS1 | VC1 |
---|---|
1999-12-31-23.59.59.000000 |
1999-12-31 23:59:59 |
2002-10-30-11.22.33.000000 |
2002-10-30 11:22:33 |
WITH temp1 (sn1) AS
(VALUES (-1234567890)
, (1234567890)
)
SELECT sn1
, VARCHAR_FORMAT(sn1,'9,999,999,999') AS vc1
FROM temp1
ORDER BY sn1;
ANSWER
SN1 | VC1 |
---|---|
-1234567890 |
-1,234,567,890 |
1234567890 |
1,234,567,890 |
7.2.228. VARCHAR_FORMAT_BIT
Returns a VARCHAR representation of a character bit-data string. See the SQL Reference for more details.
7.2.229. VARGRAPHIC
Converts the input (1st argument) to a VARGRAPHIC data type. The output length (2nd argument) is optional.
7.2.230. VERIFY_GROUP_FOR_USER
Returns a value that indicates whether the groups associated with the authorization ID identified by the SESSION_USER special register are in the group names specified by the list of group-name-expression arguments.
7.2.231. VERIFY_ROLE_FOR_USER
Returns a value that indicates whether any of the roles associated with the authorization ID identified by the SESSION_USER special register are in (or contain any of) the role names specified by the list of role-name-expression arguments.
7.2.232. VERIFY_TRUSTED_CONTEXT_ROLE_FOR_USER
Returns a value that indicates whether the authorization ID identified by the SESSION_USER special register has acquired a role under a trusted connection associated with some trusted context and that role is in (or part of) the role names specified by the list of role-name-expression arguments.
7.2.233. WEEK
Returns a value in the range 1 to 53 or 54 that represents the week of the year, where a week begins on a Sunday, or on the first day of the year. Valid input types are a date, a timestamp, or an equivalent character value. The output is of type integer.
SELECT WEEK(DATE('2000-01-01')) AS w1
, WEEK(DATE('2000-01-02')) AS w2
, WEEK(DATE('2001-01-02')) AS w3
, WEEK(DATE('2000-12-31')) AS w4
, WEEK(DATE('2040-12-31')) AS w5
FROM sysibm.sysdummy1;
ANSWER
W1 | W2 | W3 | W4 | W5 |
---|---|---|---|---|
1 |
2 |
1 |
54 |
53 |
Both the first and last week of the year may be partial weeks. Likewise, from one year to the next, a particular day will often be in a different week (see Comparing Weeks).
7.2.234. WEEK_ISO
Returns an integer value, in the range 1 to 53, that is the "ISO" week number. An ISO week differs from an ordinary week in that it begins on a Monday and it neither ends nor begins at the exact end of the year. Instead, week 1 is the first week of the year to contain a Thursday. Therefore, it is possible for up to three days at the beginning of the year to appear in the last week of the previous year. As with ordinary weeks, not all ISO weeks contain seven days.
WEEK_ISO function example
WITH temp1 (n) AS
(VALUES (0)
UNION ALL
SELECT n+1
FROM temp1
WHERE n < 10
), temp2 (dt2) AS
(SELECT DATE('1998-12-27') + y.n YEARS + d.n DAYS
FROM temp1 y
, temp1 d
WHERE y.n IN (0,2)
)
SELECT CHAR(dt2,ISO) AS dte
, SUBSTR(DAYNAME(dt2),1,3) AS dy
, WEEK(dt2) AS wk
, DAYOFWEEK(dt2) AS dy
, WEEK_ISO(dt2) AS wi
, DAYOFWEEK_ISO(dt2) AS di
FROM temp2
ORDER BY 1;
ANSWER
DTE | DY | WK | DY | WI | DI |
---|---|---|---|---|---|
1998-12-27 |
Sun |
53 |
1 |
52 |
7 |
1998-12-28 |
Mon |
53 |
2 |
53 |
1 |
1998-12-29 |
Tue |
53 |
3 |
53 |
2 |
1998-12-30 |
Wed |
53 |
4 |
53 |
3 |
1998-12-31 |
Thu |
53 |
5 |
53 |
4 |
1999-01-01 |
Fri |
1 |
6 |
53 |
5 |
1999-01-02 |
Sat |
1 |
7 |
53 |
6 |
1999-01-03 |
Sun |
2 |
1 |
53 |
7 |
1999-01-04 |
Mon |
2 |
2 |
1 |
1 |
1999-01-05 |
Tue |
2 |
3 |
1 |
2 |
1999-01-06 |
Wed |
2 |
4 |
1 |
3 |
2000-12-27 |
Wed |
53 |
4 |
52 |
3 |
2000-12-28 |
Thu |
53 |
5 |
52 |
4 |
2000-12-29 |
Fri |
53 |
6 |
52 |
5 |
2000-12-30 |
Sat |
53 |
7 |
52 |
6 |
2000-12-31 |
Sun |
54 |
1 |
52 |
7 |
2001-01-01 |
Mon |
1 |
2 |
1 |
1 |
2001-01-02 |
Tue |
1 |
3 |
1 |
2 |
2001-01-03 |
Wed |
1 |
4 |
1 |
3 |
2001-01-04 |
Thu |
1 |
5 |
1 |
4 |
2001-01-05 |
Fri |
1 |
6 |
1 |
5 |
2001-01-06 |
Sat |
1 |
7 |
1 |
6 |
7.2.235. WEEKS_BETWEEN
Returns the number of full weeks between the specified arguments
7.2.236. WIDTH_BUCKET
Is used to create equal-width histograms.
7.2.237. XMLATTRIBUTES
Constructs XML attributes from the arguments.
7.2.238. XMLCOMMENT
Returns an XML value with a single XQuery comment node with the input argument as the content.
7.2.239. XMLCONCAT
Returns a sequence containing the concatenation of a variable number of XML input arguments
7.2.240. XMLDOCUMENT
Returns an XML value with a single XQuery document node with zero or more children nodes.
7.2.241. XMLELEMENT
Returns an XML value that is an XQuery element node.
7.2.242. XMLFOREST
Returns an XML value that is a sequence of XQuery element nodes.
7.2.243. XMLNAMESPACES
Constructs namespace declarations from the arguments.
7.2.244. XMLPARSE
Parses the argument as an XML document and returns an XML value.
7.2.245. XMLPI
Returns an XML value with a single XQuery processing instruction node.
7.2.246. XMLQUERY
Returns an XML value from the evaluation of an XQuery expression possibly using specified input arguments as XQuery variables.
7.2.247. XMLROW
Returns an XML value with a single XQuery document node containing one top-level element node.
7.2.248. XMLSERIALIZE
Returns a serialized XML value of the specified data type generated from the XML-expression argument.
7.2.249. XMLTEXT
Returns an XML value with a single XQuery text node having the input argument as the content.
7.2.250. XMLVALIDATE
Returns a copy of the input XML value augmented with information obtained from XML schema validation, including default values.
7.2.251. XMLXSROBJECTID
Returns the XSR object identifier of the XML schema used to validate the XML document specified in the argument. The XSR object identifier is returned as a BIGINT value and provides the key to a single row in SYSCAT.XSROBJECTS.
7.2.252. XSLTRANSFORM
Use XSLTRANSFORM to convert XML data into other formats, including the conversion of XML documents that conform to one XML schema into documents that conform to another schema.
7.2.253. YEAR
Returns a four-digit year value in the range 0001 to 9999 that represents the year (including the century). The input is a date or timestamp (or equivalent) value. The output is integer.
SELECT dt1
, YEAR(dt1) AS yr
, WEEK(dt1) AS wk
FROM scalar;
ANSWER
DT1 | YR | WK |
---|---|---|
1996-04-22 |
1996 |
17 |
1996-08-15 |
1996 |
33 |
0001-01-01 |
1 |
1 |
7.2.254. YEARS_BETWEEN
Returns the number of full years between the specified arguments.
7.2.255. YMD_BETWEEN
Returns a numeric value that specifies the number of full years, full months, and full days between two datetime values.
7.2.256. "+" PLUS
The PLUS function is same old plus sign that you have been using since you were a kid. One can use it the old fashioned way, or as if it were normal a Db2 function - with one or two input items. If there is a single input item, then the function acts as the unary "plus" operator. If there are two items, the function adds them:
SELECT id
, salary
, "+"(salary) AS s2
, "+"(salary, id) AS s3
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SALARY | S2 | S3 |
---|---|---|---|
10 |
98357.50 |
98357.50 |
98367.50 |
20 |
78171.25 |
78171.25 |
78191.25 |
30 |
77506.75 |
77506.75 |
77536.75 |
Both the PLUS and MINUS functions can be used to add and subtract numbers, and also date and time values. For the latter, one side of the equation has to be a date/time value, and the other either a date or time duration (a numeric representation of a date/time), or a specified date/time type. To illustrate, below are three different ways to add one year to a date:
SELECT empno
, CHAR(birthdate,ISO) AS bdate1
, CHAR(birthdate + 1 YEAR,ISO) AS bdate2
, CHAR("+"(birthdate,DEC(00010000,8)),ISO) AS bdate3
, CHAR("+"(birthdate,DOUBLE(1),SMALLINT(1)),ISO) AS bdate4
FROM employee
WHERE empno < '000040'
ORDER BY empno;
ANSWER
EMPNO | BDATE1 | BDATE2 | BDATE3 | BDATE4 |
---|---|---|---|---|
000010 |
1933-08-24 |
1934-08-24 |
1934-08-24 |
1934-08-24 |
000020 |
1948-02-02 |
1949-02-02 |
1949-02-02 |
1949-02-02 |
000030 |
1941-05-11 |
1942-05-11 |
1942-05-11 |
1942-05-11 |
7.2.257. "-" MINUS
The MINUS works the same way as the PLUS function, but does the opposite:
SELECT id
, salary
, "-"(salary) AS s2
, "-"(salary, id) AS s3
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SALARY | S2 | S3 |
---|---|---|---|
10 |
98357.50 |
-98357.50 |
98347.50 |
20 |
78171.25 |
-78171.25 |
78151.25 |
30 |
77506.75 |
-77506.75 |
77476.75 |
7.2.258. "*" MULTIPLY
The MULTIPLY function is used to multiply two numeric values:
SELECT id
, salary
, salary * id AS s2
, "*"(salary, id) AS s3
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SALARY | S2 | S3 |
---|---|---|---|
10 |
98357.50 |
983575.00 |
983575.00 |
20 |
78171.25 |
1563425.00 |
1563425.00 |
30 |
77506.75 |
2325202.50 |
2325202.50 |
7.2.259. "/" DIVIDE
The DIVIDE function is used to divide two numeric values:
SELECT id
, salary
, salary / id AS s2
, "/"(salary, id) AS s3
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SALARY | S2 | S3 |
---|---|---|---|
10 |
98357.50 |
9835.75 |
9835.75 |
20 |
78171.25 |
3908.56 |
3908.56 |
30 |
77506.75 |
2583.55 |
2583.55 |
7.2.260. "||" CONCAT
Same as the CONCAT function:
SELECT id
, name || 'Z' AS n1
, name CONCAT 'Z' AS n2
, "||"(name,'Z') AS n3
, CONCAT(name,'Z') AS n4
FROM staff
WHERE LENGTH(name) < 5
ORDER BY id;
ANSWER
ID | N1 | N2 | N3 | N4 |
---|---|---|---|---|
110 |
NganZ |
NganZ |
NganZ |
NganZ |
210 |
LuZ |
LuZ |
LuZ |
LuZ |
270 |
LeaZ |
LeaZ |
LeaZ |
LeaZ |
7.3. User Defined Functions
Many problems that are really hard to solve using raw SQL become surprisingly easy to address, once one writes a simple function. This chapter will cover some of the basics of user defined functions. These can be very roughly categorized by their input source, their output type, and the language used:
-
External scalar functions use an external process (e.g. a C program), and possibly also an external data source, to return a single value.
-
External table functions use an external process, and possibly also an external data source, to return a set of rows and columns.
-
Internal sourced functions are variations of an existing Db2 function
-
Internal scalar functions use compound SQL code to return a single value.
-
Internal table functions use compound SQL code to return a set of rows and columns.
This chapter will briefly go over the last three types of function listed above. See the official Db2 documentation for more details.
As of the time of writing, there is a known bug in Db2 that causes the prepare cost of a dynamic SQL statement to go up exponentially when a user defined function that is written in the SQL language is referred to multiple times in a single SQL statement. |
7.3.1. Sourced Functions
A sourced function is used to redefine an existing Db2 function so as to in some way restrict or enhance its applicability. Below is a scalar function that is a variation on the standard DIGITS function, but which only works on small integer fields:
CREATE FUNCTION digi_int (SMALLINT)
RETURNS CHAR(5)
SOURCE SYSIBM.DIGITS(SMALLINT);
Here is an example of the function in use:
SELECT id AS ID
, DIGITS(id) AS I2
, digi_int(id) AS I3
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | I2 | I3 |
---|---|---|
10 |
00010 |
00010 |
20 |
00020 |
00020 |
30 |
00030 |
00030 |
By contrast, the following statement will fail because the input is an integer field:
SELECT id
, digi_int(INT(id))
FROM staff
WHERE id < 50;
ANSWER: Error
Sourced functions are especially useful when one has created a distinct (data) type, because these do not come with any of the usual Db2 functions. To illustrate, in the following example a distinct type is created, then a table using the type, then two rows are inserted:
Create distinct type and test table
CREATE DISTINCT TYPE us_dollars AS DEC(7,2) WITH COMPARISONS;
CREATE TABLE customers
( ID SMALLINT NOT NULL
, balance us_dollars NOT NULL);
INSERT INTO customers VALUES (1, 111.11), (2, 222.22);
SELECT *
FROM customers
ORDER BY ID;
ANSWER
ID | balance |
---|---|
1 |
111.11 |
2 |
222.22 |
The next query will fail because there is currently no multiply function for "us_dollars":
Do multiply - fails
SELECT id
, balance * 10
FROM customers
ORDER BY id;
ANSWER: Error
The enable the above, we have to create a sourced function:
CREATE FUNCTION "*" (us_dollars, INT)
RETURNS us_dollars
SOURCE SYSIBM."*"(DECIMAL, INT);
Now we can do the multiply:
SELECT id
, balance * 10 AS newbal
FROM customers
ORDER BY id;
ANSWER
ID | NEWBAL |
---|---|
1 |
1111.10 |
2 |
2222.20 |
For the record, here is another way to write the same:
SELECT id
, "*"(balance,10) AS newbal
FROM customers
ORDER BY id;
ANSWER
ID | NEWBAL |
---|---|
1 |
1111.10 |
2 |
2222.20 |
7.3.2. Scalar Functions
A scalar function has as input a specific number of values (i.e. not a table) and returns a single output item.
If a function returns a value (as opposed to a table), that value will always be nullable, regardless of whether or not the returned value can ever actually be null. This may cause problems if one is not prepared to handle a null indicator. To illustrate, the following function will return a nullable value that never be null:
CREATE FUNCTION Test()
RETURNS CHAR(5)
RETURN 'abcde';
Input and Output Limits
One can have multiple scalar functions with the same name and different input/output data types, but not with the same name and input/output types, but with different lengths. So if one wants to support all possible input/output lengths for, say, varchar data, one has to define the input and output lengths to be the maximum allowed for the field type. For varchar input, one would need an output length of 32,672 bytes to support all possible input values. But this is a problem, because it is very close to the maximum allowable table (row) length in Db2, which is 32,677 bytes. Decimal field types are even more problematic, because one needs to define both a length and a scale. To illustrate, imagine that one defines the input as being of type decimal(31,12). The following input values would be treated thus:
-
A decimal(10,5) value would be fine.
-
A decimal(31,31) value would lose precision.
-
A decimal(31,0) value may fail because it is too large.
-
See Convert Number to Character for a detailed description of this problem.
Examples
Below is a very simple scalar function - that always returns zero:
CREATE FUNCTION returns_zero() RETURNS SMALLINT RETURN 0;
SELECT id AS id
, returns_zero() AS zz
FROM staff
WHERE id = 10;
ANSWER
ID | ZZ |
---|---|
10 |
0 |
Two functions can be created with the same name. Which one is used depends on the input type that is provided:
CREATE FUNCTION calc(inval SMALLINT) RETURNS INT RETURN inval * 10;
CREATE FUNCTION calc(inval INTEGER) RETURNS INT RETURN inval * 5;
SELECT id AS id
, calc(SMALLINT(id)) AS c1
, calc(INTEGER (id)) AS C2
FROM staff
WHERE id < 30
ORDER BY id;
ANSWER
ID | C1 | C2 |
---|---|---|
10 |
100 |
50 |
20 |
200 |
100 |
DROP FUNCTION calc(SMALLINT); DROP FUNCTION calc(INTEGER);
Below is an example of a function that is not deterministic, which means that the function result can not be determined based on the input:
CREATE FUNCTION rnd(inval INT)
RETURNS SMALLINT
NOT DETERMINISTIC
RETURN RAND() * 50;
SELECT id AS id
, rnd(1) AS RND
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | RND |
---|---|
10 |
37 |
20 |
8 |
30 |
42 |
The next function uses a query to return a single row/column value:
CREATE FUNCTION get_sal(inval SMALLINT)
RETURNS DECIMAL(7, 2)
RETURN SELECT salary
FROM staff
WHERE id = inval;
SELECT id AS id
, get_sal(id) AS salary
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SALARY |
---|---|
10 |
98357.50 |
20 |
78171.25 |
30 |
77506.75 |
More complex SQL statements are also allowed - as long as the result (in a scalar function) is just one row/column value. In the next example, the either the maximum salary in the same department is obtained, or the maximum salary for the same year - whatever is higher:
CREATE FUNCTION max_sal(inval SMALLINT)
RETURNS DECIMAL(7, 2)
RETURN
WITH ddd (max_sal) AS
(SELECT MAX(S2.salary)
FROM staff S1
, staff S2
WHERE S1.id = inval
AND S1.dept = s2.dept
)
, yyy (max_sal) AS
(SELECT MAX(S2.salary)
FROM staff S1
, staff S2
WHERE S1.id = inval
AND S1.years = s2.years
)
SELECT
CASE
WHEN ddd.max_sal > yyy.max_sal THEN ddd.max_sal
ELSE yyy.max_sal
END
FROM ddd, yyy;
SELECT id AS id
, salary AS SAL1
, max_sal(id) AS SAL2
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SAL1 | SAL2 |
---|---|---|
10 |
98357.50 |
98357.50 |
20 |
78171.25 |
98357.50 |
30 |
77506.75 |
79260.25 |
A scalar or table function cannot change any data, but it can be used in a DML statement. In the next example, a function is used to remove all "e" characters from the name column:
CREATE FUNCTION remove_e(instr VARCHAR(50))
RETURNS VARCHAR(50)
RETURN replace(instr,'e','');
UPDATE staff
SET name = remove_e(name)
WHERE id < 40;
7.3.3. Compound SQL Usage
A function can use compound SQL, with the following limitations:
-
The statement delimiter, if needed, cannot be a semi-colon.
-
No DML statements are allowed.
Below is an example of a scalar function that uses compound SQL to reverse the contents of a text string:
--#SET DELIMITER !
CREATE FUNCTION reverse(instr VARCHAR(50))
RETURNS VARCHAR(50)
BEGIN ATOMIC
DECLARE outstr VARCHAR(50) DEFAULT '';
DECLARE curbyte SMALLINT
DEFAULT 0;
SET curbyte = LENGTH(RTRIM(instr));
WHILE curbyte >= 1 DO
SET outstr = outstr || SUBSTR(instr,curbyte,1);
SET curbyte = curbyte - 1;
END WHILE;
RETURN outstr;
END!
SELECT id AS id
, name AS name1
, reverse(name) AS name2
FROM staff
WHERE id < 40
ORDER BY id!
ANSWER
ID | NAME1 | NAME2 |
---|---|---|
10 |
Sanders |
srednaS |
20 |
Pernal |
lanreP |
30 |
Marenghi |
ihgneraM |
This example uses an "!" as the stmt delimiter. |
Because compound SQL is a language with basic logical constructs, one can add code that does different things, depending on what input is provided. To illustrate, in the next example the possible output values are as follows:
-
If the input is null, the output is set to null.
-
If the length of the input string is less than 6, an error is flagged.
-
If the length of the input string is less than 7, the result is set to -1.
-
Otherwise, the result is the length of the input string.
Now for the code:
--#SET DELIMITER !
CREATE FUNCTION check_len(instr VARCHAR(50))
RETURNS SMALLINT
BEGIN ATOMIC
IF instr IS NULL THEN
RETURN NULL;
END IF;
IF length(instr) < 6 THEN
SIGNAL SQLSTATE '75001'
SET MESSAGE_TEXT = 'Input string is < 6';
ELSEIF length(instr) < 7 THEN
RETURN -1;
END IF;
RETURN length(instr);
END!
SELECT id AS id
, name AS name1
, check_len(name) AS name2
FROM staff
WHERE id < 60
ORDER BY id!
This example uses an "!" as the stmt delimiter. |
ANSWER
ID | NAME1 | NAME2 |
---|---|---|
10 |
Sanders |
7 |
20 |
Pernal |
-1 |
30 |
Marenghi |
8 |
40 |
O’Brien |
7 |
The above query failed when it got to the name "Hanes", which is less than six bytes long.
8. Table Functions
A table function is very similar to a scalar function, except that it returns a set of rows and columns, rather than a single value. Here is an example:
CREATE FUNCTION get_staff()
RETURNS TABLE ( ID SMALLINT
, name VARCHAR(9)
, YR SMALLINT)
RETURN
SELECT id
, name
, years
FROM staff;
SELECT *
FROM TABLE(get_staff()) AS s
WHERE id < 40
ORDER BY id;
ANSWER
ID | NAME | YR |
---|---|---|
10 |
Sanders |
7 |
20 |
Pernal |
8 |
30 |
Marenghi |
5 |
Note the following:
-
The TABLE keyword, the function name (obviously), the two sets of parenthesis , and a correlation name, are all required.
-
If the function has input parameters, they are all required, and their type must match.
-
Optionally, one can list all of the columns that are returned by the function, giving each an assigned name.
Below is an example of a function that uses all of the above features:
CREATE FUNCTION get_st(inval INTEGER)
RETURNS
TABLE ( id SMALLINT
, name VARCHAR(9)
, yr SMALLINT)
RETURN
SELECT id
, name
, years
FROM staff
WHERE id = inval;
SELECT *
FROM TABLE(get_st(30)) AS sss (id, nnn, yy) *
ANSWER
ID | NNN | YY |
---|---|---|
30 |
Marenghi |
5 |
Examples
A table function returns a table, but it doesn’t have to touch a table. To illustrate, the following function creates the data on the fly:
CREATE FUNCTION make_data()
RETURNS
TABLE ( KY SMALLINT
, DAT CHAR(5))
RETURN
WITH temp1 (k#) AS
(VALUES (1),(2),(3))
SELECT k#
, DIGITS(SMALLINT(k#))
FROM temp1;
SELECT *
FROM TABLE(make_data()) AS ttt;
ANSWER
KY | DAT |
---|---|
1 |
00001 |
2 |
00002 |
3 |
00003 |
The next example uses compound SQL to first flag an error if one of the input values is too low, then find the maximum salary and related ID in the matching set of rows, then fetch the same rows - returning the two previously found values at the same time:
This example uses an "!" as the stmt delimiter. |
CREATE FUNCTION staff_list(lo_key INTEGER
, lo_sal INTEGER)
RETURNS
TABLE ( id SMALLINT
, salary DECIMAL(7,2)
, max_sal DECIMAL(7,2)
, id_max SMALLINT)
LANGUAGE SQL
READS SQL DATA
EXTERNAL ACTION
DETERMINISTIC
BEGIN ATOMIC
DECLARE hold_sal DECIMAL(7,2) DEFAULT 0;
DECLARE hold_key SMALLINT;
IF lo_sal < 0 THEN
SIGNAL SQLSTATE '75001'
SET MESSAGE_TEXT = 'Salary too low';
END IF;
FOR get_max AS
SELECT id AS in_key
, salary As in_sal
FROM staff
WHERE id >= lo_key
DO
IF in_sal > hold_sal THEN
SET hold_sal = in_sal;
SET hold_key = in_key;
END IF;
END FOR;
RETURN
SELECT id
, salary
, hold_sal
, hold_key
FROM staff
WHERE id >= lo_key;
END!
SELECT *
FROM TABLE(staff_list(66,1)) AS ttt
WHERE id < 111
ORDER BY id!
ANSWER
ID | SALARY | MAX_SAL | ID_MAX |
---|---|---|---|
70 |
76502.83 |
91150.00 |
140 |
80 |
43504.60 |
91150.00 |
140 |
90 |
38001.75 |
91150.00 |
140 |
100 |
78352.80 |
91150.00 |
140 |
110 |
42508.20 |
91150.00 |
140 |
8.1. BASE_TABLE
Returns both the object name and schema name of the object found after any alias chains have been resolved.
8.2. UNNEST
Returns a result table that includes a row for each element of the specified array. If there are multiple ordinary array arguments specified, the number of rows will match the array with the largest cardinality.
8.3. XMLTABLE
Returns a result table from the evaluation of XQuery expressions, possibly using specified input arguments as XQuery variables. Each sequence item in the result sequence of the row XQuery expression represents a row of the result table.
9. Useful User-Defined Functions
In this section we will describe some simple functions that are generally useful, and that people have asked for over the years. In addition to the functions listed here, there are also the following elsewhere in this book:
-
Check character input is a numeric value - Check Numeric function
-
Convert numeric data to character (right justified) - User-defined functions - convert integer to character.
-
Like-column predicate evaluation - LIKE_COLUMN Function.
-
Locate string in input, a block at a time - LOCATE_BLOCK user defined function.
-
Pause SQL statement (by looping) for "n" seconds - Function to Pause for "n" Seconds.
9.1. Julian Date Functions
The function below converts a Db2 date into a Julian date (format) value:
CREATE FUNCTION julian_out(inval DATE)
RETURNS
CHAR(7)
RETURN
RTRIM(CHAR(YEAR(inval)))
||
SUBSTR(DIGITS(DAYOFYEAR(inval)),8);
SELECT empno
, CHAR(hiredate, ISO) AS h_date
, JULIAN_OUT(hiredate) AS j_date
FROM employee
WHERE empno < '000050'
ORDER BY empno;
ANSWER
EMPNO | H_DATE | J_DATE |
---|---|---|
000010 |
1995-01-01 |
1995001 |
000020 |
2003-10-10 |
2003283 |
000030 |
2005-04-05 |
2005095 |
The next function does the opposite:
CREATE FUNCTION julian_in(inval CHAR(7))
RETURNS
DATE
RETURN
DATE('0001-01-01')
+ (INT(SUBSTR(inval,1,4)) - 1) YEARS
+ (INT(SUBSTR(inval,5,3)) - 1) DAYS;
9.2. Get Prior Date
Imagine that one wanted to get all rows where some date is for the prior year - relative to the current year. This is easy to code:
SELECT empno
, hiredate
FROM employee
WHERE YEAR(hiredate) = YEAR(CURRENT DATE) - 1;
9.3. Get Prior Month
One can use the DAYS function to get the same data for the prior day. But one cannot use the MONTH function to do the equivalent for the prior month because at the first of the year the month number goes back to one.
One can address this issue by writing a simple function that multiplies the year-number by 12, and then adds the month-number:
CREATE FUNCTION year_month(inval DATE)
RETURNS
INTEGER
RETURN
(YEAR(inval) * 12) + MONTH(inval);
We can use this function thus:
SELECT empno
, hiredate
FROM employee
WHERE YEAR_MONTH(hiredate) = YEAR_MONTH(CURRENT DATE) - 1;
9.4. Get Prior Week
Selecting rows for the prior week is complicated by the fact that both the US and ISO definitions of a week begin at one at the start of the year (see Comparing Weeks). If however we choose to define a week as a set of seven contiguous days, regardless of the date, we can create a function to do the job. In the example below we shall assume that a week begins on a Sunday:
CREATE FUNCTION sunday_week(inval DATE)
RETURNS
INTEGER
RETURN
DAYS(inval) / 7;
The next function assumes that a week begins on a Monday:
CREATE FUNCTION monday_week(inval DATE)
RETURNS
INTEGER
RETURN
(DAYS(inval) - 1) / 7;
Both the above functions convert the input date into a day-number value, then subtract (if needed) to get to the right day of the week, then divide by seven to get a week-number. The result is the number of weeks since the beginning of the current era. The next query shows the two functions in action:
WITH temp1 (num,dt) AS
(VALUES (1, DATE('2004-12-29'))
UNION ALL
SELECT num + 1
, dt + 1 DAY
FROM temp1
WHERE num < 15
),
temp2 (dt, dy) AS
(SELECT dt
, SUBSTR(DAYNAME(dt),1,3)
FROM temp1
)
SELECT CHAR(dt, ISO) AS date
, dy AS day
, WEEK(dt) AS wk
, WEEK_ISO(dt) AS is
, sunday_week(dt) AS sun_wk
, monday_week(dt) AS mon_wk
FROM temp2
ORDER BY 1;
ANSWER
DATE | DAY | WK | IS | SUN_WK | MON_WK |
---|---|---|---|---|---|
2004-12-29 |
Wed |
53 |
53 |
104563 |
104563 |
2004-12-30 |
Thu |
53 |
53 |
104563 |
104563 |
2004-12-31 |
Fri |
53 |
53 |
104563 |
104563 |
2005-01-01 |
Sat |
1 |
53 |
104563 |
104563 |
2005-01-02 |
Sun |
2 |
53 |
104564 |
104563 |
2005-01-03 |
Mon |
2 |
1 |
104564 |
104564 |
2005-01-04 |
Tue |
2 |
1 |
104564 |
104564 |
2005-01-05 |
Wed |
2 |
1 |
104564 |
104564 |
2005-01-06 |
Thu |
2 |
1 |
104564 |
104564 |
2005-01-07 |
Fri |
2 |
1 |
104564 |
104564 |
2005-01-08 |
Sat |
2 |
1 |
104564 |
104564 |
2005-01-09 |
Sun |
3 |
1 |
104565 |
104564 |
2005-01-10 |
Mon |
3 |
2 |
104565 |
104565 |
2005-01-11 |
Tue |
3 |
2 |
104565 |
104565 |
2005-01-12 |
Wed |
3 |
2 |
104565 |
104565 |
9.4.1. Generating Numbers
The next function returns a table of rows. Each row consists of a single integer value , starting at zero, and going up to the number given in the input. At least one row is always returned. If the input value is greater than zero, the number of rows returned equals the input value plus one:
CREATE FUNCTION NumList(max_num INTEGER)
RETURNS
TABLE(num INTEGER)
LANGUAGE SQL
RETURN
WITH temp1 (num) AS
(VALUES (0)
UNION ALL
SELECT num + 1
FROM temp1
WHERE num < max_num
)
SELECT num
FROM temp1;
Below are some queries that use the above function:
SELECT *
FROM TABLE(NumList(-1)) AS xxx;
ANSWER
1 |
---|
0 |
SELECT *
FROM TABLE(NumList(+0)) AS xxx;
1 |
---|
0 |
SELECT *
FROM TABLE(NumList(+3)) AS xxx;
1 |
---|
0 |
1 |
2 |
3 |
SELECT *
FROM TABLE(NumList(CAST(NULL AS INTEGER))) AS xxx;
1 |
---|
0 |
If this function did not always return one row, we might have to use a left-outer-join when joining to it. Otherwise the calling row might disappear from the answer-set because no row was returned. |
To illustrate the function’s usefulness, consider the following query, which returns the start and end date for a given set of activities:
SELECT actno
, emstdate
, emendate
, DAYS(emendate) - DAYS(emstdate) AS #days
FROM emp_act act
WHERE empno = '000260'
AND projno = 'AD3113'
AND actno < 100
AND emptime = 0.5
ORDER BY actno;
ANSWER
ACTNO | EMSTDATE | EMENDATE | #DAYS |
---|---|---|---|
70 |
2002-06-15 |
2002-07-01 |
16 |
80 |
2002-03-01 |
2002-04-15 |
45 |
Imagine that we wanted take the above output, and generate a row for each day between the start and end dates. To do this we first have to calculate the number of days between a given start and end, and then join to the function using that value:
SELECT actno
, #days
, num
, emstdate + num DAYS AS new_date
FROM
(SELECT actno
, emstdate
, emendate
, DAYS(emendate) - DAYS(emstdate) AS #days
FROM emp_act act
WHERE empno = '000260'
AND projno = 'AD3113'
AND actno < 100
AND emptime = 0.5
) AS aaa
, TABLE(NumList(#days)) AS ttt
ORDER BY actno
, num;
ANSWER
ACTNO | #DAYS | NUM | NEW_DATE |
---|---|---|---|
70 |
16 |
0 |
2002-06-15 |
70 |
16 |
1 |
2002-06-16 |
70 |
16 |
2 |
2002-06-17 |
70 |
16 |
3 |
2002-06-18 |
70 |
16 |
4 |
2002-06-19 |
70 |
16 |
5 |
2002-06-20 |
70 |
16 |
6 |
2002-06-21 |
70 |
16 |
7 |
2002-06-22 |
70 |
16 |
8 |
2002-06-23 |
70 |
16 |
9 |
2002-06-24 |
70 |
16 |
10 |
2002-06-25 |
etc… |
In the above query the #days value equals the number of days between the start and end dates. If the two dates equal, the #days value will be zero. In this case we will still get a row because the function will return a single zero value. If this were not the case (i.e. the function returned no rows if the input value was less than one), we would have to code a left-outer-join with a fake ON statement:
SELECT actno
, #days
, num
, emstdate + num DAYS AS new_date
FROM
(SELECT actno
, emstdate
, emendate
, DAYS(emendate)- DAYS(emstdate) AS #days
FROM emp_act act
WHERE empno = '000260'
AND projno = 'AD3113'
AND actno < 100
AND emptime = 0.5
) AS aaa
LEFT OUTER JOIN
TABLE(NumList(#days)) AS ttt
ON 1 = 1
ORDER BY actno
, num;
ACTNO | #DAYS | NUM | NEW_DATE |
---|---|---|---|
70 |
16 |
0 |
2002-06-15 |
70 |
16 |
1 |
2002-06-16 |
70 |
16 |
2 |
2002-06-17 |
70 |
16 |
3 |
2002-06-18 |
70 |
16 |
4 |
2002-06-19 |
70 |
16 |
5 |
2002-06-20 |
70 |
16 |
6 |
2002-06-21 |
70 |
16 |
7 |
2002-06-22 |
70 |
16 |
8 |
2002-06-23 |
70 |
16 |
9 |
2002-06-24 |
70 |
16 |
10 |
2002-06-25 |
etc… |
9.5. Check Data Value Type
The following function checks to see if an input value is character, where character is defined as meaning that all bytes are "A" through "Z" or blank. It converts (if possible) all bytes to blank using the TRANSLATE function, and then checks to see if the result is blank:
CREATE FUNCTION ISCHAR (inval VARCHAR(250))
RETURNS
SMALLINT
LANGUAGE SQL
RETURN
CASE
WHEN TRANSLATE(UPPER(inval),' ','ABCDEFGHIJKLMNOPQRSTUVWXYZ') = ' ' THEN 1
ELSE 0
END;
The next function is similar to the prior, except that it looks to see if all bytes in the input are in the range of "0" through "9", or blank:
CREATE FUNCTION ISNUM (inval VARCHAR(250))
RETURNS
SMALLINT
LANGUAGE SQL
RETURN
CASE
WHEN TRANSLATE(inval,' ','01234567890') = ' ' THEN 1
ELSE 0
END;
Below is an example of the above two functions in action:
WITH temp (indata) AS
(VALUES ('ABC')
, ('123')
, ('3.4')
, ('-44')
, ('A1 ')
, (' ')
)
SELECT indata AS indata
, ISCHAR(indata) AS c
, ISNUM(indata) AS n
FROM temp;
ANSWER
INDATA | C | N |
---|---|---|
ABC |
1 |
0 |
123 |
0 |
1 |
3.4 |
0 |
0 |
-44 |
0 |
0 |
A1 |
0 |
0 |
1 |
1 |
The above ISNUM function is a little simplistic. It doesn’t check for all-blanks, or embedded blanks, decimal input, or sign indicators. The next function does all of this, and also indicates what type of number was found:
CREATE FUNCTION ISNUM2 (inval VARCHAR(255))
RETURNS
CHAR(4)
LANGUAGE SQL
RETURN
CASE
WHEN inval = ' ' THEN ' '
WHEN LOCATE(' ',RTRIM(LTRIM(inval))) > 0 THEN ' '
WHEN TRANSLATE(inval,' ','01234567890') = inval THEN ' '
WHEN TRANSLATE(inval,' ','01234567890') = ' ' THEN 'INT '
WHEN TRANSLATE(inval,' ','+01234567890') = ' '
AND LOCATE('+',LTRIM(inval)) = 1
AND LENGTH(REPLACE(inval,'+','')) = LENGTH(inval) - 1 THEN 'INT+'
WHEN TRANSLATE(inval,' ','-01234567890') = ' '
AND LOCATE('-',LTRIM(inval)) = 1
AND LENGTH(REPLACE(inval,'-','')) = LENGTH(inval) - 1 THEN 'INT-'
WHEN TRANSLATE(inval,' ','.01234567890') = ' '
AND LENGTH(REPLACE(inval,'.','')) = LENGTH(inval) - 1 THEN 'DEC '
WHEN TRANSLATE(inval,' ','+.01234567890') = ' '
AND LOCATE('+',LTRIM(inval)) = 1
AND LENGTH(REPLACE(inval,'+','')) = LENGTH(inval) - 1
AND LENGTH(REPLACE(inval,'.','')) = LENGTH(inval) - 1 THEN 'DEC+'
WHEN TRANSLATE(inval,' ','-.01234567890') = ' '
AND LOCATE('-',LTRIM(inval)) = 1
AND LENGTH(REPLACE(inval,'-','')) = LENGTH(inval) - 1
AND LENGTH(REPLACE(inval,'.','')) = LENGTH(inval) - 1 THEN 'DEC-'
ELSE ' '
END;
The first three WHEN checks above are looking for non-numeric input:
-
The input is blank.
-
The input has embedded blanks.
-
The input does not contain any digits.
The final five WHEN checks look for a specific types of numeric input. They are all similar in design, so we can use the last one (looking of negative decimal input) to illustrate how they all work:
-
Check that the input consists only of digits, dots, the minus sign, and blanks.
-
Check that the minus sign is the left-most non-blank character.
-
Check that there is only one minus sign in the input.
-
Check that there is only one dot in the input.
Below is an example of the above function in use:
WITH temp (indata) AS
(VALUES ('ABC')
, ('123')
, ('3.4')
, ('-44')
, ('+11')
, ('-1-')
, ('12+')
, ('+.1')
, ('-0.')
, (' ')
, ('1 1')
, (' . ')
)
SELECT indata AS indata
, ISNUM2(indata) AS type
, CASE
WHEN ISNUM2(indata) <> '' THEN DEC(indata,5,2)
ELSE NULL
END AS number
FROM temp;
ANSWER
INDATA | TYPE | NUMBER |
---|---|---|
ABC |
- |
|
123 |
INT |
123.00 |
3.4 |
DEC |
|
3.40 |
-44 |
INT |
-44.00 |
+11 |
INT+ |
11.00 |
-1- |
|
- |
12+ |
|
- |
+.1 |
DEC+ |
0.10 |
-0. |
DEC |
0.00 |
||
- |
1 1 |
|
- |
. |
9.6. Hash Function
The following hash function is a little crude, but it works. It accepts a VARCHAR string as input, then walks the string and, one byte at a time, manipulates a floating point number. At the end of the process the floating point value is translated into BIGINT.
This example uses an "!" as the stmt delimiter. |
CREATE FUNCTION HASH_STRING (instr VARCHAR(30000))
RETURNS
BIGINT
DETERMINISTIC
CONTAINS SQL
NO EXTERNAL ACTION
BEGIN ATOMIC
DECLARE inlen SMALLINT;
DECLARE curbit SMALLINT DEFAULT 1;
DECLARE outnum DOUBLE DEFAULT 0;
SET inlen = LENGTH(instr);
WHILE curbit <= inlen
DO
SET outnum = (outnum * 123) + ASCII(SUBSTR(instr,curbit));
IF outnum > 1E10 THEN
SET outnum = outnum / 1.2345E6;
END IF;
SET curbit = curbit + 1;
END WHILE;
RETURN BIGINT(TRANSLATE(CHAR(outnum),'01','.E'));
END!
Below is an example of the function in use:
SELECT id
, name
, HASH_STRING(name) AS hash_val
FROM staff s
WHERE id < 70
ORDER BY id!
ANSWER
ID | NAME | HASH_VAL |
---|---|---|
10 |
Sanders |
203506538768383718 |
20 |
Pernal |
108434258721263716 |
30 |
Marenghi |
201743899927085914 |
40 |
O’Brien |
202251277018590318 |
50 |
Hanes |
103496977706763914 |
60 |
Quigley |
202990889019520318 |
One way to judge a hash function is to look at the number of distinct values generated for a given number of input strings. Below is a very simple test:
WITH
temp1 (col1) AS
(VALUES (1)
UNION ALL
SELECT col1 + 1
FROM temp1
WHERE col1 < 100000
)
SELECT COUNT(*) AS #rows
, COUNT(DISTINCT HASH_STRING(CHAR(col1))) AS #hash1
, COUNT(DISTINCT HASH_STRING(DIGITS(col1))) AS #hash2
FROM temp1!
ANSWER
#ROWS | #HASH1 | #HASH2 |
---|---|---|
100000 |
100000 |
100000 |
10. Order By, Group By, and Having
10.1. Order By
The ORDER BY statement is used to sequence output rows.
Notes
One can order on any one of the following:
-
A named column, or an expression, neither of which need to be in the select list.
-
An unnamed column - identified by its number in the list of columns selected.
-
The ordering sequence of a specific nested subselect.
-
For an insert, the order in which the rows were inserted (see Insert Examples).
Also note:
-
One can have multiple ORDER BY statements in a query, but only one per subselect.
-
Specifying the same field multiple times in an ORDER BY list is allowed, but silly. Only the first specification of the field will have any impact on the output order.
-
If the ORDER BY column list does not uniquely identify each row, any rows with duplicate values will come out in random order. This is almost always the wrong thing to do when the data is being displayed to an end-user.
-
Use the TRANSLATE function to order data regardless of case. Note that this trick may not work consistently with some European character sets.
-
NULL values sort high.
Sample Data
The following view is used throughout this section:
CREATE VIEW SEQ_DATA(col1,col2) AS
VALUES ('ab','xy')
, ('AB','xy')
, ('ac','XY')
, ('AB','XY')
, ('Ab','12');
10.1.1. Order by Examples
The following query presents the output in ascending order:
SELECT col1
, col2
FROM seq_data
ORDER BY col1 ASC
, col2;
SEQ_DATA
COL1 | COL2 |
---|---|
ab |
xy |
AB |
xy |
ac |
XY |
AB |
XY |
Ab |
12 |
ANSWER
COL1 | COL2 |
---|---|
AB |
XY |
AB |
xy |
Ab |
12 |
ab |
xy |
ac |
XY |
In the above example, all of the lower case data comes before any of the upper case data. Use the TRANSLATE function to display the data in case-independent order:
SELECT col1
, col2
FROM seq_data
ORDER BY TRANSLATE(col1) ASC
, TRANSLATE(col2) ASC
ANSWER
COL1 | COL2 |
---|---|
Ab |
12 |
ab |
xy |
AB |
xy |
AB |
XY |
ac |
XY |
One does not have to specify the column in the ORDER BY in the select list though, to the end-user, the data may seem to be random order if one leaves it out:
SELECT col2
FROM seq_data
ORDER BY col1
, col2;
ANSWER
COL2 |
---|
XY |
xy |
12 |
xy |
XY |
In the next example, the data is (primarily) sorted in descending sequence, based on the second byte of the first column:
SELECT col1
, col2
FROM seq_data
ORDER BY SUBSTR(col1,2) DESC
, col2
, 1;
ANSWER
COL1 | COL2 |
---|---|
ac |
XY |
Ab |
12 |
ab |
xy |
AB |
XY |
AB |
xy |
The standard ASCII collating sequence defines upper-case characters as being lower than lower-case (i.e. 'A' < 'a'), so upper-case characters display first if the data is ascending order. In the next example, this is illustrated using the HEX function is used to display character data in bit-data order:
SELECT col1
, HEX(col1) AS hex1
, col2
, HEX(col2) AS hex2
FROM seq_data
ORDER BY HEX(col1)
, HEX(col2)
ANSWER
COL1 | HEX1 | COL2 | HEX2 |
---|---|---|---|
AB |
4142 |
XY |
5859 |
AB |
4142 |
xy |
7879 |
Ab |
4162 |
12 |
3132 |
ab |
6162 |
xy |
7879 |
ac |
6163 |
XY |
5859 |
10.1.2. ORDER BY subselect
One can order by the result of a nested ORDER BY, thus enabling one to order by a column that is not in the input - as is done below:
-ORDER BY nested ORDER BY
SELECT col1
FROM
(SELECT col1
FROM seq_data
ORDER BY col2
) AS xxx
ORDER BY ORDER OF xxx;
ANSWER
COL1 |
---|
Ab |
ac |
AB |
ab |
AB |
In the next example the ordering of the innermost subselect is used, in part, to order the final output. This is done by first referring it to directly, and then indirectly:
SELECT *
FROM
(SELECT *
FROM
(SELECT *
FROM seq_data
ORDER BY col2
) AS xxx
ORDER BY ORDER OF xxx
, SUBSTR(col1, 2)
) AS yyy
ORDER BY ORDER OF yyy
, col1;
ANSWER
COL1 | COL2 |
---|---|
Ab |
12 |
AB |
XY |
ac |
XY |
AB |
xy |
ac |
xy |
10.1.3. ORDER BY inserted rows
One can select from an insert statement (see Insert Examples) to see what was inserted. Order by the INSERT SEQUENCE to display the rows in the order that they were inserted:
SELECT empno
, projno AS prj
, actno AS act
, ROW_NUMBER() OVER() AS r#
FROM
FINAL TABLE
(INSERT INTO emp_act (empno, projno, actno)
VALUES ('400000','ZZZ',999)
, ('400000','VVV',111)
)
ORDER BY INPUT SEQUENCE;
ANSWER
EMPNO | PRJ | ACT | R# |
---|---|---|---|
400000 |
ZZZ |
999 |
1 |
400000 |
VVV |
111 |
2 |
The INPUT SEQUENCE phrase only works in an insert statement. It can be listed in the ORDER BY part of the statement, but not in the SELECT part. The select cannot be a nested table expression. |
10.1.4. Group By and Having
The GROUP BY and GROUPING SETS statements are used to group individual rows into combined sets based on the value in one, or more, columns. The related ROLLUP and CUBE statements are short-hand forms of particular types of GROUPING SETS statement.
Rules and Restrictions
-
There can only be one GROUP BY per SELECT. Multiple select statements in the same query can each have their own GROUP BY.
-
Every field in the SELECT list must either be specified in the GROUP BY, or must have a column function applied against it.
-
The result of a simple GROUP BY is always a distinct set of rows, where the unique identifier is whatever fields were grouped on.
-
Only expressions returning constant values (e.g. a column name, a constant) can be referenced in a GROUP BY. For example, one cannot group on the RAND function as its result varies from one call to the next. To reference such a value in a GROUP BY, resolve it beforehand using a nested-table-expression.
-
Variable length character fields with differing numbers on trailing blanks are treated as equal in the GROUP. The number of trailing blanks, if any, in the result is unpredictable.
-
When grouping, all null values in the GROUP BY fields are considered equal.
-
There is no guarantee that the rows resulting from a GROUP BY will come back in any particular order. If this is a problem, use an ORDER BY.
10.1.5. GROUP BY Flavors
A typical GROUP BY that encompasses one or more fields is actually a subset of the more general GROUPING SETS command. In a grouping set, one can do the following:
-
Summarize the selected data by the items listed such that one row is returned per unique combination of values. This is an ordinary GROUP BY.
-
Summarize the selected data using multiple independent fields. This is equivalent to doing multiple independent GROUP BY statements - with the separate results combined into one using UNION ALL statements.
-
Summarize the selected data by the items listed such that one row is returned per unique combination of values, and also get various sub-totals, plus a grand-total. Depending on what exactly is wanted, this statement can be written as a ROLLUP, or a CUBE.
To illustrate the above concepts, imagine that we want to group some company data by team, department, and division. The possible sub-totals and totals that we might want to get are:
GROUP BY division, department, team
GROUP BY division, department
GROUP BY division
GROUP BY division, team
GROUP BY department, team
GROUP BY department
GROUP BY team
GROUP BY () <= grand-total
If we wanted to get the first three totals listed above, plus the grand-total, we could write the statement one of three ways:
GROUP BY division, department, team
UNION ALL
GROUP BY division, department
UNION ALL
GROUP BY division
UNION ALL
GROUP BY ()
GROUP BY GROUPING SETS ((division, department, team)
, (division, department)
, (division)
, ())
GROUP BY ROLLUP (division, department, team)
Usage Warnings
Before we continue, be aware of the following:
-
Single vs. double parenthesis is a very big deal in grouping sets. When using the former, one is listing multiple independent groupings, while with the latter one is listing the set of items in a particular grouping.
-
Repetition matters - sometimes. In an ordinary GROUP BY duplicate references to the same field has no impact on the result. By contrast, in a GROUPING SET, ROLLUP, or CUBE statement, duplicate references can often result in the same set of data being retrieved multiple times.
10.1.6. GROUP BY Sample Data
The following view will be used throughout this section:
CREATE VIEW employee_view (d1, dept, sex, salary) AS
VALUES ('A', 'A00', 'F', 52750)
, ('A', 'A00', 'M', 29250)
, ('A', 'A00', 'M', 46500)
, ('B', 'B01', 'M', 41250)
, ('C', 'C01', 'F', 23800)
, ('C', 'C01', 'F', 28420)
, ('C', 'C01', 'F', 38250)
, ('D', 'D11', 'F', 21340)
, ('D', 'D11', 'F', 22250)
, ('D', 'D11', 'F', 29840)
, ('D', 'D11', 'M', 18270)
, ('D', 'D11', 'M', 20450)
, ('D', 'D11', 'M', 24680)
, ('D', 'D11', 'M', 25280)
, ('D', 'D11', 'M', 27740)
, ('D', 'D11', 'M', 32250);
VIEW CONTENTS
D1 | DEPT | SEX | SALARY |
---|---|---|---|
A |
A00 |
F |
52750 |
A |
A00 |
M |
29250 |
A |
A00 |
M |
46500 |
B |
B01 |
M |
41250 |
C |
C01 |
F |
23800 |
C |
C01 |
F |
28420 |
C |
C01 |
F |
38250 |
D |
D11 |
F |
21340 |
D |
D11 |
F |
22250 |
D |
D11 |
F |
29840 |
D |
D11 |
M |
18270 |
D |
D11 |
M |
20450 |
D |
D11 |
M |
24680 |
D |
D11 |
M |
25280 |
D |
D11 |
M |
27740 |
D |
D11 |
M |
32250 |
10.1.7. Simple GROUP BY Statements
A simple GROUP BY is used to combine individual rows into a distinct set of summary rows.
Sample Queries
In this first query we group our sample data by the leftmost three fields in the view:
SELECT d1
, dept
, sex
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
FROM employee_view
WHERE dept <> 'ABC'
GROUP BY d1
, dept
, sex
HAVING dept > 'A0'
AND (SUM(salary) > 100
OR MIN(salary) > 10
OR COUNT(*) <> 22)
ORDER BY d1
, dept
, sex;
ANSWER
D1 | DEPT | SEX | SALARY | #ROWS |
---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
A |
A00 |
M |
75750 |
2 |
B |
B01 |
M |
41250 |
1 |
C |
C01 |
F |
90470 |
3 |
D |
D11 |
F |
73430 |
3 |
D |
D11 |
M |
148670 |
6 |
There is no need to have a field in the GROUP BY in the SELECT list, but the answer really doesn’t make much sense if one does this:
SELECT sex
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
FROM employee_view
WHERE sex IN ('F','M')
GROUP BY dept
, sex
ORDER BY sex;
ANSWER
SEX | SALARY | #ROWS |
---|---|---|
F |
52750 |
1 |
F |
90470 |
3 |
F |
73430 |
3 |
M |
75750 |
2 |
M |
41250 |
1 |
M |
148670 |
6 |
One can also do a GROUP BY on a derived field, which may, or may not be, in the statement SELECT list. This is an amazingly stupid thing to do:
SELECT SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
FROM employee_view
WHERE d1 <> 'X'
GROUP BY SUBSTR(dept,3,1)
HAVING COUNT(*) <> 99;
ANSWER
SALARY | #ROWS |
---|---|
128500 |
3 |
353820 |
13 |
One can not refer to the name of a derived column in a GROUP BY statement. Instead, one has to repeat the actual derivation code. One can however refer to the new column name in an ORDER BY:
SELECT SUBSTR(dept, 3, 1) AS wpart
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
FROM employee_view
GROUP BY SUBSTR(dept, 3, 1)
ORDER BY wpart DESC;
ANSWER
WPART | SALARY | #ROWS |
---|---|---|
1 |
353820 |
13 |
0 |
128500 |
3 |
10.1.8. GROUPING SETS Statement
The GROUPING SETS statement enables one to get multiple GROUP BY result sets using a single statement. It is important to understand the difference between nested (i.e. in secondary parenthesis), and non-nested GROUPING SETS sub-phrases:
-
A nested list of columns works as a simple GROUP BY.
-
A non-nested list of columns works as separate simple GROUP BY statements, which are then combined in an implied UNION ALL.
GROUP BY GROUPING SETS ((A,B,C)) is equivalent to GROUP BY A , B , C
GROUP BY GROUPING SETS (A,B,C) is equivalent to GROUP BY A UNION ALL
GROUP BY B UNION ALL GROUP BY C
GROUP BY GROUPING SETS (A,(B,C)) is equivalent to GROUP BY A UNION ALL
GROUP BY B , C
Multiple GROUPING SETS in the same GROUP BY are combined together as if they were simple fields in a GROUP BY list:
GROUP BY GROUPING SETS (A) is equivalent to GROUP BY A
, GROUPING SETS (B) , B
, GROUPING SETS (C) , C
GROUP BY GROUPING SETS (A) is equivalent to GROUP BY A
, GROUPING SETS ((B,C)) , B
, C
GROUP BY GROUPING SETS (A) is equivalent to GROUP BY A
, GROUPING SETS (B,C) , B
UNION ALL
GROUP BY A
, C
One can mix simple expressions and GROUPING SETS in the same GROUP BY:
GROUP BY A is equivalent to GROUP BY A
, GROUPING SETS ((B,C)) , B
, C
Repeating the same field in two parts of the GROUP BY will result in different actions depending on the nature of the repetition. The second field reference is ignored if a standard GROUP BY is being made, and used if multiple GROUP BY statements are implied:
GROUP BY A is equivalent to GROUP BY A
, B , B
, GROUPING SETS ((B,C)) , C
GROUP BY A is equivalent to GROUP BY A
, B , B
, GROUPING SETS (B,C) , C
UNION ALL
GROUP BY A
, B
GROUP BY A is equivalent to GROUP BY A
, B , B
, C , C
, GROUPING SETS (B,C) UNION ALL
GROUP BY A
, B
, C
A single GROUPING SETS statement can contain multiple sets of (implied) GROUP BY phrases. These are combined using implied UNION ALL statements:
GROUP BY GROUPING SETS ((A,B,C) is equivalent to GROUP BY A
, (A,B) , B
, (C)) , C
UNION ALL
GROUP BY A
, B
UNION ALL
GROUP BY C
GROUP BY GROUPING SETS ((A) is equivalent to GROUP BY A
, (B,C) UNION ALL
, (A) GROUP BY B
, A , C
, ((C))) UNION ALL
GROUP BY A
UNION ALL
GROUP BY A
UNION ALL
GROUP BY C
The null-field list "( )" can be used to get a grand total. This is equivalent to not having the GROUP BY at all.
GROUP BY GROUPING SETS ((A,B,C) is equivalent to GROUP BY A
, (A,B) , B
, (A) , C
, ()) UNION ALL
GROUP BY A
, B
is equivalent to UNION ALL
GROUP BY A
UNION ALL
ROLLUP(A, B, C) grand-totl
The above GROUPING SETS statement is equivalent to a ROLLUP(A,B,C), while the next is equivalent to a CUBE(A,B,C):
GROUP BY GROUPING SETS ((A,B,C) is equivalent to GROUP BY A
, (A,B) , B
, (A,C) , C
, (B,C) UNION ALL
, (A) GROUP BY A
, (B) , B
, (C) UNION ALL
, ()) GROUP BY A
, C
UNION ALL
GROUP BY B
is equivalent to , C
UNION ALL
GROUP BY A
UNION ALL
CUBE(A,B,C) GROUP BY B
UNION ALL
GROUP BY C
UNION ALL
grand-totl
10.1.9. SQL Examples
This first example has two GROUPING SETS. Because the second is in nested parenthesis, the result is the same as a simple three-field group by:
SELECT d1
, dept
, sex
, SUM(salary) AS sal
, SMALLINT(COUNT(*)) AS #r
, GROUPING(d1) AS f1
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY GROUPING SETS (d1)
, GROUPING SETS ((dept, sex))
ORDER BY d1
, dept
, sex;
ANSWER
D1 | DEPT | SEX | SAL | #R | DF | WF | SF |
---|---|---|---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
0 |
0 |
0 |
A |
A00 |
M |
75750 |
2 |
0 |
0 |
0 |
B |
B01 |
M |
41250 |
1 |
0 |
0 |
0 |
C |
C01 |
F |
90470 |
3 |
0 |
0 |
0 |
D |
D11 |
F |
73430 |
3 |
0 |
0 |
0 |
D |
D11 |
M |
148670 |
6 |
0 |
0 |
0 |
The GROUPING(field-name) column function is used in these examples to identify what rows come from which particular GROUPING SET. A value of 1 indicates that the corresponding data field is null because the row is from of a GROUPING SET that does not involve this row. Otherwise, the value is zero. |
In the next query, the second GROUPING SET is not in nested-parenthesis. The query is therefore equivalent to GROUP BY D1, DEPT UNION ALL GROUP BY D1, SEX:
SELECT d1
, dept
, sex
, SUM(salary) AS sal
, SMALLINT(COUNT(*)) AS #r
, GROUPING(d1) AS f1
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY GROUPING SETS (d1)
, GROUPING SETS (dept, sex)
ORDER BY d1
, dept
, sex;
ANSWER
D1 | DEPT | SEX | SAL | #R | F1 | FD | FS |
---|---|---|---|---|---|---|---|
A |
A00 |
- |
128500 |
3 |
0 |
0 |
1 |
A |
- |
F |
52750 |
1 |
0 |
1 |
0 |
A |
- |
M |
75750 |
2 |
0 |
1 |
0 |
B |
B01 |
- |
41250 |
1 |
0 |
0 |
1 |
B |
- |
M |
41250 |
1 |
0 |
1 |
0 |
C |
C01 |
- |
90470 |
3 |
0 |
0 |
1 |
C |
- |
F |
90470 |
3 |
0 |
1 |
0 |
D |
D11 |
- |
222100 |
9 |
0 |
0 |
1 |
D |
- |
F |
73430 |
3 |
0 |
1 |
0 |
D |
- |
M |
148670 |
6 |
0 |
1 |
0 |
It is generally unwise to repeat the same field in both ordinary GROUP BY and GROUPING SETS statements, because the result is often rather hard to understand. To illustrate, the following two queries differ only in their use of nested-parenthesis. Both of them repeat the DEPT field:
-
In the first, the repetition is ignored, because what is created is an ordinary GROUP BY on all three fields.
-
In the second, repetition is important, because two GROUP BY statements are implicitly generated. The first is on D1 and DEPT. The second is on D1, DEPT, and SEX.
SELECT d1
, dept
, sex
, SUM(salary) AS sal
, SMALLINT(COUNT(*)) AS #r
, GROUPING(d1) AS f1
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY d1
, dept
, GROUPING SETS ((dept, sex))
ORDER BY d1
, dept
, sex;
ANSWER
D1 | DEPT | SEX | SAL | #R | F1 | FD | FS |
---|---|---|---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
0 |
0 |
0 |
A |
A00 |
M |
75750 |
2 |
0 |
0 |
0 |
B |
B01 |
M |
41250 |
1 |
0 |
0 |
0 |
C |
C01 |
F |
90470 |
3 |
0 |
0 |
0 |
D |
D11 |
F |
73430 |
3 |
0 |
0 |
0 |
D |
D11 |
M |
148670 |
6 |
0 |
0 |
0 |
SELECT d1
, dept
, sex
, SUM(salary) AS sal
, SMALLINT(COUNT(*)) AS #r
, GROUPING(d1) AS f1
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY d1
, dept
, GROUPING SETS (dept, sex)
ORDER BY d1
, dept
, sex;
ANSWER
D1 | DEPT | SEX | SAL | #R | F1 | FD | FS |
---|---|---|---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
0 |
0 |
0 |
A |
A00 |
M |
75750 |
2 |
0 |
0 |
0 |
A |
A00 |
128500 |
3 |
0 |
0 |
1 |
|
B |
B01 |
M |
41250 |
1 |
0 |
0 |
0 |
B |
B01 |
41250 |
1 |
0 |
0 |
1 |
|
C |
C01 |
F |
90470 |
3 |
0 |
0 |
0 |
C |
C01 |
90470 |
3 |
0 |
0 |
1 |
|
D |
D11 |
73430 |
3 |
0 |
0 |
0 |
|
D |
D11 |
M |
148670 |
6 |
0 |
0 |
0 |
D |
D11 |
222100 |
9 |
0 |
0 |
1 |
The above two queries can be rewritten as follows:
GROUP BY d1 is equivalent to GROUP BY d1
, dept , dept
, GROUPING SETS ((dept, sex)) , sex
GROUP BY d1 is equivalent to GROUP BY d1
, dept , dept
, GROUPING SETS (dept, sex) , sex
UNION ALL
GROUP BY d1
, dept
, dept
Repetitions of the same field in a GROUP BY (as is done above) are ignored during query processing. Therefore GROUP BY D1, DEPT, DEPT, SEX is the same as GROUP BY D1, DEPT, SEX. |
10.1.10. ROLLUP Statement
A expression displays sub-totals for the specified fields. This is equivalent to doing the original GROUP BY, and also doing more groupings on sets of the left-most columns.
GROUP BY ROLLUP(A,B,C) ===> GROUP BY GROUPING SETS((A,B,C)
, (A,B)
, (A)
, ())
GROUP BY ROLLUP(C,B) ===> GROUP BY GROUPING SETS((C,B)
, (C)
, ())
GROUP BY ROLLUP(A) ===> GROUP BY GROUPING SETS((A)
, ())
Imagine that we wanted to GROUP BY, but not ROLLUP one field in a list of fields. To do this, we simply combine the field to be removed with the next more granular field:
GROUP BY ROLLUP(A,(B,C)) ===> GROUP BY GROUPING SETS((A,B,C)
, (A)
, ())
Multiple ROLLUP statements in the same GROUP BY act independently of each other:
GROUP BY ROLLUP(A) ===> GROUP BY GROUPING SETS((A,B,C)
, ROLLUP(B,C) , (A,B)
, (A)
, (B, C)
, (B)
, ())
One way to understand the above is to convert the two ROLLUP statement into equivalent grouping sets, and them "multiply" them - ignoring any grand-totals except when they are on both sides of the equation:
ROLLUP(A) * ROLLUP(B,C) = GROUPING SETS((A,B,C)
, (A,B)
, (A)
GROUPING SETS((A) * GROUPING SETS((B,C) = , (B,C)
, ()) , (B) , (B)
, ()) , (())
SQL Examples
Here is a standard GROUP BY that gets no sub-totals:
SELECT dept
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
FROM employee_view
GROUP BY dept
ORDER BY dept;
ANSWER
DEPT | SALARY | #ROWS | FD |
---|---|---|---|
A00 |
128500 |
3 |
0 |
B01 |
41250 |
1 |
0 |
C01 |
0470 |
3 |
0 |
D11 |
222100 |
9 |
0 |
Imagine that we wanted to also get a grand total for the above. Below is an example of using the ROLLUP statement to do this:
SELECT dept
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS FD
FROM employee_view
GROUP BY ROLLUP(dept)
ORDER BY dept;
ANSWER
DEPT | SALARY | #ROWS | FD |
---|---|---|---|
A00 |
128500 |
3 |
0 |
B01 |
41250 |
1 |
0 |
C01 |
90470 |
3 |
0 |
D11 |
222100 |
9 |
0 |
- |
482320 |
16 |
1 |
The GROUPING(field-name) function that is selected in the above example returns a one when the output row is a summary row, else it returns a zero. |
Alternatively, we could do things the old-fashioned way and use a UNION ALL to combine the original GROUP BY with an all-row summary:
SELECT dept
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
FROM employee_view
GROUP BY dept
UNION ALL
SELECT CAST(NULL AS CHAR(3)) AS dept
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, CAST(1 AS INTEGER) AS fd
FROM employee_view
ORDER BY dept;
ANSWER
DEPT | SALARY | #ROWS | FD |
---|---|---|---|
A00 |
128500 |
3 |
0 |
B01 |
41250 |
1 |
0 |
C01 |
90470 |
3 |
0 |
D11 |
222100 |
9 |
0 |
- |
482320 |
16 |
1 |
Specifying a field both in the original GROUP BY, and in a ROLLUP list simply results in every data row being returned twice. In other words, the result is garbage:
SELECT dept
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
FROM employee_view
GROUP BY dept
, ROLLUP(dept)
ORDER BY dept;
ANSWER
DEPT | SALARY | #ROWS | FD |
---|---|---|---|
A00 |
128500 |
3 |
0 |
A00 |
128500 |
3 |
0 |
B01 |
41250 |
1 |
0 |
B01 |
41250 |
1 |
0 |
C01 |
90470 |
3 |
0 |
C01 |
90470 |
3 |
0 |
D11 |
222100 |
9 |
0 |
D11 |
222100 |
9 |
0 |
Below is a graphic representation of why the data rows were repeated above. Observe that two GROUP BY statements were, in effect, generated:
GROUP BY dept => GROUP BY dept => GROUP BY dept
, ROLLUP(dept) , GROUPING SETS((dept) UNION ALL
, ()) GROUP BY dept
, ()
In the next example the GROUP BY, is on two fields, with the second also being rolled up:
SELECT dept
, sex
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY dept
, ROLLUP(sex)
ORDER BY dept
, sex;
ANSWER
DEPT | SEX | SALARY | #ROWS | FD | FS |
---|---|---|---|---|---|
A00 |
F |
52750 |
1 |
0 |
0 |
A00 |
M |
75750 |
2 |
0 |
0 |
A00 |
- |
128500 |
3 |
0 |
1 |
B01 |
M |
41250 |
1 |
0 |
0 |
B01 |
- |
41250 |
1 |
0 |
1 |
C01 |
F |
90470 |
3 |
0 |
0 |
C01 |
- |
90470 |
3 |
0 |
1 |
D11 |
F |
73430 |
3 |
0 |
0 |
D11 |
M |
148670 |
6 |
0 |
0 |
D11 |
- |
222100 |
9 |
0 |
1 |
The next example does a ROLLUP on both the DEPT and SEX fields, which means that we will get rows for the following:
-
The work-department and sex field combined (i.e. the original raw GROUP BY).
-
A summary for all sexes within an individual work-department.
-
A summary for all work-departments (i.e. a grand-total).
SELECT dept
, sex
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY ROLLUP(dept , sex)
ORDER BY dept , sex;
ANSWER
DEPT | SEX | SALARY | #ROWS | FD | FS |
---|---|---|---|---|---|
A00 |
F |
52750 |
1 |
0 |
0 |
A00 |
M |
75750 |
2 |
0 |
0 |
A00 |
- |
128500 |
3 |
0 |
1 |
B01 |
M |
41250 |
1 |
0 |
0 |
B01 |
- |
41250 |
1 |
0 |
1 |
C01 |
F |
90470 |
3 |
0 |
0 |
C01 |
- |
90470 |
3 |
0 |
1 |
D11 |
F |
73430 |
3 |
0 |
0 |
D11 |
M |
148670 |
6 |
0 |
0 |
D11 |
- |
222100 |
9 |
0 |
1 |
- |
- |
482320 |
16 |
1 |
1 |
In the next example we have reversed the ordering of fields in the ROLLUP statement. To make things easier to read, we have also altered the ORDER BY sequence. Now get an individual row for each sex and work-department value, plus a summary row for each sex:, plus a grand-total row:
SELECT sex
, dept
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY ROLLUP(sex
, dept)
ORDER BY sex
, dept;
ANSWER
SEX | DEPT | SALARY | #ROWS | FD | FS |
---|---|---|---|---|---|
F |
A00 |
52750 |
1 |
0 |
0 |
F |
C01 |
90470 |
3 |
0 |
0 |
F |
D11 |
73430 |
3 |
0 |
0 |
F |
216650 |
7 |
1 |
0 |
|
M |
A00 |
75750 |
2 |
0 |
0 |
M |
B01 |
41250 |
1 |
0 |
0 |
M |
D11 |
148670 |
6 |
0 |
0 |
M |
- |
265670 |
9 |
1 |
0 |
- |
- |
482320 |
16 |
1 |
1 |
The next statement is the same as the prior, but it uses the logically equivalent GROUPING SETS syntax:
SELECT sex
, dept
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY GROUPING SETS ((sex, dept)
, (sex)
, ())
ORDER BY sex
, dept;
ANSWER
SEX | DEPT | SALARY | #ROWS | FD | FS |
---|---|---|---|---|---|
F |
A00 |
52750 |
1 |
0 |
0 |
F |
C01 |
90470 |
3 |
0 |
0 |
F |
D11 |
73430 |
3 |
0 |
0 |
F |
- |
216650 |
7 |
1 |
0 |
M |
A00 |
75750 |
2 |
0 |
0 |
M |
B01 |
41250 |
1 |
0 |
0 |
M |
D11 |
148670 |
6 |
0 |
0 |
M |
- |
265670 |
9 |
1 |
0 |
- |
- |
482320 |
16 |
1 |
1 |
The next example has two independent rollups:
-
The first generates a summary row for each sex.
-
The second generates a summary row for each work-department.
The two together make a (single) combined summary row of all matching data. This query is the same as a UNION of the two individual rollups, but it has the advantage of being done in a single pass of the data. The result is the same as a CUBE of the two fields:
SELECT sex
, dept
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY ROLLUP(sex)
, ROLLUP(dept)
ORDER BY sex
, dept;
ANSWER
SEX | DEPT | SALARY | #ROWS | FD | FS |
---|---|---|---|---|---|
F |
A00 |
52750 |
1 |
0 |
0 |
F |
C01 |
90470 |
3 |
0 |
0 |
F |
D11 |
73430 |
3 |
0 |
0 |
F |
- |
216650 |
7 |
1 |
0 |
M |
A00 |
75750 |
2 |
0 |
0 |
M |
B01 |
41250 |
1 |
0 |
0 |
M |
D11 |
148670 |
6 |
0 |
0 |
M |
- |
265670 |
9 |
1 |
0 |
- |
A00 |
128500 |
3 |
0 |
1 |
- |
B01 |
41250 |
1 |
0 |
1 |
- |
C01 |
90470 |
3 |
0 |
1 |
- |
D11 |
222100 |
9 |
0 |
1 |
- |
- |
482320 |
16 |
1 |
1 |
Below we use an inner set of parenthesis to tell the ROLLUP to treat the two fields as one, which causes us to only get the detailed rows, and the grand-total summary:
SELECT dept
, sex
, SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY ROLLUP((dept,sex))
ORDER BY dept
, sex;
ANSWER
DEPT | SEX | SALARY | #ROWS | FD | FS |
---|---|---|---|---|---|
A00 |
F |
52750 |
1 |
0 |
0 |
A00 |
M |
75750 |
2 |
0 |
0 |
B01 |
M |
41250 |
1 |
0 |
0 |
C01 |
F |
90470 |
3 |
0 |
0 |
D11 |
F |
73430 |
3 |
0 |
0 |
D11 |
M |
148670 |
6 |
0 |
0 |
- |
- |
482320 |
16 |
1 |
1 |
The HAVING statement can be used to refer to the two GROUPING fields. For example, in the following query, we eliminate all rows except the grand total:
SELECT SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
FROM employee_view
GROUP BY ROLLUP(sex
, dept)
HAVING GROUPING(dept) = 1 AND
GROUPING(sex) = 1
ORDER BY salary;
ANSWER
SALARY | #ROWS |
---|---|
482320 |
16 |
Below is a logically equivalent SQL statement:
SELECT SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
FROM employee_view
GROUP BY GROUPING SETS(());
ANSWER
SALARY | #ROWS |
---|---|
482320 |
16 |
Here is another:
SELECT SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
FROM employee_view
GROUP BY ();
ANSWER
SALARY | #ROWS |
---|---|
482320 |
16 |
And another:
SELECT SUM(salary) AS salary
, SMALLINT(COUNT(*)) AS #rows
FROM employee_view;
ANSWER
SALARY | #ROWS |
---|---|
482320 |
16 |
10.1.11. CUBE Statement
A CUBE expression displays a cross-tabulation of the sub-totals for any specified fields. As such, it generates many more totals than the similar ROLLUP.
GROUP BY CUBE(A,B,C) ===> GROUP BY GROUPING SETS((A,B,C)
, (A,B)
, (A,C)
, (B,C)
, (A)
, (B)
, (C)
, ())
GROUP BY CUBE(C,B) ===> GROUP BY GROUPING SETS((C,B)
, (C)
, (B)
, ())
GROUP BY CUBE(A) ===> GROUP BY GROUPING SETS((A)
, ())
As with the ROLLLUP statement, any set of fields in nested parenthesis is treated by the CUBE as a single field:
GROUP BY CUBE(A,(B,C)) ===> GROUP BY GROUPING SETS((A,B,C)
, (B,C)
, (A)
, ())
Having multiple CUBE statements is allowed, but very, very silly:
GROUP BY CUBE(A,B) ==> GROUPING SETS((A,B,C),(A,B),(A,B,C),(A,B)
, CUBE(B,C) , (A,B,C),(A,B),(A,C),(A)
, (B,C),(B),(B,C),(B)
, (B,C),(B),(C),())
Obviously, the above is a lot of GROUPING SETS, and even more underlying GROUP BY statements. Think of the query as the Cartesian Product of the two CUBE statements, which are first resolved down into the following two GROUPING SETS:
((A,B),(A),(B),())
((B,C),(B),(C),())
10.1.12. SQL Examples
Below is a standard CUBE statement:
SELECT d1
, dept
, sex
, INT(SUM(salary)) AS sal
, SMALLINT(COUNT(*)) AS #r
, GROUPING(d1) AS f1
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY CUBE(d1, dept, sex)
ORDER BY d1
, dept
, sex;
ANSWER
D1 | DEPT | SEX | SAL | #R | F1 | FD | FS |
---|---|---|---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
0 |
0 |
0 |
A |
A00 |
M |
75750 |
2 |
0 |
0 |
0 |
A |
A00 |
- |
128500 |
3 |
0 |
0 |
1 |
A |
F |
52750 |
1 |
0 |
1 |
0 |
|
A |
M |
75750 |
2 |
0 |
1 |
0 |
|
A |
128500 |
3 |
0 |
1 |
1 |
||
B |
B01 |
M |
41250 |
1 |
0 |
0 |
0 |
B |
B01 |
41250 |
1 |
0 |
0 |
1 |
|
B |
M |
41250 |
1 |
0 |
1 |
0 |
|
B |
41250 |
1 |
0 |
1 |
1 |
||
C |
C01 |
F |
90470 |
3 |
0 |
0 |
0 |
C |
C01 |
90470 |
3 |
0 |
0 |
1 |
|
C |
F |
90470 |
3 |
0 |
1 |
0 |
|
C |
90470 |
3 |
0 |
1 |
1 |
||
D |
D11 |
F |
73430 |
3 |
0 |
0 |
0 |
D |
D11 |
M |
148670 |
6 |
0 |
0 |
0 |
D |
D11 |
222100 |
9 |
0 |
0 |
1 |
|
D |
F |
73430 |
3 |
0 |
1 |
0 |
|
D |
M |
148670 |
6 |
0 |
1 |
0 |
|
D |
222100 |
9 |
0 |
1 |
1 |
||
- |
A00 |
F |
2750 |
1 |
1 |
0 |
0 |
- |
A00 |
M |
75750 |
2 |
1 |
0 |
0 |
- |
A00 |
128500 |
3 |
1 |
0 |
1 |
|
- |
B01 |
M |
41250 |
1 |
1 |
0 |
0 |
- |
B01 |
41250 |
1 |
1 |
0 |
1 |
|
- |
C01 |
F |
90470 |
3 |
1 |
0 |
0 |
- |
C01 |
90470 |
3 |
1 |
0 |
1 |
|
- |
D11 |
F |
73430 |
3 |
1 |
0 |
0 |
- |
D11 |
M |
148670 |
6 |
1 |
0 |
0 |
- |
D11 |
222100 |
9 |
1 |
0 |
1 |
|
- |
F |
216650 |
7 |
1 |
1 |
0 |
|
- |
M |
265670 |
9 |
1 |
1 |
0 |
|
- |
482320 |
16 |
1 |
1 |
1 |
Here is the same query expressed as GROUPING SETS;
SELECT d1
, dept
, sex
, INT(SUM(salary)) AS sal
, SMALLINT(COUNT(*)) AS #r
, GROUPING(d1) AS f1
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_view
GROUP BY GROUPING SETS ((d1, dept, sex)
, (d1,dept)
, (d1,sex)
, (dept,sex)
, (d1)
, (dept)
, (sex)
, ())
ORDER BY d1
, dept
, sex;
ANSWER
D1 | DEPT | SEX | SAL | #R | F1 | FD | FS |
---|---|---|---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
0 |
0 |
0 |
A |
A00 |
M |
75750 |
2 |
0 |
0 |
0 |
etc... (same as prior query)
A CUBE on a list of columns in nested parenthesis acts as if the set of columns was only one field. The result is that one gets a standard GROUP BY (on the listed columns), plus a row with the grand-totals:
SELECT d1
, dept
, sex
, INT(SUM(salary)) AS sal
, SMALLINT(COUNT(*)) AS #r
, GROUPING(d1) AS f1
, GROUPING(dept) AS fd
, GROUPING(sex) AS fs
FROM employee_VIEW
GROUP BY CUBE((d1, dept, sex))
ORDER BY d1
, dept
, sex;
ANSWER
D1 | DEPT | SEX | SAL | #R | F1 | FD | FS |
---|---|---|---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
0 |
0 |
0 |
A |
A00 |
M |
75750 |
2 |
0 |
0 |
0 |
B |
B01 |
M |
41250 |
1 |
0 |
0 |
0 |
C |
C01 |
F |
90470 |
3 |
0 |
0 |
0 |
D |
D11 |
F |
73430 |
3 |
0 |
0 |
0 |
D |
D11 |
M |
148670 |
6 |
0 |
0 |
0 |
- |
482320 |
16 |
1 |
1 |
1 |
The above query is resolved thus:
GROUP BY CUBE((A,B,C)) => GROUP BY GROUPING SETS((A,B,C) => GROUP BY A
, ()) , B
, C
UNION ALL
GROUP BY()
Complex Grouping Sets - Done Easy
Many of the more complicated SQL statements illustrated above are essentially unreadable because it is very hard to tell what combinations of fields are being rolled up, and what are not. There ought to be a more user-friendly way and, fortunately, there is. The CUBE command can be used to roll up everything. Then one can use ordinary SQL predicates to select only those totals and sub-totals that one wants to display.
Queries with multiple complicated ROLLUP and/or GROUPING SET statements sometimes fail to compile. In which case, this method can be used to get the answer. |
To illustrate this technique, consider the following query. It summarizes the data in the sample view by three fields:
SELECT d1 AS d1
, dept AS dpt
, sex AS sx
, INT(SUM(salary)) AS sal
, SMALLINT(COUNT(*)) AS r
FROM employee_VIEW
GROUP BY d1
, dept
, sex
ORDER BY 1,2,3;
ANSWER
D1 | DPT | SX | SAL | R |
---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
A |
A00 |
M |
75750 |
2 |
B |
B01 |
M |
41250 |
1 |
C |
C01 |
F |
90470 |
3 |
D |
D11 |
F |
73430 |
3 |
D |
D11 |
M |
148670 |
6 |
Now imagine that we want to extend the above query to get the following sub-total rows:
DESIRED SUB-TOTALS EQUIVILENT TO
D1, DEPT, and SEX. GROUP BY GROUPING SETS ((d1,dept,sex)
D1 and DEPT. , (d1,dept)
D1 and SEX. , (d1,sex)
D1. , (d1)
SEX. , (sex)
Grand total. EQUIVILENT TO , ())
GROUP BY ROLLUP(d1,dept)
, ROLLUP(sex)
Rather than use either of the syntaxes shown on the right above, below we use the CUBE expression to get all sub-totals, and then select those that we want:
SELECT *
FROM (SELECT d1 AS d1
, dept AS dpt
, sex AS sx
, INT(SUM(salary)) AS sal
, SMALLINT(COUNT(*)) AS #r
, SMALLINT(GROUPING(d1)) AS g1
, SMALLINT(GROUPING(dept)) AS gd
, SMALLINT(GROUPING(sex)) AS gs
FROM EMPLOYEE_VIEW
GROUP BY CUBE(d1,dept,sex)
) AS xxx
WHERE (g1,gd,gs) = (0,0,0)
OR (g1,gd,gs) = (0,0,1)
OR (g1,gd,gs) = (0,1,0)
OR (g1,gd,gs) = (0,1,1)
OR (g1,gd,gs) = (1,1,0)
OR (g1,gd,gs) = (1,1,1)
ORDER BY 1,2,3;
ANSWER
D1 | DPT | SX | SAL | #R | G1 | GD | GS |
---|---|---|---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
0 |
0 |
0 |
A |
A00 |
M |
75750 |
2 |
0 |
0 |
0 |
A |
A00 |
- |
128500 |
3 |
0 |
0 |
1 |
A |
F |
52750 |
1 |
0 |
1 |
0 |
|
A |
M |
75750 |
2 |
0 |
1 |
0 |
|
A |
- |
128500 |
3 |
0 |
1 |
1 |
|
B |
B01 |
M |
41250 |
1 |
0 |
0 |
0 |
B |
B01 |
41250 |
1 |
0 |
0 |
1 |
|
B |
M |
41250 |
1 |
0 |
1 |
0 |
|
B |
41250 |
1 |
0 |
1 |
1 |
||
C |
C01 |
F |
90470 |
3 |
0 |
0 |
0 |
C |
C01 |
90470 |
3 |
0 |
0 |
1 |
|
C |
F |
90470 |
3 |
0 |
1 |
0 |
|
C |
90470 |
3 |
0 |
1 |
1 |
||
D |
D11 |
F |
73430 |
3 |
0 |
0 |
0 |
D |
D11 |
M |
148670 |
6 |
0 |
0 |
0 |
D |
D11 |
- |
222100 |
9 |
0 |
0 |
1 |
D |
F |
73430 |
3 |
0 |
1 |
0 |
|
D |
M |
148670 |
6 |
0 |
1 |
0 |
|
D |
- |
222100 |
9 |
0 |
1 |
1 |
|
- |
F |
216650 |
7 |
1 |
1 |
0 |
|
- |
M |
265670 |
9 |
1 |
1 |
0 |
|
- |
- |
482320 |
16 |
1 |
1 |
1 |
In the above query, the GROUPING function (see GROUPING) is used to identify what fields are being summarized on each row. A value of one indicates that the field is being summarized; while a value of zero means that it is not. Only the following combinations are kept:
(G1,GD,GS) = (0,0,0) <== D1, DEPT, SEX (G1,GD,GS) = (0,0,1) <== D1, DEPT (G1,GD,GS) = (0,1,0) <== D1, SEX (G1,GD,GS) = (0,1,1) <== D1, (G1,GD,GS) = (1,1,0) <== SEX, (G1,GD,GS) = (1,1,1) <== grand total
Here is the same query written using two ROLLUP expressions. You can be the judge as to which is the easier to understand:
SELECT d1
, dept
, sex
, INT(SUM(salary)) AS sal
, SMALLINT(COUNT(*)) AS #r
FROM employee_view
GROUP BY ROLLUP(d1, dept)
, ROLLUP(sex)
ORDER BY 1,2,3;
ANSWER
D1 | DEPT | SEX | SAL | #R |
---|---|---|---|---|
A |
A00 |
F |
52750 |
1 |
A |
A00 |
M |
75750 |
2 |
A |
A00 |
128500 |
3 |
|
A |
F |
52750 |
1 |
|
A |
M |
75750 |
2 |
|
A |
128500 |
3 |
||
B |
B01 |
M |
41250 |
1 |
B |
B01 |
41250 |
1 |
|
B |
M |
41250 |
1 |
|
B |
41250 |
1 |
||
C |
C01 |
F |
90470 |
3 |
C |
C01 |
90470 |
3 |
|
C |
F |
90470 |
3 |
|
C |
90470 |
3 |
||
D |
D11 |
F |
73430 |
3 |
D |
D11 |
M |
148670 |
6 |
D |
D11 |
222100 |
9 |
|
D |
F |
73430 |
3 |
|
D |
M |
148670 |
6 |
|
D |
222100 |
9 |
||
- |
F |
216650 |
7 |
|
- |
M |
265670 |
9 |
|
- |
482320 |
16 |
10.1.13. Group By and Order By
One should never assume that the result of a GROUP BY will be a set of appropriately ordered rows because Db2 may choose to use a "strange" index for the grouping so as to avoid doing a row sort. For example, if one says "GROUP BY C1, C2" and the only suitable index is on C2 descending and then C1, the data will probably come back in index-key order.
SELECT dept
, job
, COUNT(*)
FROM staff
GROUP BY dept, job
ORDER BY dept, job;
Always code an ORDER BY if there is a need for the rows returned from the query to be specifically ordered - which there usually is._ |
10.1.14. Group By in Join
We want to select those rows in the STAFF table where the average SALARY for the employee’s DEPT is greater than $18,000. Answering this question requires using a JOIN and GROUP BY in the same statement. The GROUP BY will have to be done first, then its' result will be joined to the STAFF table. There are two syntactically different, but technically similar, ways to write this query. Both techniques use a temporary table, but the way by which this is expressed differs. In the first example, we shall use a common table expression:
WITH staff2 (dept, avgsal) AS
(SELECT dept
, AVG(salary)
FROM staff
GROUP BY dept
HAVING AVG(salary) > 18000
)
SELECT a.id
, a.name
, a.dept
FROM staff a
, staff2 b
WHERE a.dept = b.dept
ORDER BY a.id;
ANSWER
ID | NAME | DEPT |
---|---|---|
160 |
Molinare |
10 |
210 |
Lu |
10 |
240 |
Daniels |
10 |
260 |
Jones |
10 |
In the next example, we shall use a fullselect:
SELECT a.id
, a.name
, a.dept
FROM staff a
, (SELECT dept AS dept
, AVG(salary) AS avgsal
FROM staff
GROUP BY dept
HAVING AVG(salary) > 18000
) AS b
WHERE a.dept = b.dept
ORDER BY a.id;
ANSWER
ID | NAME | DEPT |
---|---|---|
160 |
Molinare |
10 |
210 |
Lu |
10 |
240 |
Daniels |
10 |
260 |
Jones |
10 |
10.1.15. COUNT and No Rows
When there are no matching rows, the value returned by the COUNT depends upon whether this is a GROUP BY in the SQL statement or not:
SELECT COUNT(*) AS c1
FROM staff
WHERE id < 1;
ANSWER ==> 0
SELECT COUNT(*) AS c1
FROM staff
WHERE id < 1
GROUP BY id;
ANSWER ==> no row
see No Rows Match for a comprehensive discussion of what happens when no rows match.
11. Joins
A join is used to relate sets of rows in two or more logical tables. The tables are always joined on a row-by-row basis using whatever join criteria are provided in the query. The result of a join is always a new, albeit possibly empty, set of rows. In a join, the matching rows are joined side-by-side to make the result table. By contrast, in a union (see Union, Intersect, and Except) the matching rows are joined (in a sense) one-above-the-other to make the result table.
11.1. Why Joins Matter
The most important data in a relational database is not that stored in the individual rows. Rather, it is the implied relationships between sets of related rows. For example, individual rows in an EMPLOYEE table may contain the employee ID and salary - both of which are very important data items. However, it is the set of all rows in the same table that gives the gross wages for the whole company, and it is the (implied) relationship between the EMPLOYEE and DEPARTMENT tables that enables one to get a breakdown of employees by department and/or division. Joins are important because one uses them to tease the relationships out of the database. They are also important because they are very easy to get wrong.
11.2. Sample Views
CREATE VIEW staff_v1 AS
SELECT id
, name
FROM staff
WHERE id BETWEEN 10 AND 30;
CREATE VIEW staff_v2 AS
SELECT id
, job
FROM staff
WHERE id BETWEEN 20 AND 50
UNION ALL
SELECT id
, 'Clerk' AS job
FROM staff
WHERE id = 30;
STAFF_V1
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
STAFF_V2
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
Observe that the above two views have the following characteristics:
-
Both views contain rows that have no corresponding ID in the other view.
-
In the V2 view, there are two rows for ID of 30.
11.3. Join Syntax
Db2 SQL comes with two quite different ways to represent a join. Both syntax styles will be shown throughout this section though, in truth, ne of the styles is usually the better, depending upon the situation.
The first style, which is only really suitable for inner joins, involves listing the tables to be joined in a FROM statement. A comma separates each table name. A subsequent WHERE statement constrains the join.
Here are some sample joins:
SELECT v1.id
, v1.name
, v2.job
FROM staff_v1 v1
, staff_v2 v2
WHERE v1.id = v2.id
ORDER BY v1.id
, v2.job;
JOIN ANSWER
ID | NAME | JOB |
---|---|---|
20 |
Pernal |
Sales |
30 |
Marenghi |
Clerk |
30 |
Marenghi |
Mgr |
SELECT v1.id
, v2.job
, v3.name
FROM staff_v1 v1
, staff_v2 v2
, staff_v1 v3
WHERE v1.id = v2.id AND
v2.id = v3.id AND
v3.name LIKE 'M%'
ORDER BY v1.name
, v2.job;
JOIN ANSWER
ID | JOB | NAME |
---|---|---|
30 |
Clerk |
Marenghi |
30 |
Mgr |
Marenghi |
The second join style, which is suitable for both inner and outer joins, involves joining the tables two at a time, listing the type of join as one goes. ON conditions constrain the join (note: there must be at least one), while WHERE conditions are applied after the join and constrain the result.
The following sample joins are logically equivalent to the two given above:
SELECT v1.id
, v1.name
, v2.job
FROM staff_v1 v1
INNER JOIN staff_v2 v2
ON v1.id = v2.id
ORDER BY v1.id
, v2.job;
JOIN ANSWER
ID | NAME | JOB |
---|---|---|
20 |
Pernal |
Sales |
30 |
Marenghi |
Clerk |
30 |
Marenghi |
Mgr |
SELECT v1.id
, v2.job
, v3.name
FROM staff_v1 v1
JOIN staff_v2 v2
ON v1.id = v2.id
JOIN staff_v1 v3
ON v2.id = v3.id
WHERE v3.name LIKE 'M%'
ORDER BY v1.name
, v2.job;
STAFF_V1
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
JOIN ANSWER
ID | JOB | NAME |
---|---|---|
30 |
Clerk |
Marenghi |
30 |
Mgr |
Marenghi |
STAFF_V2
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
11.3.1. Query Processing Sequence
The following table lists the sequence with which various parts of a query are executed:
FROM clause
JOIN ON clause
WHERE clause
GROUP BY and aggregate
HAVING clause
SELECT list
ORDER BY clause
FETCH FIRST
Observe that ON predicates (e.g. in an outer join) are always processed before any WHERE predicates (in the same join) are applied. Ignoring this processing sequence can cause what looks like an outer join to run as an inner join - see ON and WHERE Usage.
11.3.2. ON vs. WHERE
A join written using the second syntax style shown above can have either, or both, ON and WHERE checks. These two types of check work quite differently:
-
WHERE checks are used to filter rows, and to define the nature of the join. Only those rows that match all WHERE checks are returned.
-
ON checks define the nature of the join. They are used to categorize rows as either joined or not-joined, rather than to exclude rows from the answer-set, though they may do this in some situations.
Let illustrate this difference with a simple, if slightly silly, left outer join:
SELECT *
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON 1 = 1
AND v1.id = v2.id
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
- |
- |
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
Now lets replace the second ON check with a WHERE check:
SELECT *
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON 1 = 1
WHERE v1.id = v2.id
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
In the first example above, all rows were retrieved from the V1 view. Then, for each row, the two ON checks were used to find matching rows in the V2 view. In the second query, all rows were again retrieved from the V1 view. Then each V1 row was joined to every row in the V2 view using the (silly) ON check. Finally, the WHERE check (which is always done after the join) was applied to filter out all pairs that do not match on ID. Can an ON check ever exclude rows? The answer is complicated:
-
In an inner join, an ON check can exclude rows because it is used to define the nature of the join and, by definition, in an inner join only matching rows are returned.
-
In a partial outer join, an ON check on the originating table does not exclude rows. It simply categorizes each row as participating in the join or not.
-
In a partial outer join, an ON check on the table to be joined to can exclude rows because if the row fails the test, it does not match the join.
-
In a full outer join, an ON check never excludes rows. It simply categorizes them as matching the join or not.
-
Each of the above principles will be demonstrated as we look at the different types of join.
11.3.3. Join Types
A generic join matches one row with another to create a new compound row. Joins can be categorized by the nature of the match between the joined rows. In this section we shall discuss each join type and how to code it in SQL.
Inner Join
An inner-join is another name for a standard join in which two sets of columns are joined by matching those rows that have equal data values. Most of the joins that one writes will probably be of this kind and, assuming that suitable indexes have been created, they will almost always be very efficient.
STAFF_V1
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
STAFF_V2
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
Join on ID
INNER-JOIN ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
SELECT *
FROM staff_v1 v1
, staff_v2 v2
WHERE v1.id = v2.id
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
SELECT *
FROM staff_v1 v1
INNER JOIN staff_v2 v2
ON v1.id = v2.id
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
11.3.4. ON and WHERE Usage
In an inner join only, an ON and a WHERE check work much the same way. Both define the nature of the join, and because in an inner join, only matching rows are returned, both act to exclude all rows that do not match the join. Below is an inner join that uses an ON check to exclude managers:
SELECT *
FROM staff_v1 v1
INNER JOIN staff_v2 v2
ON v1.id = v2.id
AND v2.job <> 'Mgr'
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
Here is the same query written using a WHERE check
SELECT *
FROM staff_v1 v1
INNER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v2.job <> 'Mgr'
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
11.3.5. Left Outer Join
A left outer join is the same as saying that I want all of the rows in the first table listed, plus any matching rows in the second table:
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
LEFT-OUTER-JOIN ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
- |
|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
SELECT *
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
ORDER BY 1,4;
It is possible to code a left outer join using the standard inner join syntax (with commas between tables), but it is a lot of work:
SELECT v1.* -- (1)
, v2.*
FROM staff_v1 v1
, staff_v2 v2
WHERE v1.id = v2.id
UNION
SELECT v1.* -- (2)
, CAST(NULL AS SMALLINT) AS id
, CAST(NULL AS CHAR(5)) AS job
FROM staff_v1 v1
WHERE v1.id NOT IN
(SELECT id FROM staff_v2)
ORDER BY 1,4;
(1) This join gets all rows in STAFF_V1 that match rows in STAFF_V2. (2) This query gets all the rows in STAFF_V1 with no matching rows in STAFF_V2.
11.3.6. ON and WHERE Usage
In any type of join, a WHERE check works as if the join is an inner join. If no row matches, then no row is returned, regardless of what table the predicate refers to. By contrast, in a left or right outer join, an ON check works differently, depending on what table field it refers to:
-
If it refers to a field in the table being joined to, it determines whether the related row matches the join or not.
-
If it refers to a field in the table being joined from, it determines whether the related row finds a match or not. Regardless, the row will be returned.
In the next example, those rows in the table being joined to (i.e. the V2 view) that match on ID, and that are not for a manager are joined to:
SELECT *
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
AND v2.job <> 'Mgr'
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
If we rewrite the above query using a WHERE check we will lose a row (of output) because the check is applied after the join is done, and a null JOB does not match:
SELECT *
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v2.job <> 'Mgr'
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
We could make the WHERE equivalent to the ON, if we also checked for nulls:
SELECT *
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE (v2.job <> 'Mgr' OR
v2.job IS NULL)
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
In the next example, those rows in the table being joined from (i.e. the V1 view) that match on ID and have a NAME > 'N' participate in the join. Note however that V1 rows that do not participate in the join (i.e. ID = 30) are still returned:
SELECT *
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
AND v1.name > 'N'
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
- |
- |
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
- |
- |
If we rewrite the above query using a WHERE check (on NAME) we will lose a row because now the check excludes rows from the answer-set, rather than from participating in the join:
SELECT *
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v1.name > 'N'
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
Unlike in the previous example, there is no way to alter the above WHERE check to make it logically equivalent to the prior ON check. The ON and the WHERE are applied at different times and for different purposes, and thus do completely different things.
11.3.7. Right Outer Join
A right outer join is the inverse of a left outer join. One gets every row in the second table listed, plus any matching rows in the first table:
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
RIGHT-OUTER-JOIN ANSWER .Example of Right Outer Join
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
40 |
Sales |
||
50 |
Mgr |
SELECT *
FROM staff_v1 v1
RIGHT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
ORDER BY v2.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
- |
40 |
Sales |
|
- |
50 |
Mgr |
It is also possible to code a right outer join using the standard inner join syntax:
SELECT v1.*
, v2.*
FROM staff_v1 v1
, staff_v2 v2
WHERE v1.id = v2.id
UNION
SELECT CAST(NULL AS SMALLINT) AS id
, CAST(NULL AS VARCHAR(9)) AS name
, v2.*
FROM staff_v2 v2
WHERE v2.id NOT IN
(SELECT id FROM staff_v1)
ORDER BY 3, 4;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
40 |
Sales |
||
50 |
Mgr |
ON and WHERE Usage
The rules for ON and WHERE usage are the same in a right outer join as they are for a left outer join (see Left Outer Join), except that the relevant tables are reversed.
11.3.8. Full Outer Joins
A full outer join occurs when all of the matching rows in two tables are joined, and there is also returned one copy of each non-matching row in both tables.
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
FULL-OUTER-JOIN ANSWER .Example of Full Outer Join
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
40 |
Sales |
||
50 |
Mgr |
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
ORDER BY v1.id
, v2.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
´ |
40 |
Sales |
|
50 |
Mgr |
Here is the same done using the standard inner join syntax:
SELECT v1.*
, v2.*
FROM staff_v1 v1
, staff_v2 v2
WHERE v1.id = v2.id
UNION
SELECT v1.*
, CAST(NULL AS SMALLINT) AS id
, CAST(NULL AS CHAR(5)) AS job
FROM staff_v1 v1
WHERE v1.id NOT IN
(SELECT id FROM staff_v2)
UNION
SELECT CAST(NULL AS SMALLINT) AS id
, CAST(NULL AS VARCHAR(9)) AS name
, v2.*
FROM staff_v2 v2
WHERE v2.id NOT IN
(SELECT id FROM staff_v1)
ORDER BY 1,3,4;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
40 |
Sales |
||
50 |
Mgr |
The above is reasonably hard to understand when two tables are involved, and it goes down hill fast as more tables are joined. Avoid.
ON and WHERE Usage
In a full outer join, an ON check is quite unlike a WHERE check in that it never results in a row being excluded from the answer set. All it does is categorize the input row as being either matching or non-matching. For example, in the following full outer join, the ON check joins those rows with equal key values:
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
ORDER BY v1.id
, v2.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
40 |
Sales |
||
50 |
Mgr |
In the next example, we have deemed that only those IDs that match, and that also have a value greater than 20, are a true match:
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
AND v1.id > 20
ORDER BY v1.id
, v2.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
||
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
20 |
Sales |
||
40 |
Sales |
||
50 |
Mgr |
Observe how in the above statement we added a predicate, and we got more rows! This is because in an outer join an ON predicate never removes rows. It simply categorizes them as being either matching or non-matching. If they match, it joins them. If they don’t, it passes them through.
In the next example, nothing matches. Consequently, every row is returned individually. This query is logically similar to doing a UNION ALL on the two views:
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
AND +1 = -1
ORDER BY v1.id
, v2.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
||
30 |
Marenghi |
||
20 |
Sales |
||
30 |
Clerk |
||
30 |
Mgr |
||
40 |
Sales |
||
50 |
Mgr |
ON checks are somewhat like WHERE checks in that they have two purposes. Within a table, they are used to categorize rows as being either matching or non-matching. Between tables, they are used to define the fields that are to be joined on. In the prior example, the first ON check defined the fields to join on, while the second join identified those fields that matched the join. Because nothing matched (due to the second predicate), everything fell into the "outer join" category. This means that we can remove the first ON check without altering the answer set:
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON +1 = -1
ORDER BY v1.id
, v2.id
, v2.job;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
||
30 |
Marenghi |
||
20 |
Sales |
||
30 |
Clerk |
||
30 |
Mgr |
||
40 |
Sales |
||
50 |
Mgr |
What happens if everything matches and we don’t identify the join fields? The result in a Cartesian Product:
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON +1 <> -1
ORDER BY v1.id
, v2.id
, v2.job;
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
ANSWER .Full Outer Join, don’t match on keys (all rows match)
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
20 |
Sales |
10 |
Sanders |
30 |
Clerk |
10 |
Sanders |
30 |
Mgr |
10 |
Sanders |
40 |
Sales |
10 |
Sanders |
50 |
Mgr |
20 |
Pernal |
20 |
Sales |
20 |
Pernal |
30 |
Clerk |
20 |
Pernal |
30 |
Mgr |
20 |
Pernal |
40 |
Sales |
20 |
Pernal |
50 |
Mgr |
30 |
Marenghi |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
30 |
Marenghi |
40 |
Sales |
30 |
Marenghi |
50 |
Mgr |
In an outer join, WHERE predicates behave as if they were written for an inner join. In particular, they always do the following:
-
WHERE predicates defining join fields enforce an inner join on those fields.
-
WHERE predicates on non-join fields are applied after the join, which means that when they are used on not-null fields, they negate the outer join.
Here is an example of a WHERE join predicate turning an outer join into an inner join:
SELECT *
FROM staff_v1 v1
FULL JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v1.id = v2.id
ORDER BY 1,3,4;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
To illustrate some of the complications that WHERE checks can cause, imagine that we want to do a FULL OUTER JOIN on our two test views (see below), limiting the answer to those rows where the "V1 ID" field is less than 30. There are several ways to express this query, each giving a different answer:
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
OUTER-JOIN CRITERIA .Outer join V1.ID < 30, sample data
V1.ID = V2.ID V1.ID < 30 ANSWER ???, *_DEPENDS_*
In our first example, the "V1.ID < 30" predicate is applied after the join, which effectively eliminates all "V2" rows that don’t match (because their "V1.ID" value is null):
SELECT *
FROM staff_v1 v1
FULL JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v1.id < 30
ORDER BY 1,3,4;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
In the next example the "V1.ID < 30" check is done during the outer join where it does not any eliminate rows, but rather limits those that match in the two views:
SELECT *
FROM staff_v1 v1
FULL JOIN staff_v2 v2
ON v1.id = v2.id
AND v1.id < 30
ORDER BY 1,3,4;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
||
30 |
Clerk |
||
30 |
Mgr |
||
40 |
Sales |
||
50 |
Mgr |
Imagine that what really wanted to have the "V1.ID < 30" check to only apply to those rows in the "V1" table. Then one has to apply the check before the join, which requires the use of a nested-table expression:
SELECT *
FROM (SELECT *
FROM staff_v1
WHERE id < 30) AS v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
ORDER BY 1,3,4;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Clerk |
||
30 |
Mgr |
||
40 |
Sales |
||
50 |
Mgr |
Observe how in the above query we still got a row back with an ID of 30, but it came from the "V2" table. This makes sense, because the WHERE condition had been applied before we got to this table. There are several incorrect ways to answer the above question. In the first example, we shall keep all non-matching V2 rows by allowing to pass any null V1.ID values:
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v1.id < 30
OR v1.id IS NULL
ORDER BY 1,3,4;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
40 |
Sales |
||
50 |
Mgr |
There are two problems with the above query: First, it is only appropriate to use when the V1.ID field is defined as not null, which it is in this case. Second, we lost the row in the V2 table where the ID equaled 30. We can fix this latter problem, by adding another check, but the answer is still wrong:
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v1.id < 30
OR v1.id = v2.id
OR v1.id IS NULL
ORDER BY 1,3,4;
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
40 |
Sales |
||
50 |
Mgr |
The last two checks in the above query ensure that every V2 row is returned. But they also have the affect of returning the NAME field from the V1 table whenever there is a match. Given our intentions, this should not happen.
SUMMARY: Query WHERE conditions are applied after the join. When used in an outer join, this means that they applied to all rows from all tables. In effect, this means that any WHERE conditions in a full outer join will, in most cases, turn it into a form of inner join.
11.3.9. Cartesian Product
A Cartesian Product is a form of inner join, where the join predicates either do not exist, or where they do a poor job of matching the keys in the joined tables.
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
20 |
Sales |
10 |
Sanders |
30 |
Clerk |
10 |
Sanders |
30 |
Mgr |
10 |
Sanders |
40 |
Sales |
10 |
Sanders |
50 |
Mgr |
20 |
Pernal |
20 |
Sales |
20 |
Pernal |
30 |
Clerk |
20 |
Pernal |
30 |
Mgr |
20 |
Pernal |
40 |
Sales |
20 |
Pernal |
50 |
Mgr |
30 |
Marenghi |
20 |
Sales |
30 |
Marenghi |
30 |
Clerk |
30 |
Marenghi |
30 |
Mgr |
30 |
Marenghi |
40 |
Sales |
30 |
Marenghi |
50 |
Mgr |
Writing a Cartesian Product is simplicity itself. One simply omits the WHERE conditions:
SELECT *
FROM staff_v1 v1
, staff_v2 v2
ORDER BY v1.id
, v2.id
, v2.job;
One way to reduce the likelihood of writing a full Cartesian Product is to always use the inner/outer join style. With this syntax, an ON predicate is always required. There is however no guarantee that the ON will do any good. Witness the following example:
SELECT *
FROM staff_v1 v1
INNER JOIN staff_v2 v2
ON 'A' <> 'B'
ORDER BY v1.id
, v2.id
, v2.job;
A Cartesian Product is almost always the wrong result. There are very few business situations where it makes sense to use the kind of SQL shown above. The good news is that few people ever make the mistake of writing the above. But partial Cartesian Products are very common, and they are also almost always incorrect. Here is an example:
SELECT v2a.id
, v2a.job
, v2b.id
FROM staff_v2 v2a
, staff_v2 v2b
WHERE v2a.job = v2b.job
AND v2a.id < 40
ORDER BY v2a.id
, v2b.id;
ANSWER
ID | JOB | ID |
---|---|---|
20 |
Sales |
20 |
20 |
Sales |
40 |
30 |
Clerk |
30 |
30 |
Mgr |
30 |
30 |
Mgr |
50 |
In the above example we joined the two views by JOB, which is not a unique key. The result was that for each JOB value, we got a mini Cartesian Product.
Cartesian Products are at their most insidious when the result of the (invalid) join is feed into a GROUP BY or DISTINCT statement that removes all of the duplicate rows. Below is an example where the only clue that things are wrong is that the count is incorrect:
SELECT v2.job
, COUNT(*) AS #rows
FROM staff_v1 v1
, staff_v2 v2
GROUP BY v2.job
ORDER BY #rows
, v2.job;
ANSWER
JOB | #ROWS |
---|---|
Clerk |
3 |
Mgr |
6 |
Sales |
6 |
To really mess up with a Cartesian Product you may have to join more than one table. Note however that big tables are not required. For example, a Cartesian Product of five 100-row tables will result in 10,000,000,000 rows being returned.
A good rule of thumb to use when writing a join is that for all of the tables (except one) there should be equal conditions on all of the fields that make up the various unique keys. If this is not true then it is probable that some kind Cartesian Product is being done and the answer may be wrong. |
11.3.10. Join Notes
Using the COALESCE Function
If you don’t like working with nulls, but you need to do outer joins, then life is tough. In an outer join, fields in non-matching rows are given null values as placeholders. Fortunately, these nulls can be eliminated using the COALESCE function. The COALESCE function can be used to combine multiple fields into one, and/or to eliminate null values where they occur. The result of the COALESCE is always the first non-null value encountered. In the following example, the two ID fields are combined, and any null NAME values are replaced with a question mark.
SELECT COALESCE(v1.id, v2.id) AS id
, COALESCE(v1.name, '?') AS name
, v2.job
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | JOB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
30 |
Marenghi |
Clerk |
30 |
Marenghi |
Mgr |
40 |
? |
Sales |
50 |
? |
Mgr |
Listing non-matching rows only
Imagine that we wanted to do an outer join on our two test views, only getting those rows that do not match. This is a surprisingly hard query to write.
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
NON-MATCHING OUTER-JOIN ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
40 |
Sales |
||
50 |
Mgr |
One way to express the above is to use the standard inner-join syntax:
SELECT v1.* -- (1)
, CAST(NULL AS SMALLINT) AS id
, CAST(NULL AS CHAR(5)) AS job
FROM staff_v1 v1
WHERE v1.id NOT IN
(SELECT id FROM staff_v2)
UNION
SELECT CAST(NULL AS SMALLINT) AS id -- (2)
, CAST(NULL AS VARCHAR(9)) AS name
, v2.*
FROM staff_v2 v2
WHERE v2.id NOT IN
(SELECT id FROM staff_v1)
ORDER BY 1,3,4;
(1) Get all the rows in STAFF_V1 that have no matching row in STAFF_V2.
(2) Get all the rows in STAFF_V2 that have no matching row in STAFF_V1.
The above question can also be expressed using the outer-join syntax, but it requires the use of two nested-table expressions. These are used to assign a label field to each table. Only those rows where either of the two labels are null are returned:
SELECT *
FROM (SELECT v1.*
, 'V1' AS flag
FROM staff_v1 v1) AS v1
FULL OUTER JOIN
(SELECT v2.*
, 'V2' AS flag
FROM staff_v2 v2) AS v2
ON v1.id = v2.id
WHERE v1.flag IS NULL
OR v2.flag IS NULL
ORDER BY v1.id
, v2.id
, v2.job;
ANSWER
ID | NAME | FLAG | ID | JOB | FLAG |
---|---|---|---|---|---|
10 |
Sanders |
V1 |
|||
40 |
Sales |
V2 |
|||
50 |
Mgr |
V2 |
Alternatively, one can use two common table expressions to do the same job:
WITH v1 AS
(SELECT v1.*
, 'V1' AS flag
FROM staff_v1 v1)
, v2 AS
(SELECT v2.*
, 'V2' AS flag
FROM staff_v2 v2)
SELECT *
FROM v1 v1
FULL OUTER JOIN v2 v2
ON v1.id = v2.id
WHERE v1.flag IS NULL
OR v2.flag IS NULL
ORDER BY v1.id
, v2.id
, v2.job;
ANSWER
ID | NAME | FLAG | ID | JOB | FLAG |
---|---|---|---|---|---|
10 |
Sanders |
V1 |
|||
40 |
Sales |
V2 |
|||
50 |
Mgr |
V2 |
If either or both of the input tables have a field that is defined as not null, then label fields can be discarded. For example, in our test tables, the two ID fields will suffice:
SELECT *
FROM staff_v1 v1
FULL OUTER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v1.id IS NULL
OR v2.id IS NULL
ORDER BY v1.id
, v2.id
, v2.job;
Join in SELECT Phrase
Imagine that we want to get selected rows from the V1 view, and for each matching row, get the corresponding JOB from the V2 view - if there is one:
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
V1.ID = V2.ID V1.ID <> 30
ANSWER
ID | NAME | ID | JOB |
---|---|---|---|
10 |
Sanders |
||
20 |
Pernal |
20 |
Sales |
Here is one way to express the above as a query:
SELECT v1.id
, v1.name
, v2.job
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v1.id <> 30
ORDER BY v1.id ;
ANSWER
ID | NAME | JOB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
Below is a logically equivalent left outer join with the join placed in the SELECT phrase of the SQL statement. In this query, for each matching row in STAFF_V1, the join (i.e. the nested table expression) will be done:
SELECT v1.id
, v1.name
, (SELECT v2.job
FROM staff_v2 v2
WHERE v1.id = v2.id) AS jb
FROM staff_v1 v1
WHERE v1.id <> 30
ORDER BY v1.id;
ANSWER
ID | NAME | JB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
Certain rules apply when using the above syntax:
-
The nested table expression in the SELECT is applied after all other joins and sub-queries (i.e. in the FROM section of the query) are done.
-
The nested table expression acts as a left outer join.
-
Only one column and row (at most) can be returned by the expression.
-
If no row is returned, the result is null.
Given the above restrictions, the following query will fail because more than one V2 row is returned for every V1 row (for ID = 30):
SELECT v1.id
, v1.name
, (SELECT v2.job
FROM staff_v2 v2
WHERE v1.id = v2.id) AS jb
FROM staff_v1 v1
ORDER BY v1.id;
ANSWER
ID | NAME | JB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
To make the above query work for all IDs, we have to decide which of the two matching JOB values for ID 30 we want. Let us assume that we want the maximum:
SELECT v1.id
, v1.name
, (SELECT MAX(v2.job)
FROM staff_v2 v2
WHERE v1.id = v2.id) AS jb
FROM staff_v1 v1
ORDER BY v1.id;
ANSWER
ID | NAME | JB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
30 |
Marenghi |
Mgr |
The above is equivalent to the following query:
SELECT v1.id
, v1.name
, MAX(v2.job) AS jb
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
GROUP BY v1.id
, v1.name
ORDER BY v1.id ;
ANSWER
ID | NAME | JB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
30 |
Marenghi |
Mgr |
The above query is rather misleading because someone unfamiliar with the data may not understand why the NAME field is in the GROUP BY. Obviously, it is not there to remove any rows, it simply needs to be there because of the presence of the MAX function. Therefore, the preceding query is better because it is much easier to understand. It is also probably more efficient.
CASE Usage
The SELECT expression can be placed in a CASE statement if needed. To illustrate, in the following query we get the JOB from the V2 view, except when the person is a manager, in which case we get the NAME from the corresponding row in the V1 view:
SELECT v2.id
, CASE
WHEN v2.job <> 'Mgr' THEN v2.job
ELSE (SELECT v1.name
FROM staff_v1 v1
WHERE v1.id = v2.id)
END AS j2
FROM staff_v2 v2
ORDER BY v2.id
, j2;
ANSWER
ID | J2 |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Marenghi |
40 |
Sales |
50 |
- |
Multiple Columns
If you want to retrieve two columns using this type of join, you need to have two independent nested table expressions:
SELECT v2.id
, v2.job
, (SELECT v1.name
FROM staff_v1 v1
WHERE v2.id = v1.id)
, (SELECT LENGTH(v1.name) AS n2
FROM staff_v1 v1
WHERE v2.id = v1.id)
FROM staff_v2 v2
ORDER BY v2.id
, v2.job;
ANSWER
ID | JOB | NAME | N2 |
---|---|---|---|
20 |
Sales |
Pernal |
6 |
30 |
Clerk |
Marenghi |
8 |
30 |
Mgr |
Marenghi |
8 |
40 |
Sales |
||
50 |
Mgr |
An easier way to do the above is to write an ordinary left outer join with the joined columns in the SELECT list. To illustrate this, the next query is logically equivalent to the prior:
SELECT v2.id
, v2.job
, v1.name
, LENGTH(v1.name) AS n2
FROM staff_v2 v2
LEFT OUTER JOIN staff_v1 v1
ON v2.id = v1.id
ORDER BY v2.id
, v2.job;
ANSWER
ID | JOB | NAME | N2 |
---|---|---|---|
20 |
Sales |
Pernal |
6 |
30 |
Clerk |
Marenghi |
8 |
30 |
Mgr |
Marenghi |
8 |
40 |
Sales |
||
50 |
Mgr |
Column Functions
This join style lets one easily mix and match individual rows with the results of column functions. For example, the following query returns a running SUM of the ID column:
SELECT v1.id
, v1.name
, (SELECT SUM(x1.id)
FROM staff_v1 x1
WHERE x1.id <= v1.id
)AS sum_id
FROM staff_v1 v1
ORDER BY v1.id
, v2.job;
ID | NAME | SUM_ID |
---|---|---|
10 |
Sanders |
10 |
20 |
Pernal |
30 |
30 |
Marenghi |
60 |
An easier way to do the same as the above is to use an OLAP function:
SELECT v1.id
, v1.name
, SUM(id) OVER(ORDER BY id) AS sum_id
FROM staff_v1 v1
ORDER BY v1.id;
ANSWER
ID | NAME | SUM_ID |
---|---|---|
10 |
Sanders |
10 |
20 |
Pernal |
30 |
30 |
Marenghi |
60 |
Predicates and Joins, a Lesson
Imagine that one wants to get all of the rows in STAFF_V1, and to also join those matching rows in STAFF_V2 where the JOB begins with an 'S':
_STAFF_V1_
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
30 |
Marenghi |
_STAFF_V2_
ID | JOB |
---|---|
20 |
Sales |
30 |
Clerk |
30 |
Mgr |
40 |
Sales |
50 |
Mgr |
OUTER-JOIN CRITERIA
V1.ID = V2.ID
V2.JOB LIKE 'S%'
ANSWER
ID | NAME | JOB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
30 |
Marenghi |
The first query below gives the wrong answer. It is wrong because the WHERE is applied after the join, so eliminating some of the rows in the STAFF_V1 table:
SELECT v1.id
, v1.name
, v2.job
FROM staff_v1 v1
LEFT OUTER JOIN staff_v2 v2
ON v1.id = v2.id
WHERE v2.job LIKE 'S%'
ORDER BY v1.id
, v2.job;
ANSWER (WRONG)
ID | NAME | JOB |
---|---|---|
20 |
Pernal |
Sales |
In the next query, the WHERE is moved into a nested table expression - so it is done before the join (and against STAFF_V2 only), thus giving the correct answer:
SELECT v1.id
, v1.name
, v2.job
FROM staff_v1 v1
LEFT OUTER JOIN
(SELECT *
FROM staff_v2
WHERE job LIKE 'S%'
) AS v2
ON v1.id = v2.id
ORDER BY v1.id
, v2.job;
ANSWER
ID | NAME | JOB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
30 |
Marenghi |
The next query does the join in the SELECT phrase. In this case, whatever predicates are in the nested table expression apply to STAFF_V2 only, so we get the correct answer:
SELECT v1.id
, v1.name
, (SELECT v2.job
FROM staff_v2 v2
WHERE v1.id = v2.id
AND v2.job LIKE 'S%')
FROM staff_v1 v1
ORDER BY v1.id
, job;
ANSWER
ID | NAME | JOB |
---|---|---|
10 |
Sanders |
|
20 |
Pernal |
Sales |
30 |
Marenghi |
Joins - Things to Remember
-
You get nulls in an outer join, whether you want them or not, because the fields in nonmatching rows are set to null. If they bug you, use the COALESCE function to remove them. See Using the COALESCE Function for an example.
-
From a logical perspective, all WHERE conditions are applied after the join. For performance reasons, Db2 may apply some checks before the join, especially in an inner join, where doing this cannot affect the result set.
-
All WHERE conditions that join tables act as if they are doing an inner join, even when they are written in an outer join.
-
The ON checks in a full outer join never remove rows. They simply determine what rows are matching versus not (see ON and WHERE Usage). To eliminate rows in an outer join, one must use a WHERE condition.
-
The ON checks in a partial outer join work differently, depending on whether they are against fields in the table being joined to, or joined from (ON and WHERE Usage).
-
A Cartesian Product is not an outer join. It is a poorly matching inner join. By contrast, a true outer join gets both matching rows, and non-matching rows.
-
The NODENUMBER and PARTITION functions cannot be used in an outer join. These functions only work on rows in real tables.
-
When the join is defined in the SELECT part of the query (see ON and WHERE Usage), it is done after any other joins and/or sub-queries specified in the FROM phrase. And it acts as if it is a left outer join.
11.3.11. Complex Joins
When one joins multiple tables using an outer join, one must consider carefully what exactly what one wants to do, because the answer that one gets will depend upon how one writes the query. To illustrate, the following query first gets a set of rows from the employee table, and then joins (from the employee table) to both the activity and photo tables:
SELECT eee.empno
, aaa.projno
, aaa.actno
, ppp.photo_format AS format
FROM employee eee
LEFT OUTER JOIN emp_act aaa
ON eee.empno = aaa.empno
AND aaa.emptime = 1
AND aaa.projno LIKE 'M%1%'
LEFT OUTER JOIN emp_photo ppp
ON eee.empno = ppp.empno
AND ppp.photo_format LIKE 'b%'
WHERE eee.lastname LIKE '%A%'
AND eee.empno < '000170'
AND eee.empno <> '000030'
ORDER BY eee.empno;
ANSWER
EMPN O | PROJNO | ACTNO | FORMAT |
---|---|---|---|
000010 |
MA2110 |
10 |
- |
000070 |
- |
- |
- |
000130 |
- |
- |
bitmap |
000150 |
MA2112 |
60 |
bitmap |
000150 |
MA2112 |
180 |
bitmap |
000160 |
MA2113 |
60 |
- |
Observe that we got photo data, even when there was no activity data. This is because both tables were joined directly from the employee table. In the next query, we will again start at the employee table, then join to the activity table, and then from the activity table join to the photo table. We will not get any photo data, if the employee has no activity:
SELECT eee.empno
, aaa.projno
, aaa.actno
, ppp.photo_format AS format
FROM employee eee
LEFT OUTER JOIN emp_act aaa
ON eee.empno = aaa.empno
AND aaa.emptime = 1
AND aaa.projno LIKE 'M%1%'
LEFT OUTER JOIN emp_photo ppp
ON aaa.empno = ppp.empno
AND ppp.photo_format LIKE 'b%'
WHERE eee.lastname LIKE '%A%'
AND eee.empno < '000170'
AND eee.empno <> '000030'
ORDER BY eee.empno;
ANSWER
EMPNO | PROJNO | ACTNO | FORMAT |
---|---|---|---|
000010 |
MA2110 |
10 |
|
000070 |
- |
||
000130 |
|||
000150 |
MA2112 |
60 |
bitmap |
000150 |
MA2112 |
180 |
bitmap |
000160 |
MA2113 |
60 |
- |
The only difference between the above two queries is the first line of the second ON.
Outer Join followed by Inner Join
Mixing and matching inner and outer joins in the same query can cause one to get the wrong answer. To illustrate, the next query has an outer join, followed by an inner join. We are trying to do the following:
-
Get a list of matching employees - based on some local predicates.
-
For each employee found, list their matching activities, if any (i.e. left outer join).
-
For each activity found, only list it if its project-name contains the letter "Q" (i.e. inner join between activity and project).
Below is the wrong way to write this query. It is wrong because the final inner join (between activity and project) turns the preceding outer join into an inner join. This causes an employee to not show when there are no matching projects:
SELECT eee.workdept AS dp#
, eee.empno
, aaa.projno
, ppp.prstaff AS staff
FROM
(SELECT *
FROM employee
WHERE lastname LIKE '%A%'
AND job <> 'DESIGNER'
AND workdept BETWEEN 'B' AND 'E'
) AS eee
LEFT OUTER JOIN emp_act aaa
ON aaa.empno = eee.empno
AND aaa.emptime <= 0.5
INNER JOIN project ppp
ON aaa.projno = ppp.projno
AND ppp.projname LIKE '%Q%'
ORDER BY eee.workdept
, eee.empno
, aaa.projno;
DP# | EMPNO | PROJNO | STAFF |
---|---|---|---|
C01 |
000030 |
IF1000 |
2.00 |
C01 |
000130 |
IF1000 |
2.00 |
As was stated above, we really want to get all matching employees, and their related activities (projects). If an employee has no matching activates, we still want to see the employee.
The next query gets the correct answer by putting the inner join between the activity and project tables in parenthesis, and then doing an outer join to the combined result:
SELECT eee.workdept AS dp#
, eee.empno
, xxx.projno
, xxx.prstaff AS staff
FROM
(SELECT *
FROM employee
WHERE lastname LIKE '%A%'
AND job <> 'DESIGNER'
AND workdept BETWEEN 'B' AND 'E'
) AS eee
LEFT OUTER JOIN
(SELECT aaa.empno
, aaa.emptime
, aaa.projno
, ppp.prstaff
FROM emp_act aaa
INNER JOIN project ppp
ON aaa.projno = ppp.projno
AND ppp.projname LIKE '%Q%'
) AS xxx
ON xxx.empno = eee.empno
AND xxx.emptime <= 0.5
ORDER BY eee.workdept
, eee.empno
, xxx.projno;
DP# | EMPNO | PROJNO | STAFF |
---|---|---|---|
C01 |
000030 |
IF1000 |
2.00 |
C01 |
000130 |
IF1000 |
2.00 |
D21 |
000070 |
||
D21 |
000240 |
The lesson to be learnt here is that if a subsequent inner join acts upon data in a preceding outer join, then it, in effect, turns the former into an inner join.
11.3.12. Simplified Nested Table Expression
The next query is the same as the prior, except that the nested-table expression has no select list, nor correlation name. In this example, any columns in tables that are inside of the nested table expression are referenced directly in the rest of the query:
SELECT eee.workdept AS dp#
, eee.empno
, aaa.projno
, ppp.prstaff AS staff
FROM
(SELECT *
FROM employee
WHERE lastname LIKE '%A%'
AND job <> 'DESIGNER'
AND workdept BETWEEN 'B' AND 'E'
) AS eee
LEFT OUTER JOIN
(SELECT aaa.empno
, aaa.emptime
, aaa.projno
, ppp.prstaff
FROM emp_act aaa
INNER JOIN project ppp
ON aaa.projno = ppp.projno
AND ppp.projname LIKE '%Q%'
) AS xxx
ON xxx.empno = eee.empno
AND xxx.emptime <= 0.5
ORDER BY eee.workdept
, eee.empno
, xxx.projno;
DP# | EMPNO | PROJNO | STAFF |
---|---|---|---|
C01 |
000030 |
IF1000 |
2.00 |
C01 |
000130 |
IF1000 |
2.00 |
D21 |
000070 |
||
D21 |
000240 |
12. Sub-Query
Sub-queries are hard to use, tricky to tune, and often do some strange things. Consequently, a lot of people try to avoid them, but this is stupid because sub-queries are really, really, useful. Using a relational database and not writing sub-queries is almost as bad as not doing joins. A sub-query is a special type of fullselect that is used to relate one table to another without actually doing a join. For example, it lets one select all of the rows in one table where some related value exists, or does not exist, in another table.
12.1. Sample Tables
Two tables will be used in this section. Please note that the second sample table has a mixture of null and not-null values:
CREATE TABLE table1
( t1a CHAR(1) NOT NULL
, t1b CHAR(2) NOT NULL
, PRIMARY KEY(t1a));
COMMIT;
CREATE TABLE table2
( t2a CHAR(1) NOT NULL
, t2b CHAR(1) NOT NULL
, t2c CHAR(1));
INSERT INTO table1 VALUES ('A', 'AA')
, ('B','BB')
, ('C','CC');
INSERT INTO table2 VALUES ('A', 'A', 'A')
, ('B', 'A', NULL);
TABLE1
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
TABLE2
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
B |
A |
- |
"-" = null
A sub-query compares an expression against a fullselect. The type of comparison done is a function of which, if any, keyword is used.
The result of doing a sub-query check can be any one of the following:
-
True, in which case the current row being processed is returned.
-
False, in which case the current row being processed is rejected.
-
Unknown, which is functionally equivalent to false.
-
A SQL error, due to an invalid comparison.
12.2. No Keyword Sub-Query
One does not have to provide a SOME, or ANY, or IN, or any other keyword, when writing a sub-query. But if one does not, there are three possible results:
-
If no row in the sub-query result matches, the answer is false.
-
If one row in the sub-query result matches, the answer is true.
-
If more than one row in the sub-query result matches, you get a SQL error.
In the example below, the T1A field in TABLE1 is checked to see if it equals the result of the sub-query (against T2A in TABLE2). For the value "A" there is a match, while for the values "B" and "C" there is no match:
SELECT *
FROM table1
WHERE t1a =
(SELECT t2a
FROM table2
WHERE t2a = 'A');
ANSWER
T1A | T1B |
---|---|
A |
AA |
SUB-Q RESLT
T2A |
---|
A |
TABLE1
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
TABLE2
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
B |
A |
- |
"-" = null
The next example gets a SQL error. The sub-query returns two rows, which the "=l" check cannot process. Had an "= ANY" or an "= SOME" check been used instead, the query would have worked fine:
SELECT *
FROM table1
WHERE t1a =
(SELECT t2a
FROM table2);
ANSWER =⇒ error
SUB-Q RESLT
T2A |
---|
A |
B |
TABLE1
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
TABLE2
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
B |
A |
- |
"-" = null
There is almost never a valid reason for coding a sub-query that does not use an appropriate sub-query keyword. Do not do the above. |
12.3. SOME/ANY Keyword Sub-Query
When a SOME or ANY sub-query check is used, there are two possible results:
-
If any row in the sub-query result matches, the answer is true.
-
If the sub-query result is empty, or all nulls, the answer is false.
-
If no value found in the sub-query result matches, the answer is also false.
-
The query below compares the current T1A value against the sub-query result three times.
The first row (i.e. T1A = "A") fails the test, while the next two rows pass:
SELECT *
FROM table1
WHERE t1a > ANY
(SELECT t2a
FROM table2);
ANSWER
T1A | T1B |
---|---|
B |
BB |
C |
CC |
SUB-Q RESLT
T2A |
---|
A |
B |
TABLE1
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
TABLE2
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
B |
A |
- |
"-" = null
When an ANY or ALL sub-query check is used with a "greater than" or similar expression (as opposed to an "equal" or a "not equal" expression) then the check can be considered similar to evaluating the MIN or the MAX of the sub-query result set. The following table shows what type of sub-query check equates to what type of column function: .ANY and ALL vs. column functions
SUB-QUERY CHECK | EQUIVALENT COLUMN FUNCTION |
---|---|
> ANY(sub-query) |
> MINIMUM(sub-query results) |
< ANY(sub-query) |
< MAXIMUM(sub-query results) |
> ALL(sub-query) |
> MAXIMUM(sub-query results) |
< ALL(sub-query) |
< MINIMUM(sub-query results) |
12.3.1. All Keyword Sub-Query
When an ALL sub-query check is used, there are two possible results:
-
If all rows in the sub-query result match, the answer is true.
-
If there are no rows in the sub-query result, the answer is also true.
-
If any row in the sub-query result does not match, or is null, the answer is false.
Below is a typical example of the ALL check usage. Observe that a TABLE1 row is returned only if the current T1A value equals all of the rows in the sub-query result:
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2b
FROM table2
WHERE t2b >= 'A');
ANSWER
T1A | T1B |
---|---|
A |
AA |
SUB-Q RESLT
T2B |
---|
A |
A |
When the sub-query result consists of zero rows (i.e. an empty set) then all rows processed in TABLE1 are deemed to match:
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2b
FROM table2
WHERE t2b >= 'X');
ANSWER
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
SUB-Q RESLT
T2B |
---|
The above may seem a little unintuitive, but it actually makes sense, and is in accordance with how the NOT EXISTS sub-query (see [sub.query]) handles a similar situation.
Imagine that one wanted to get a row from TABLE1 where the T1A value matched all of the sub-query result rows, but if the latter was an empty set (i.e. no rows), one wanted to get a non-match. Try this:
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2b
FROM table2
WHERE t2b >= 'X')
AND 0 <>
(SELECT COUNT(*)
FROM table2
WHERE t2b >= 'X');
ANSWER =⇒ 0 rows
SQ-#1 RESLT
T2B |
---|
SQ-#2 RESLT
(*) |
---|
0 |
TABLE1
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
TABLE2
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
B |
A |
- |
"-" = null
Two sub-queries are done above: The first looks to see if all matching values in the sub-query equal the current T1A value. The second confirms that the number of matching values in the sub-query is not zero.
Observe that the ANY sub-query check returns false when used against an empty set, while a similar ALL check returns true. |
12.4. EXISTS Keyword Sub-Query
So far, we have been taking a value from the TABLE1 table and comparing it against one or more rows in the TABLE2 table. The EXISTS phrase does not compare values against rows, rather it simply looks for the existence or non-existence of rows in the sub-query result set:
-
If the sub-query matches on one or more rows, the result is true.
-
If the sub-query matches on no rows, the result is false.
Below is an EXISTS check that, given our sample data, always returns true:
SELECT *
FROM table1
WHERE EXISTS
(SELECT *
FROM table2);
ANSWER
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
Below is an EXISTS check that, given our sample data, always returns false:
SELECT *
FROM table1
WHERE EXISTS
(SELECT *
FROM table2
WHERE t2b >= 'X');
ANSWER =⇒ 0 rows
When using an EXISTS check, it doesn’t matter what field, if any, is selected in the sub-query SELECT phrase. What is important is whether the sub-query returns a row or not. If it does, the sub-query returns true. Having said this, the next query is an example of an EXISTS subquery that will always return true, because even when no matching rows are found in the subquery, the SELECT COUNT(*) statement will return something (i.e. a zero). Arguably, this query is logically flawed:
SELECT *
FROM table1
WHERE EXISTS
(SELECT COUNT(*)
FROM table2
WHERE t2b = 'X');
ANSWER
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
12.5. NOT EXISTS Keyword Sub-query
The NOT EXISTS phrases looks for the non-existence of rows in the sub-query result set:
-
If the sub-query matches on no rows, the result is true.
-
If the sub-query has rows, the result is false.
We can use a NOT EXISTS check to create something similar to an ALL check, but with one very important difference. The two checks will handle nulls differently. To illustrate, consider the following two queries, both of which will return a row from TABLE1 only when it equals all of the matching rows in TABLE2:
SELECT *
FROM table1
WHERE NOT EXISTS
(SELECT *
FROM table2
WHERE t2c >= 'A'
AND t2c <> t1a);
ANSWERS
T1A | T1B |
---|---|
A |
AA |
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2c
FROM table2
WHERE t2c >= 'A');
The above two queries are very similar. Both define a set of rows in TABLE2 where the T2C value is greater than or equal to "A", and then both look for matching TABLE2 rows that are not equal to the current T1A value. If a row is found, the sub-query is false. What happens when no TABLE2 rows match the ">=" predicate? As is shown below, both of our test queries treat an empty set as a match:
-- NOT EXISTS
SELECT *
FROM table1
WHERE NOT EXISTS
(SELECT *
FROM table2
WHERE t2c >= 'X'
AND t2c <> t1a);
-- ALL
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2c
FROM table2
WHERE t2c >= 'X');
ANSWERS
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
One might think that the above two queries are logically equivalent, but they are not. As is shown below, they return different results when the sub-query answer set can include nulls:
-- NOT EXISTS
SELECT *
FROM table1
WHERE NOT EXISTS
(SELECT *
FROM table2
WHERE t2c <> t1a);
-- ALL
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2c
FROM table2);
ANSWER NOT EXISTS
T1A | T1B |
---|---|
A |
AA |
ANSWER_ALL =⇒ no rows
A sub-query can only return true or false, but a Db2 field value can either match (i.e. be true), or not match (i.e. be false), or be unknown. It is the differing treatment of unknown values that is causing the above two queries to differ: * In the ALL sub-query, each value in T1A is checked against all of the values in T2C. The null value is checked, deemed to differ, and so the sub-query always returns false. * In the NOT EXISTS sub-query, each value in T1A is used to find those T2C values that are not equal. For the T1A values "B" and "C", the T2C value "A" does not equal, so the NOT EXISTS check will fail. But for the T1A value "A", there are no "not equal" values in T2C, because a null value does not "not equal" a literal. So the NOT EXISTS check will pass.
The following three queries list those T2C values that do "not equal" a given T1A value:
SELECT * -- (a)
FROM table2
WHERE t2c <> 'A';
SELECT * -- (b)
FROM table2
WHERE t2c <> 'B';
SELECT * -- (c)
FROM table2
WHERE t2c <> 'C';
ANSWER (a) (no rows)
T2A | T2B | T2C |
---|
ANSWER (b)
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
ANSWER (c)
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
To make a NOT EXISTS sub-query that is logically equivalent to the ALL sub-query that we have used above, one can add an additional check for null T2C values:
SELECT *
FROM table1
WHERE NOT EXISTS
(SELECT *
FROM table2
WHERE t2c <> t1a
OR t2c IS NULL);
ANSWER =⇒ no rows
One problem with the above query is that it is not exactly obvious. Another is that the two T2C predicates will have to be fenced in with parenthesis if other predicates (on TABLE2) exist. For these reasons, use an ALL sub-query when that is what you mean to do.
12.6. IN Keyword Sub-Query
The IN sub-query check is similar to the ANY and SOME checks:
-
If any row in the sub-query result matches, the answer is true.
-
If the sub-query result is empty, the answer is false.
-
If no row in the sub-query result matches, the answer is also false.
-
If all of the values in the sub-query result are null, the answer is false.
Below is an example that compares the T1A and T2A columns. Two rows match:
SELECT *
FROM table1
WHERE t1a IN
(SELECT t2a
FROM table2);
ANSWER
T1A | T1B |
---|---|
A |
AA |
B |
BB |
In the next example, no rows match because the sub-query result is an empty set:
SELECT *
FROM table1
WHERE t1a IN
(SELECT t2a
FROM table2
WHERE t2a >= 'X');
ANSWER =⇒ 0 rows
The IN, ANY, SOME, and ALL checks all look for a match. Because one null value does not equal another null value, having a null expression in the "top" table causes the sub-query to always returns false:
-- IN
SELECT *
FROM table2
WHERE t2c IN
(SELECT t2c
FROM table2);
-- = ANY
SELECT *
FROM table2
WHERE t2c = ANY
(SELECT t2c
FROM table2);
ANSWERS
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
12.7. NOT IN Keyword Sub-Queries
Sub-queries that look for the non-existence of a row work largely as one would expect, except when a null value in involved. To illustrate, consider the following query, where we want to see if the current T1A value is not in the set of T2C values:
SELECT *
FROM table1
WHERE t1a NOT IN
(SELECT t2c
FROM table2);
ANSWER =⇒ 0 rows
Observe that the T1A values "B" and "C" are obviously not in T2C, yet they are not returned. The sub-query result set contains the value null, which causes the NOT IN check to return unknown, which equates to false. The next example removes the null values from the sub-query result, which then enables the NOT IN check to find the non-matching values:
SELECT *
FROM table1
WHERE t1a NOT IN
(SELECT t2c
FROM table2
WHERE t2c IS NOT NULL);
ANSWER
T1A | T1B |
---|---|
B |
BB |
C |
CC |
Another way to find the non-matching values while ignoring any null rows in the sub-query, is to use an EXISTS check in a correlated sub-query:
SELECT *
FROM table1
WHERE NOT EXISTS
(SELECT *
FROM table2
WHERE t1a = t2c);
ANSWER
T1A | T1B |
---|---|
B |
BB |
C |
CC |
12.8. Correlated vs. Uncorrelated Sub-Queries
An uncorrelated sub-query is one where the predicates in the sub-query part of SQL statement have no direct relationship to the current row being processed in the "top" table (hence uncorrelated). The following sub-query is uncorrelated:
SELECT *
FROM table1
WHERE t1a IN
(SELECT t2a
FROM table2);
ANSWER
T1A | T1B |
---|---|
A |
AA |
B |
BB |
A correlated sub-query is one where the predicates in the sub-query part of the SQL statement cannot be resolved without reference to the row currently being processed in the "top" table (hence correlated). The following query is correlated:
SELECT *
FROM table1
WHERE t1a IN
(SELECT t2a
FROM table2
WHERE t1a = t2a);
ANSWER
T1A | T1B |
---|---|
A |
AA |
B |
BB |
Below is another correlated sub-query. Because the same table is being referred to twice, correlation names have to be used to delineate which column belongs to which table:
SELECT *
FROM table2 aa
WHERE EXISTS
(SELECT *
FROM table2 bb
WHERE aa.t2a = bb.t2b);
ANSWER
T2A | T2B | T2C |
---|---|---|
A |
A |
A |
12.9. Which is Faster
In general, if there is a suitable index on the sub-query table, use a correlated sub-query. Else, use an uncorrelated sub-query. However, there are several very important exceptions to this rule, and some queries can only be written one way. NOTE: The Db2 optimizer is not as good at choosing the best access path for sub-queries as it is with joins. Be prepared to spend some time doing tuning.
12.10. Multi-Field Sub-Queries
Imagine that you want to compare multiple items in your sub-query. The following examples use an IN expression and a correlated EXISTS sub-query to do two equality checks:
SELECT *
FROM table1
WHERE (t1a,t1b) IN
(SELECT t2a, t2b
FROM table2);
SELECT *
FROM table1
WHERE EXISTS
(SELECT *
FROM table2
WHERE t1a = t2a
AND t1b = t2b);
ANSWER =⇒ 0 rows
Observe that to do a multiple-value IN check, you put the list of expressions to be compared in parenthesis, and then select the same number of items in the sub-query. An IN phrase is limited because it can only do an equality check. By contrast, use whatever predicates you want in an EXISTS correlated sub-query to do other types of comparison:
SELECT *
FROM table1
WHERE EXISTS
(SELECT *
FROM table2
WHERE t1a = t2a
AND t1b >= t2b);
ANSWER
T1A | T1B |
---|---|
A |
AA |
B |
BB |
12.11. Nested Sub-Queries
Some business questions may require that the related SQL statement be written as a series of nested sub-queries. In the following example, we are after all employees in the EMPLOYEE table who have a salary that is greater than the maximum salary of all those other employees that do not work on a project with a name beginning 'MA'.
SELECT empno
, lastname
, salary
FROM employee
WHERE salary >
(SELECT MAX(salary)
FROM employee
WHERE empno NOT IN
(SELECT empno
FROM emp_act
WHERE projno LIKE 'MA%'
)
)
ORDER BY 1;
ANSWER
EMPNO | LASTNAME | SALARY |
---|---|---|
000010 |
HAAS |
52750.00 |
000110 |
LUCCHESSI |
46500.00 |
12.12. Usage Examples
In this section we will use various sub-queries to compare our two test tables - looking for those rows where none, any, ten, or all values match.
12.12.1. Beware of Nulls
The presence of null values greatly complicates sub-query usage. Not allowing for them when they are present can cause one to get what is arguably a wrong answer. And do not assume that just because you don’t have any nullable fields that you will never therefore encounter a null value. The DEPTNO table in the Department table is defined as not null, but in the following query, the maximum DEPTNO that is returned will be null:
SELECT COUNT(*) AS #rows
, MAX(deptno) AS maxdpt
FROM department
WHERE deptname LIKE 'Z%'
ORDER BY 1;
ANSWER
#ROWS | MAXDEPT |
---|---|
0 |
null |
12.12.2. True if NONE Match
Find all rows in TABLE1 where there are no rows in TABLE2 that have a T2C value equal to the current T1A value in the TABLE1 table:
SELECT *
FROM table1 t1
WHERE 0 =
(SELECT COUNT(*)
FROM table2 t2
WHERE t1.t1a = t2.t2c);
SELECT *
FROM table1 t1
WHERE NOT EXISTS
(SELECT *
FROM table2 t2
WHERE t1.t1a = t2.t2c);
SELECT *
FROM table1
WHERE t1a NOT IN
(SELECT t2c
FROM table2
WHERE t2c IS NOT NULL);
ANSWER
T1A | T1B |
---|---|
B |
BB |
C |
CC |
Observe that in the last statement above we eliminated the null rows from the sub-query. Had this not been done, the NOT IN check would have found them and then returned a result of "unknown" (i.e. false) for all of rows in the TABLE1A table.
12.12.3. Using a Join
Another way to answer the same problem is to use a left outer join, going from TABLE1 to TABLE2 while matching on the T1A and T2C fields. Get only those rows (from TABLE1) where the corresponding T2C value is null:
SELECT t1.*
FROM table1 t1
LEFT OUTER JOIN table2 t2
ON t1.t1a = t2.t2c
WHERE t2.t2c IS NULL;
ANSWER
T1A | T1B |
---|---|
B |
BB |
C |
CC |
12.12.4. True if ANY Match
Find all rows in TABLE1 where there are one, or more, rows in TABLE2 that have a T2C value equal to the current T1A value:
SELECT *
FROM table1 t1
WHERE EXISTS
(SELECT *
FROM table2 t2
WHERE t1.t1a = t2.t2c);
SELECT *
FROM table1 t1
WHERE 1 <=
(SELECT COUNT(*)
FROM table2 t2
WHERE t1.t1a = t2.t2c);
SELECT *
FROM table1
WHERE t1a = ANY
(SELECT t2c
FROM table2);
ANSWER
T1A | T1B |
---|---|
A |
AA |
SELECT *
FROM table1
WHERE t1a = SOME
(SELECT t2c
FROM table2);
SELECT *
FROM table1
WHERE t1a IN
(SELECT t2c
FROM table2);
Of all of the above queries, the second query is almost certainly the worst performer. All of the others can, and probably will, stop processing the sub-query as soon as it encounters a single matching value. But the sub-query in the second statement has to count all of the matching rows before it return either a true or false indicator.
12.12.5. Using a Join
This question can also be answered using an inner join. The trick is to make a list of distinct T2C values, and then join that list to TABLE1 using the T1A column. Several variations on this theme are given below:
WITH t2 AS
(SELECT DISTINCT t2c
FROM table2
)
SELECT t1.*
FROM table1 t1
, t2
WHERE t1.t1a = t2.t2c;
SELECT t1.*
FROM table1 t1
, (SELECT DISTINCT t2c
FROM table2
) AS t2
WHERE t1.t1a = t2.t2c;
ANSWER
T1A | T1B |
---|---|
A |
AA |
SELECT t1.* FROM table1 t1 INNER JOIN (SELECT DISTINCT t2c FROM table2 )AS t2 ON t1.t1a = t2.t2c;
12.13. True if TEN Match
Find all rows in TABLE1 where there are exactly ten rows in TABLE2 that have a T2B value equal to the current T1A value in the TABLE1 table:
SELECT *
FROM table1 t1
WHERE 10 =
(SELECT COUNT(*)
FROM table2 t2
WHERE t1.t1a = t2.t2b);
SELECT *
FROM table1
WHERE EXISTS
(SELECT t2b
FROM table2
WHERE t1a = t2b
GROUP BY t2b
HAVING COUNT(*) = 10);
SELECT *
FROM table1
WHERE t1a IN
(SELECT t2b
FROM table2
GROUP BY t2b
HAVING COUNT(*) = 10);
ANSWER =⇒ 0 rows
The first two queries above use a correlated sub-query. The third is uncorrelated. The next query, which is also uncorrelated, is guaranteed to befuddle your coworkers. It uses a multifield IN (see Multi-Field Sub-Queries for more notes) to both check T2B and the count at the same time:
SELECT *
FROM table1
WHERE (t1a,10) IN
(SELECT t2b
, COUNT(*)
FROM table2
GROUP BY t2b);
ANSWER =⇒ 0 rows
12.13.1. Using a Join
To answer this generic question using a join, one simply builds a distinct list of T2B values that have ten rows, and then joins the result to TABLE1:
WITH t2 AS
(SELECT t2b
FROM table2
GROUP BY t2b
HAVING COUNT(*) = 10
)
SELECT t1.*
FROM table1 t1
, t2
WHERE t1.t1a = t2.t2b;
ANSWER =⇒ 0 rows
SELECT t1.*
FROM table1 t1
, (SELECT t2b
FROM table2
GROUP BY t2b
HAVING COUNT(*) = 10) AS t2
WHERE t1.t1a = t2.t2b;
SELECT t1.*
FROM table1 t1
INNER JOIN
(SELECT t2b
FROM table2
GROUP BY t2b
HAVING COUNT(*) = 10 )AS t2
ON t1.t1a = t2.t2b;
12.14. True if ALL match
Find all rows in TABLE1 where all matching rows in TABLE2 have a T2B value equal to the current T1A value in the TABLE1 table. Before we show some SQL, we need to decide what to do about nulls and empty sets:
-
When nulls are found in the sub-query, we can either deem that their presence makes the relationship false, which is what Db2 does, or we can exclude nulls from our analysis.
-
When there are no rows found in the sub-query, we can either say that the relationship is false, or we can do as Db2 does, and say that the relationship is true.
See All Keyword Sub-Query for a detailed discussion of the above issues.
The next two queries use the basic Db2 logic for dealing with empty sets; In other words, if no rows are found by the sub-query, then the relationship is deemed to be true. Likewise, the relationship is also true if all rows found by the sub-query equal the current T1A value:
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2b
FROM table2);
SELECT *
FROM table1
WHERE NOT EXISTS
(SELECT *
FROM table2
WHERE t1a <> t2b);
ANSWER
T1A | T1B |
---|---|
A |
AA |
The next two queries are the same as the prior, but an extra predicate has been included in the sub-query to make it return an empty set. Observe that now all TABLE1 rows match:
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2b
FROM table2
WHERE t2b >= 'X');
ANSWER
T1A | T1B |
---|---|
A |
AA |
B |
BB |
C |
CC |
SELECT *
FROM table1
WHERE NOT EXISTS
(SELECT *
FROM table2
WHERE t1a <> t2b
AND t2b >= 'X');
12.15. False if no Matching Rows
The next two queries differ from the above in how they address empty sets. The queries will return a row from TABLE1 if the current T1A value matches all of the T2B values found in the sub-query, but they will not return a row if no matching values are found:
SELECT *
FROM table1
WHERE t1a = ALL
(SELECT t2b
FROM table2
WHERE t2b >= 'X')
AND 0 <>
(SELECT COUNT(*)
FROM table2
WHERE t2b >= 'X');
SELECT *
FROM table1
WHERE t1a IN
(SELECT MAX(t2b)
FROM table2
WHERE t2b >= 'X'
HAVING COUNT(DISTINCT t2b) = 1);
ANSWER =⇒ 0 rows
Both of the above statements have flaws: The first processes the TABLE2 table twice, which not only involves double work, but also requires that the sub-query predicates be duplicated. The second statement is just plain strange.
13. Union, Intersect, and Except
A UNION, EXCEPT, or INTERCEPT expression combines sets of columns into new sets of columns. An illustration of what each operation does with a given set of data is shown below:
R1 R1 R1 R1 R1 R1 R1 UNION UNION INTERSECT INTERSECT EXCEPT EXCEPT MINUS R2 ALL R2 ALL R2 ALL R2 R1 R2 R2 R2 R2 -- -- ----- ----- --------- ----- ------ ------ ----- A A A A A A E A E A A B A B A C A B C A C B C B B D A B E B B E A C C C B C D B C B E B B C C C C D E
Unlike the UNION and INTERSECT operations, the EXCEPT statement is not commutative. This means that "A EXCEPT B" is not the same as "B EXCEPT A". |
Sample Views
CREATE VIEW R1 (R1)
AS VALUES ('A'), ('A'), ('A'), ('B'), ('B'), ('C'), ('C'), ('C'), ('E');
CREATE VIEW R2 (R2)
AS VALUES ('A'), ('A'), ('B'), ('B'), ('B'), ('C'), ('D');
SELECT R1
FROM R1
ORDER BY R1;
SELECT R2
FROM R2
ORDER BY R2;
ANSWER
R1 | R2 |
---|---|
A |
A |
A |
A |
A |
B |
B |
B |
B |
B |
C |
C |
C |
D |
C |
|
E |
Usage Notes
13.1. Union & Union All
A UNION operation combines two sets of columns and removes duplicates. The UNION ALL expression does the same but does not remove the duplicates. -Union and Union All SQL
SELECT R1
FROM R1
UNION
SELECT R2
FROM R2
ORDER BY 1;
SELECT R1
FROM R1
UNION ALL
SELECT R2
FROM R2
ORDER BY 1;
R1 | R2 | UNION | UNION ALL |
---|---|---|---|
A |
A |
A |
A |
A |
A |
B |
A |
A |
B |
C |
A |
B |
B |
D |
A |
B |
B |
E |
A |
C |
C |
B |
|
C |
D |
B |
|
C |
B |
||
E |
B |
||
B |
|||
C |
|||
C |
|||
C |
|||
C |
|||
D |
|||
E |
Recursive SQL requires that there be a UNION ALL phrase between the two main parts of the statement. The UNION ALL, unlike the UNION, allows for duplicate output rows which is what often comes out of recursive processing. |
===Intersect & Intersect All
An INTERSECT operation retrieves the matching set of distinct values (not rows) from two columns. The INTERSECT ALL returns the set of matching individual rows.
SELECT R1
FROM R1
INTERSECT
SELECT R2
FROM R2
ORDER BY 1;
SELECT R1
FROM R1
INTERSECT ALL
SELECT R2
FROM R2
ORDER BY 1;
R1 | R2 | INTERSECT | INTERSECT ALL |
---|---|---|---|
A |
A |
A |
A |
A |
A |
B |
A |
A |
B |
C |
B |
B |
B |
B |
|
B |
B |
C |
|
C |
C |
||
C |
D |
||
C |
|||
E |
An INTERSECT and/or EXCEPT operation is done by matching ALL of the columns in the top and bottom result-sets. In other words, these are row, not column, operations. It is not possible to only match on the keys, yet at the same time, also fetch non-key columns. To do this, one needs to use a sub-query.
13.2. Except, Except All & Minus
An EXCEPT operation retrieves the set of distinct data values (not rows) that exist in the first the table but not in the second. The EXCEPT ALL returns the set of individual rows that exist only in the first table. The word MINUS is a synonym for EXCEPT.
SELECT R1
FROM R1
EXCEPT
SELECT R2
FROM R2
ORDER BY 1;
SELECT R1
FROM R1
EXCEPT ALL
SELECT R2
FROM R2
ORDER BY 1;
R1 | R2 | R1 EXCEPT R2 | R1 EXCEPT ALL R2 |
---|---|---|---|
A |
A |
E |
A |
A |
A |
C |
|
A |
B |
C |
|
B |
B |
E |
|
B |
B |
||
C |
C |
||
C |
D |
||
C |
|||
E |
Because the EXCEPT/MINUS operation is not commutative, using it in the reverse direction (i.e. R2 to R1 instead of R1 to R2) will give a different result:
SELECT R2
FROM R2
EXCEPT
SELECT R1
FROM R1
ORDER BY 1;
SELECT R2
FROM R2
EXCEPT ALL
SELECT R1
FROM R1
ORDER BY 1;
R1 | R2 | R2 EXCEPT R1 | R2 EXCEPT ALL R1 |
---|---|---|---|
A |
A |
D |
B |
A |
A |
D |
|
A |
B |
||
B |
B |
||
B |
B |
||
C |
C |
||
C |
D |
||
C |
|||
E |
Only the EXCEPT/MINUS operation is not commutative. Both the UNION and the INTERSECT operations work the same regardless of which table is on top or on bottom. |
13.2.1. Precedence Rules
When multiple operations are done in the same SQL statement, there are precedence rules:
-
Operations in parenthesis are done first.
-
INTERSECT operations are done before either UNION or EXCEPT.
-
Operations of equal worth are done from top to bottom.
The next example illustrates how parenthesis can be used change the processing order:
SELECT R1
FROM R1
UNION
SELECT R2
FROM R2
EXCEPT
SELECT R2
FROM R2
ORDER BY 1;
ANSWER
E |
---|
(SELECT R1
FROM R1
UNION
SELECT R2
FROM R2
)
EXCEPT
SELECT R2
FROM R2
ORDER BY 1;
ANSWER
E |
---|
SELECT R1
FROM R1
UNION
(SELECT R2
FROM R2
EXCEPT
SELECT R2
FROM R2
)
ORDER BY 1;
ANSWER
A |
---|
B |
C |
E |
13.2.2. Unions and Views
Imagine that one has a series of tables that track sales data, with one table for each year. One can define a view that is the UNION ALL of these tables, so that a user would see them as a single object. Such a view can support inserts, updates, and deletes, as long as each table in the view has a constraint that distinguishes it from all the others. Below is an example:
CREATE TABLE sales_data_2002
( sales_date DATE NOT NULL
, daily_seq# INTEGER NOT NULL
, cust_id INTEGER NOT NULL
, amount DEC(10, 2) NOT NULL
, invoice# INTEGER NOT NULL
, sales_rep CHAR(10) NOT NULL
, CONSTRAINT C CHECK (YEAR(sales_date) = 2002)
, PRIMARY KEY (sales_date, daily_seq#)
);
CREATE TABLE sales_data_2003
( sales_date DATE NOT NULL
, daily_seq# INTEGER NOT NULL
, cust_id INTEGER NOT NULL
, amount DEC(10,2) NOT NULL
, invoice# INTEGER NOT NULL
, sales_rep CHAR(10) NOT NULL
, CONSTRAINT C CHECK (YEAR(sales_date) = 2003)
, PRIMARY KEY (sales_date, daily_seq#));
CREATE VIEW sales_data AS
SELECT *
FROM sales_data_2002
UNION ALL
SELECT *
FROM sales_data_2003;
Below is some SQL that changes the contents of the above view:
INSERT INTO sales_data VALUES
('2002-11-22',1,123,100.10,996,'SUE')
, ('2002-11-22',2,123,100.10,997,'JOHN')
, ('2003-01-01',1,123,100.10,998,'FRED')
, ('2003-01-01',2,123,100.10,999,'FRED');
UPDATE sales_data
SET amount = amount / 2
WHERE sales_rep = 'JOHN';
DELETE
FROM sales_data
WHERE sales_date = '2003-01-01'
AND daily_seq# = 2;
Below is the view contents, after the above is run: .View contents after insert, update, delete
SALES_DATE | DAILY_SEQ# | CUST_ID | AMOUNT | INVOICE# | SALES_REP |
---|---|---|---|---|---|
01/01/2003 |
1 |
123 |
100.10 |
998 |
FRED |
11/22/2002 |
1 |
123 |
100.10 |
996 |
SUE |
11/22/2002 |
2 |
123 |
50.05 |
997 |
JOHN |
14. Materialized Query Tables
14.1. Introduction
A materialized query table contains the results of a query. The Db2 optimizer knows this and can, if appropriate, redirect a query that is against the source table(s) to use the materialized query table instead. This can make the query run much faster. The following statement defines a materialized query table:
CREATE TABLE staff_summary AS
(SELECT dept
, COUNT(*) AS count_rows
, SUM(id) AS sum_id
FROM staff
GROUP BY dept)
DATA INITIALLY DEFERRED REFRESH IMMEDIATE;
Below on the left is a query that is very similar to the one used in the above CREATE. The Db2 optimizer can convert this query into the optimized equivalent on the right, which uses the materialized query table. Because (in this case) the data in the materialized query table is maintained in sync with the source table, both statements will return the same answer.
ORIGINAL QUERY:
SELECT dept
, AVG(id)
FROM staff
GROUP BY dept
-- OPTIMIZED QUERY:
SELECT Q1.dept AS "dept"
, Q1.sum_id / Q1.count_rows
FROM staff_summary AS Q1
When used appropriately, materialized query tables can cause dramatic improvements in query performance. For example, if in the above STAFF table there was, on average, about 5,000 rows per individual department, referencing the STAFF_SUMMARY table instead of the STAFF table in the sample query might be about 1,000 times faster.
14.2. Db2 Optimizer Issues
In order for a materialized query table to be considered for use by the Db2 optimizer, the following has to be true:
-
The table has to be refreshed at least once.
-
The table MAINTAINED BY parameter and the related Db2 special registers must correspond. For example, if the table is USER maintained, then the CURRENT REFRESH AGE special register must be set to ANY, and the CURRENT MAINTAINED TABLE TYPES FOR OPTIMIZATION special register must be set to USER or ALL. See Optimizer Options for more details on these registers.
14.2.1. Usage Notes
A materialized query table is defined using a variation of the standard CREATE TABLE statement. Instead of providing an element list, one supplies a SELECT statement, and defines the refresh option. ===Syntax Options ====Refresh
REFRESH DEFERRED: The data is refreshed whenever one does a REFRESH TABLE. At this point, Db2 will first delete all of the existing rows in the table, then run the select statement defined in the CREATE to (you guessed it) repopulate.
REFRESH IMMEDIATE: Once created, this type of table has to be refreshed once using the REFRESH statement. From then on, Db2 will maintain the materialized query table in sync with the source table as changes are made to the latter.
Materialized query tables that are defined REFRESH IMMEDIATE are obviously more useful in that the data in them is always current. But they may cost quite a bit to maintain, and not all queries can be defined thus.
Query Optimization
ENABLE: The table is used for query optimization when appropriate. This is the default. The table can also be queried directly.
DISABLE: The table will not be used for query optimization. It can be queried directly.
Maintained By
SYSTEM: The data in the materialized query table is maintained by the system. This is the default.
USER: The user is allowed to perform insert, update, and delete operations against the materialized query table. The table cannot be refreshed. This type of table can be used when you want to maintain your own materialized query table (e.g. using triggers) to support features not provided by Db2. The table can also be defined to enable query optimization, but the optimizer will probably never use it as a substitute for a real table.
FEDERATED_TOOL: The data in the materialized query table is maintained by the replication tool. Only a REFRESH DEFERRED table can be maintained using this option.
14.2.2. Options vs. Actions
The following table compares materialized query table options to subsequent actions:
MATERIALIZED QUERY TABLE ALLOWABLE ACTIONS ON TABLE REFRESH MAINTAINED BY REFRESH TABLE INSERT/UPDATE/DELETE --------- ------------- ------------- -------------------- DEFERRED SYSTEM yes no USER no yes IMMEDIATE SYSTEM yes no
14.3. Select Statement
Various restrictions apply to the select statement that is used to define the materialized query table. In general, materialized query tables defined refresh-immediate need simpler queries than those defined refresh-deferred.
Refresh Deferred Tables
-
The query must be a valid SELECT statement.
-
Every column selected must have a name.
-
An ORDER BY is not allowed.
-
Reference to a typed table or typed view is not allowed.
-
Reference to declared temporary table is not allowed.
-
Reference to a nickname or materialized query table is not allowed.
-
Reference to a system catalogue table is not allowed. Reference to an explain table is allowed, but is impudent.
-
Reference to NODENUMBER, PARTITION, or any other function that depends on physical characteristics, is not allowed.
-
Reference to a datalink type is not allowed.
-
Functions that have an external action are not allowed.
-
Scalar functions, or functions written in SQL, are not allowed. So SUM(SALARY) is fine, but SUM(INT(SALARY)) is not allowed.
===Refresh Immediate Tables
All of the above restrictions apply, plus the following:
-
If the query references more than one table or view, it must define as inner join, yet not use the INNER JOIN syntax (i.e. must use old style).
-
If there is a GROUP BY, the SELECT list must have a COUNT() or COUNT_BIG() column.
-
Besides the COUNT and COUNT_BIG, the only other column functions supported are SUM and GROUPING - all with the DISTINCT phrase. Any field that allows nulls, and that is summed, but also have a COUNT(column name) function defined.
-
Any field in the GROUP BY list must be in the SELECT list.
-
The table must have at least one unique index defined, and the SELECT list must include (amongst other things) all the columns of this index.
-
Grouping sets, CUBE an ROLLUP are allowed. The GROUP BY items and associated GROUPING column functions in the select list must for a unique key of the result set.
-
The HAVING clause is not allowed.
-
The DISTINCT clause is not allowed.
-
Non-deterministic functions are not allowed.
-
Special registers are not allowed.
-
If REPLICATED is specified, the table must have a unique key.
14.4. Optimizer Options
A materialized query table that has been defined ENABLE QUERY OPTIMIZATION, and has been refreshed, is a candidate for use by the Db2 optimizer if, and only if, three Db2 special registers are set to match the table status:
-
CURRENT MAINTAINED TABLE TYPES FOR OPTIMIZATION.
-
CURRENT QUERY OPTIMIZATION.
-
CURRENT REFRESH AGE.
Each of the above are discussed below.
CURRENT REFRESH AGE
The refresh age special register tells the Db2 optimizer how up-to-date the data in an materialized query table has to be in order to be considered. There are only two possible values:
-
0: Only use those materialized query tables that are defined as refresh-immediate are eligible. This is the default.
-
99,999,999,999,999: Consider all valid materialized query tables. This is the same as ANY.
The above number is a 26-digit decimal value that is a timestamp duration, but without the microsecond component. The value ANY is logically equivalent. |
The database default value can be changed using the following command:
UPDATE DATABASE CONFIGURATION USING dft_refresh_age ANY;
The database default value can be overridden within a thread using the SET REFRESH AGE statement.
Below are some examples of the SET command:
SET CURRENT REFRESH AGE 0;
SET CURRENT REFRESH AGE = ANY;
SET CURRENT REFRESH AGE = 99999999999999;
CURRENT MAINTAINED TYPES
The current maintained types special register tells the Db2 optimizer what types of materialized query table that are defined refresh deferred are to be considered - assuming that the refresh-age parameter is not set to zero:
-
ALL: All refresh-deferred materialized query tables are to be considered. If this option is chosen, no other option can be used.
-
NONE: No refresh-deferred materialized query tables are to be considered. If this option is chosen, no other option can be used.
-
SYSTEM: System-maintained refresh-deferred materialized query tables are to be considered. This is the default.
-
USER: User-maintained refresh-deferred materialized query tables are to be considered.
-
FEDERATED TOOL: Federated-tool-maintained refresh-deferred materialized query tables are to be considered, but only if the CURRENT QUERY OPTIMIZATION special register is 2 or greater than 5.
-
CURRENT MAINTAINED TABLE TYPES FOR OPTIMIZATION: The existing values for this special register are used.
The database default value can be changed using the following command:
UPDATE DATABASE CONFIGURATION USING dft_refresh_age ANY;
The database default value can be overridden within a thread using the SET REFRESH AGE statement. Below are some examples of the SET command:
SET CURRENT MAINTAINED TYPES = ALL;
SET CURRENT MAINTAINED TABLE TYPES = SYSTEM;
SET CURRENT MAINTAINED TABLE TYPES FOR OPTIMIZATION = USER, SYSTEM;
CURRENT QUERY OPTIMIZATION
The current query optimization special register tells the Db2 optimizer what set of optimization techniques to use. The value can range from zero to nine - except for four or eight. A value of five or above will cause the optimizer to consider using materialized query tables.
The database default value can be changed using the following command:
UPDATE DATABASE CONFIGURATION USING DFT_QUERYOPT 5;
The database default value can be overridden within a thread using the SET CURRENT QUERY OPTIMIZATION statement.
Below are an example of the SET command:
SET CURRENT QUERY OPTIMIZATION = 9;
14.4.1. What Matches What
Assuming that the current query optimization special register is set to five or above, the Db2 optimizer will consider using a materialized query table (instead of the base table) when any of the following conditions are true:
MQT DEFINITION DATABASE/APPLICATION STATUS DB2 ========================== =================================== USE REFRESH MAINTAINED-BY REFRESH-AGE MAINTAINED-TYPE MQT ========= ============== =========== ===================== === IMMEDIATE SYSTEM - - Yes DEFERRED SYSETM ANY ALL or SYSTEM Yes DEFERRED USER ANY ALL or USER Yes DEFERRED FEDERATED-TOOL ANY ALL or FEDERATED-TOOL Yes
14.4.2. Selecting Special Registers
One can select the relevant special register to see what the values are:
SELECT CURRENT REFRESH AGE AS age_ts
, CURRENT TIMESTAMP AS current_ts
, CURRENT QUERY OPTIMIZATION AS q_opt
FROM sysibm.sysdummy1;
14.4.3. Refresh Deferred Tables
A materialized query table defined REFRESH DEFERRED can be periodically updated using the REFRESH TABLE command. Below is an example of a such a table that has one row per qualifying department in the STAFF table:
CREATE TABLE staff_names AS
(SELECT dept
, COUNT(*) AS count_rows
, SUM(salary) AS sum_salary
, AVG(salary) AS avg_salary
, MAX(salary) AS max_salary
, MIN(salary) AS min_salary
, STDDEV(salary) AS std_salary
, VARIANCE(salary) AS var_salary
, CURRENT TIMESTAMP AS last_change
FROM staff
WHERE TRANSLATE(name) LIKE '%A%'
AND salary > 10000
GROUP BY dept
HAVING COUNT(*) = 1
) DATA INITIALLY DEFERRED REFRESH DEFERRED;
14.4.4. Refresh Immediate Tables
A materialized query table defined REFRESH IMMEDIATE is automatically maintained in sync with the source table by Db2. As with any materialized query table, it is defined by referring to a query. Below is a table that refers to a single source table:
CREATE TABLE emp_summary AS
(SELECT emp.workdept
, COUNT(*) AS num_rows
, COUNT(emp.salary) AS num_salary
, SUM(emp.salary) AS sum_salary
, COUNT(emp.comm) AS num_comm
, SUM(emp.comm) AS sum_comm
FROM employee emp
GROUP BY emp.workdept
) DATA INITIALLY DEFERRED REFRESH IMMEDIATE;
Below is a query that can use the above materialized query table in place of the base table:
SELECT emp.workdept
, DEC(SUM(emp.salary),8,2) AS sum_sal
, DEC(AVG(emp.salary),7,2) AS avg_sal
, SMALLINT(COUNT(emp.comm)) AS #comms
, SMALLINT(COUNT(*)) AS #emps
FROM employee emp
WHERE emp.workdept > 'C'
GROUP BY emp.workdept
HAVING COUNT(*) <> 5
AND SUM(emp.salary) > 50000
ORDER BY sum_sal DESC;
The next query can also use the materialized query table. This time, the data returned from the materialized query table is qualified by checking against a sub-query:
SELECT emp.workdept
, COUNT(*) AS #rows
FROM employee emp
WHERE emp.workdept IN
(SELECT deptno
FROM department
WHERE deptname LIKE '%S%'
)
GROUP BY emp.workdept
HAVING SUM(salary) > 50000;
This last example uses the materialized query table in a nested table expression:
SELECT #emps
, DEC(SUM(sum_sal), 9, 2) AS sal_sal
, SMALLINT(COUNT(*)) AS #depts
FROM
(SELECT emp.workdept
, DEC(SUM(emp.salary), 8, 2) AS sum_sal
, MAX(emp.salary) AS max_sal
, SMALLINT(COUNT(*)) AS #emps
FROM employee emp
GROUP BY emp.workdept
) AS XXX
GROUP BY #emps
HAVING COUNT(*) > 1
ORDER BY #emps
FETCH FIRST 3 ROWS ONLY
OPTIMIZE FOR 3 ROWS;
14.4.5. Using Materialized Query Tables to Duplicate Data
All of the above materialized query tables have contained a GROUP BY in their definition. But this is not necessary. To illustrate, we will first create a simple table:
CREATE TABLE staff_all
( id SMALLINT NOT NULL
, name VARCHAR(9) NOT NULL
, job CHAR(5)
, salary DECIMAL(7, 2)
, PRIMARY KEY(id));
As long as the above table has a primary key, which it does, we can define a duplicate of the above using the following code:
CREATE TABLE staff_all_dup AS
(SELECT *
FROM staff_all)
DATA INITIALLY DEFERRED REFRESH IMMEDIATE;
We can also decide to duplicate only certain rows:
CREATE TABLE staff_all_dup_some AS
(SELECT *
FROM staff_all
WHERE id < 30)
DATA INITIALLY DEFERRED REFRESH IMMEDIATE;
Imagine that we had another table that listed all those staff that we are about to fire:
CREATE TABLE staff_to_fire
( id SMALLINT NOT NULL
, name VARCHAR(9) NOT NULL
, dept SMALLINT
, PRIMARY KEY(id));
We can create materialized query table that joins the above two staff tables as long as the following is true:
-
Both tables have identical primary keys (i.e. same number of columns).
-
The join is an inner join on the common primary key fields.
-
All primary key columns are listed in the SELECT.
Now for an example:
CREATE TABLE staff_combo AS
(SELECT aaa.id AS id1
, aaa.job AS job
, fff.id AS id2
, fff.dept AS dept
FROM staff_all aaa
, staff_to_fire fff
WHERE aaa.id = fff.id)
DATA INITIALLY DEFERRED REFRESH IMMEDIATE;
See Multi-table Materialized Query Tables for more examples of join usage.
14.4.6. Queries that don’t use Materialized Query Table
Below is a query that can not use the EMP_SUMMARY table because of the reference to the MAX function. Ironically, this query is exactly the same as the nested table expression above, but in the prior example the MAX is ignored because it is never actually selected:
SELECT emp.workdept
, DEC(SUM(emp.salary), 8, 2) AS sum_sal
, MAX(emp.salary) AS max_sal
FROM employee emp
GROUP BY emp.workdept;
The following query can’t use the materialized query table because of the DISTINCT clause:
SELECT emp.workdept
, DEC(SUM(emp.salary), 8, 2) AS sum_sal
, COUNT(DISTINCT salary) AS #salaries
FROM employee emp
GROUP BY emp.workdept;
14.4.7. Usage Notes and Restrictions
-
A materialized query table must be refreshed before it can be queried. If the table is defined refresh immediate, then the table will be maintained automatically after the initial refresh.
-
Make sure to commit after doing a refresh. The refresh does not have an implied commit.
-
Run RUNSTATS after refreshing a materialized query table.
-
One can not load data into materialized query tables.
-
One can not directly update materialized query tables.
To refresh a materialized query table, use either of the following commands:
REFRESH TABLE emp_summary;
COMMIT;
SET INTEGRITY FOR emp_summary iMMEDIATE CHECKED;
COMMIT;
14.4.8. Multi-table Materialized Query Tables
Single-table materialized query tables save having to look at individual rows to resolve a GROUP BY. Multi-table materialized query tables do this, and also avoid having to resolve a join.
CREATE TABLE dept_emp_summary AS
(SELECT emp.workdept
, dpt.deptname
, COUNT(*) AS num_rows
, COUNT(emp.salary) AS num_salary
, SUM(emp.salary) AS sum_salary
, COUNT(emp.comm) AS num_comm
, SUM(emp.comm) AS sum_comm
FROM employee emp
, department dpt
WHERE dpt.deptno = emp.workdept
GROUP BY emp.workdept
, dpt.deptname
) DATA INITIALLY DEFERRED REFRESH IMMEDIATE;
The following query is resolved using the above materialized query table:
SELECT d.deptname
, d.deptno
, DEC(AVG(e.salary), 7, 2) AS avg_sal
, SMALLINT(COUNT(*)) AS #emps
FROM department d
, employee e
WHERE e.workdept = d.deptno
AND d.deptname LIKE '%S%'
GROUP BY d.deptname
, d.deptno
HAVING SUM(e.comm) > 4000
ORDER BY avg_sal DESC;
Here is the SQL that Db2 generated internally to get the answer:
SELECT Q2.$C0 AS "deptname"
, Q2.$C1 AS "deptno"
, Q2.$C2 AS "avg_sal"
, Q2.$C3 AS "#emps"
FROM
(SELECT Q1.deptname AS $C0
, Q1.workdept AS $C1
, DEC((Q1.sum_salary / Q1.num_salary),7,2) AS $C2
, SMALLINT(Q1.num_rows) AS $C3
FROM dept_emp_summary AS Q1
WHERE (Q1.deptname LIKE '%S%')
AND (4000 < Q1.sum_comm)
) AS Q2
ORDER BY Q2.$C2 DESC;
14.4.9. Rules and Restrictions
-
The join must be an inner join, and it must be written in the old style syntax.
-
Every table accessed in the join (except one?) must have a unique index.
-
The join must not be a Cartesian product.
-
The GROUP BY must include all of the fields that define the unique key for every table (except one?) in the join.
14.4.10. Three-table Example
CREATE TABLE dpt_emp_act_sumry AS
(SELECT emp.workdept
, dpt.deptname
, emp.empno
, emp.firstnme
, SUM(act.emptime) AS sum_time
, COUNT(act.emptime) AS num_time
, COUNT(*) AS num_rows
FROM department dpt
, employee emp
, emp_act act
WHERE dpt.deptno = emp.workdept
AND emp.empno = act.empno
GROUP BY emp.workdept
, dpt.deptname
, emp.empno
, emp.firstnme
) DATA INITIALLY DEFERRED REFRESH IMMEDIATE;
Now for a query that will use the above:
SELECT d.deptno
, d.deptname
, DEC(AVG(a.emptime),5,2) AS avg_time
FROM department d
, employee e
, emp_act a
WHERE d.deptno = e.workdept
AND e.empno = a.empno
AND d.deptname LIKE '%S%'
AND e.firstnme LIKE '%S%'
GROUP BY d.deptno
, d.deptname
ORDER BY 3 DESC;
And here is the Db2 generated SQL:
SELECT Q4.$C0 AS "deptno"
, Q4.$C1 AS "deptname"
, Q4.$C2 AS "avg_time"
FROM
(SELECT Q3.$C3 AS $C0
, Q3.$C2 AS $C1
, DEC((Q3.$C1 / Q3.$C0),5,2) AS $C2
FROM (SELECT SUM(Q2.$C2) AS $C0
, SUM(Q2.$C3) AS $C1
, Q2.$C0 AS $C2
, Q2.$C1 AS $C3
FROM (SELECT Q1.deptname AS $C0
, Q1.workdept AS $C1
, Q1.num_time AS $C2
, Q1.sum_time AS $C3
FROM dpt_emp_act_sumry AS Q1
WHERE (Q1.firstnme LIKE '%S%')
AND (Q1.DEPTNAME LIKE '%S%')
) AS Q2
GROUP BY Q2.$C1
, Q2.$C0
) AS Q3
) AS Q4
ORDER BY Q4.$C2 DESC;
14.4.11. Indexes on Materialized Query Tables
To really make things fly, one can add indexes to the materialized query table columns. Db2 will then use these indexes to locate the required data. Certain restrictions apply:
-
Unique indexes are not allowed.
-
The materialized query table must not be in a "check pending" status when the index is defined. Run a refresh to address this problem.
Below are some indexes for the DPT_EMP_ACT_SUMRY table that was defined above:
CREATE INDEX dpt_emp_act_sumx1
ON dpt_emp_act_sumry ( workdept
, deptname
, empno
, firstnme);
CREATE INDEX dpt_emp_act_sumx2
ON dpt_emp_act_sumry (num_rows);
The next query will use the first index (i.e. on WORKDEPT):
SELECT d.deptno
, d.deptname
, e.empno
, e.firstnme
, INT(AVG(a.emptime)) AS avg_time
FROM department d
, employee e
, emp_act a
WHERE d.deptno = e.workdept
AND e.empno = a.empno
AND d.deptno LIKE 'D%'
GROUP BY d.deptno
, d.deptname
, e.empno
, e.firstnme
ORDER BY 1, 2, 3, 4;
The next query will use the second index (i.e. on NUM_ROWS):
SELECT d.deptno
, d.deptname
, e.empno
, e.firstnme
, COUNT(*) AS #acts
FROM department d
, employee e
, emp_act a
WHERE d.deptno = e.workdept
AND e.empno = a.empno
GROUP BY d.deptno
, d.deptname
, e.empno
, e.firstnme
HAVING COUNT(*) > 4
ORDER BY 1, 2, 3, 4;
14.5. Organizing by Dimensions
The following materialized query table is organized (clustered) by the two columns that are referred to in the GROUP BY. Under the covers, Db2 will also create a dimension index on each column, and a block index on both columns combined:
CREATE TABLE emp_sum AS
(SELECT workdept
, job
, SUM(salary) AS sum_sal
, COUNT(*) AS #emps
, GROUPING(workdept) AS grp_dpt
, GROUPING(job) AS grp_job
FROM employee
GROUP BY CUBE( workdept
, job)
)
DATA INITIALLY DEFERRED REFRESH DEFERRED
ORGANIZE BY DIMENSIONS (workdept, job)
IN tsempsum;
Multi-dimensional tables may perform very poorly when created in the default tablespace, or in a system-maintained tablespace. Use a database-maintained tablespace with the right extent size, and/or run the Db2EMPFA command. |
Don’t forget to run RUNSTATS!
14.5.1. Using Staging Tables
A staging table can be used to incrementally maintain a materialized query table that has been defined refresh deferred. Using a staging table can result in a significant performance saving (during the refresh) if the source table is very large, and is not changed very often.
To use a staging table, the SQL statement used to define the target materialized query table must follow the rules that apply for a table that is defined refresh immediate even though it is defined refresh deferred. |
The staging table CREATE statement has the following components:
-
The name of the staging table.
-
A list of columns (with no attributes) in the target materialized query table. The column names do not have to match those in the target table.
-
Either two or three additional columns with specific names- as provided by Db2.
-
The name of the target materialized query table.
To illustrate, below is a typical materialized query table:
CREATE TABLE emp_sumry AS
(SELECT workdept AS dept
, COUNT(*) AS #rows
, COUNT(salary) AS #sal
, SUM(salary) AS sum_sal
FROM employee emp
GROUP BY emp.workdept
) DATA INITIALLY DEFERRED REFRESH DEFERRED;
Here is a staging table for the above:
CREATE TABLE emp_sumry_s
( dept
, num_rows
, num_sal
, sum_sal
, GLOBALTRANSID
, GLOBALTRANSTIME
) FOR emp_sumry PROPAGATE IMMEDIATE;
Additional Columns
The two, or three, additional columns that every staging table must have are as follows:
-
GLOBALTRANSID: The global transaction ID for each propagated row.
-
GLOBALTRANSTIME: The transaction timestamp
-
OPERATIONTYPE: The operation type (i.e. insert, update, or delete). This column is needed if the target materialized query table does not contain a GROUP BY statement.
Using a Staging Table
To activate the staging table one must first use the SET INTEGRITY command to remove the check pending flag, and then do a full refresh of the target materialized query table. After this is done, the staging table will record all changes to the source table. Use the refresh incremental command to apply the changes recorded in the staging table to the target materialized query table.
SET INTEGRITY FOR emp_sumry_s STAGING IMMEDIATE UNCHECKED;
REFRESH TABLE emp_sumry;
<< make changes to the source table (i.e. employee) >>
REFRESH TABLE emp_sumry INCREMENTAL;
A multi-row update (or insert, or delete) uses the same CURRENT TIMESTAMP for all rows changed, and for all invoked triggers. Therefore, the #CHANGING_SQL field is only incremented when a new timestamp value is detected.
15. Identity Columns and Sequences
Imagine that one has an INVOICE table that records invoices generated. Also imagine that one wants every new invoice that goes into this table to get an invoice number value that is part of a unique and unbroken sequence of ascending values - assigned in the order that the invoices are generated. So if the highest invoice number is currently 12345, then the next invoice will get 12346, and then 12347, and so on. There are three ways to do this, up to a point:
-
Use an identity column, which generates a unique value per row in a table.
-
Use a sequence, which generates a unique value per one or more tables.
-
Do it yourself, using an insert trigger to generate the unique values.
You may need to know what values were generated during each insert. There are several ways to do this:
-
For all of the above techniques, embed the insert inside a select statement (see Selecting identity column values inserted). This is probably the best solution.
-
For identity columns, use the IDENTITY_VAL_LOCAL function (see IDENTITY_VAL_LOCAL).
-
For sequences, make a NEXTVAL or PREVVAL call (see NEXTVAL and PREVVAL - Usage Notes).
Living With Gaps
The only way that one can be absolutely certain not to have a gap in the sequence of values generated is to create your own using an insert trigger. However, this solution is probably the least efficient of those listed here, and it certainly has the least concurrency. There is almost never a valid business reason for requiring an unbroken sequence of values. So the best thing to do, if your users ask for such a feature, is to beat them up.
Living With Sequence Errors
For efficiency reasons, identity column and sequence values are usually handed out (to users doing inserts) in block of values, where the block size is defined using the CACHE option. If a user inserts a row, and then dithers for a bit before inserting another, it is possible that some other user (with a higher value) will insert first. In this case, the identity column or sequence value will be a good approximation of the insert sequence, but not right on. If the users need to know the precise order with which rows were inserted, then either set the cache size to one, which will cost, or include a current timestamp value.
15.1. Identity Columns
One can define a column in a Db2 table as an "identity column". This column, which must be numeric (note: fractional fields not allowed), will be incremented by a fixed constant each time a new row is inserted.
Below is an example of a typical invoice table that uses an identity column that starts at one, and then goes ever upwards:
CREATE TABLE invoice_data
( invoice# INTEGER NOT NULL
GENERATED ALWAYS AS IDENTITY
( START WITH 1
, INCREMENT BY 1
, NO MAXVALUE
, NO CYCLE,
ORDER)
, sale_date DATE NOT NULL
, customer_id CHAR(20) NOT NULL
, product_id INTEGER NOT NULL
, quantity INTEGER NOT NULL
, price DECIMAL(18, 2) NOT NULL
, PRIMARY KEY (invoice#));
15.2. Rules and Restrictions
Identity columns come in one of two general flavors:
-
The value is always generated by Db2.
-
The value is generated by Db2 only if the user does not provide a value (i.e. by default). This configuration is typically used when the input is coming from an external source (e.g. data propagation).
15.3. Rules
-
There can only be one identity column per table.
-
The field cannot be updated if it is defined "generated always".
-
The column type must be numeric and must not allow fractional values. Any integer type is OK. Decimal is also fine, as long as the scale is zero. Floating point is a no-no.
-
The identity column value is generated before any BEFORE triggers are applied. Use a trigger transition variable to see the value.
-
A unique index is not required on the identity column, but it is a good idea. Certainly, if the value is being created by Db2, then a non-unique index is a fairly stupid idea.
-
Unlike triggers, identity column logic is invoked and used during a LOAD. However, a load-replace will not reset the identity column value. Use the RESTART command (see below) to do this. An identity column is not affected by a REORG.
15.4. Syntax Notes
START WITH defines the start value, which can be any valid integer value. If no start value is provided, then the default is the MINVALUE for ascending sequences, and the MAXVALUE for descending sequences. If this value is also not provided, then the default is 1.
INCREMENT BY defines the interval between consecutive values. This can be any valid integer value, though using zero is pretty silly. The default is 1.
MINVALUE defines (for ascending sequences) the value that the sequence will start at if no start value is provided. It is also the value that an ascending sequence will begin again at after it reaches the maximum and loops around. If no minimum value is provided, then after reaching the maximum the sequence will begin again at the start value. If that is also not defined, then the sequence will begin again at 1, which is the default start value. For descending sequences, it is the minimum value that will be used before the sequence loops around, and starts again at the maximum value.
MAXVALUE defines (for ascending sequences) the value that a sequence will stop at, and then go back to the minimum value. For descending sequences, it is the start value (if no start value is provided), and also the restart value - if the sequence reaches the minimum and loops around.
CYCLE defines whether the sequence should cycle about when it reaches the maximum value (for an ascending sequences), or whether it should stop. The default is no cycle.
CACHE defines whether or not to allocate sequences values in chunks, and thus to save on log writes. The default is no cache, which means that every row inserted causes a log write (to save the current value). If a cache value (from 2 to 20) is provided, then the new values are assigned to a common pool in blocks. Each insert user takes from the pool, and only when all of the values are used is a new block (of values) allocated and a log write done. If the table is deactivated, either normally or otherwise, then the values in the current block are discarded, resulting in gaps in the sequence. Gaps in the sequence of values also occur when an insert is subsequently rolled back, so they cannot be avoided. But don’t use the cache if you want to try and avoid them.
ORDER defines whether all new rows inserted are assigned a sequence number in the order that they were inserted. The default is no, which means that occasionally a row that is inserted after another may get a slightly lower sequence number. This is the default.
15.5. Identity Column Examples
The following example uses all of the defaults to start an identity column at one, and then to go up in increments of one. The inserts will eventually die when they reach the maximum allowed value for the field type (i.e. for small integer = 32K).
CREATE TABLE test_data
( key# SMALLINT NOT NULL
GENERATED ALWAYS AS IDENTITY
, dat1 SMALLINT NOT NULL
, ts1 TIMESTAMP NOT NULL
, PRIMARY KEY(key#));
-- KEY# FIELD - VALUES ASSIGNED
-- 1 2 3 4 5 6 7 8 9 10 11 etc.
The next example defines an identity column that goes down in increments of -3:
CREATE TABLE test_data
( key# SMALLINT NOT NULL
GENERATED ALWAYS AS IDENTITY
( START WITH 6
, INCREMENT BY -3
, NO CYCLE
, NO CACHE
, ORDER)
, dat1 SMALLINT NOT NULL
, ts1 TIMESTAMP NOT NULL
, PRIMARY KEY(key#));
-- KEY# FIELD - VALUES ASSIGNED
-- 6 3 0 -3 -6 -9 -12 -15 etc.
The next example, which is amazingly stupid, goes nowhere fast. A primary key cannot be defined on this table:
CREATE TABLE test_data
(key# SMALLINT NOT NULL
GENERATED ALWAYS AS IDENTITY
( START WITH 123
, MAXVALUE 124
, INCREMENT BY 0
, NO CYCLE
, NO ORDER)
, dat1 SMALLINT NOT NULL
, ts1 TIMESTAMP NOT NULL);
-- KEY# VALUES ASSIGNED
-- 123 123 123 123 123 123 etc.
The next example uses every odd number up to the maximum (i.e. 6), then loops back to the minimum value, and goes through the even numbers, ad-infinitum:
CREATE TABLE test_data
( key# SMALLINT NOT NULL
GENERATED ALWAYS AS IDENTITY
( START WITH 1
, INCREMENT BY 2
, MAXVALUE 6
, MINVALUE 2
, CYCLE
, NO CACHE
, ORDER)
, dat1 SMALLINT NOT NULL
, ts1 TIMESTAMP NOT NULL);
-- KEY# VALUES ASSIGNED0
-- 1 3 5 2 4 6 2 4 6 2 4 6 etc.
15.6. Usage Examples
Below is the DDL for a simplified invoice table where the primary key is an identity column. Observe that the invoice# is always generated by Db2:
CREATE TABLE invoice_data
( invoice# INTEGER NOT NULL
GENERATED ALWAYS AS IDENTITY
( START WITH 100
, INCREMENT BY 1
, NO CYCLE
, ORDER)
, sale_date DATE NOT NULL
, customer_id CHAR(20) NOT NULL
, product_id INTEGER NOT NULL
, quantity INTEGER NOT NULL
, price DECIMAL(18, 2) NOT NULL
, PRIMARY KEY (invoice#));
One cannot provide a value for the invoice# when inserting into the above table. Therefore, one must either use a default placeholder, or leave the column out of the insert. An example of both techniques is given below. The second insert also selects the generated values:
INSERT INTO invoice_data
VALUES (DEFAULT, '2001-11-22', 'ABC', 123, 100, 10);
SELECT invoice#
FROM FINAL TABLE
(INSERT INTO invoice_data
(sale_date, customer_id, product_id, quantity,price)
VALUES ('2002-11-22', 'DEF', 123, 100, 10)
, ('2003-11-22', 'GHI', 123, 100, 10));
ANSWER
INVOICE# |
---|
101 |
102 |
Below is the state of the table after the above two inserts: .Invoice table, after inserts
INVOICE# | SALE_DATE | CUSTOMER_ID | PRODUCT_ID | QUANTITY | PRICE |
---|---|---|---|---|---|
100 |
2001-11-22 |
ABC |
123 |
100 |
10.00 |
101 |
2002-11-22 |
DEF |
123 |
100 |
10.00 |
102 |
2003-11-22 |
GHI |
123 |
100 |
10.00 |
15.7. Altering Identity Column Options
Imagine that the application is happily collecting invoices in the above table, but your silly boss is unhappy because not enough invoices, as measured by the ever-ascending invoice# value, are being generated per unit of time. We can improve things without actually fixing any difficult business problems by simply altering the invoice# current value and the increment using the ALTER TABLE … RESTART command:
ALTER TABLE invoice_data
ALTER COLUMN invoice#
RESTART WITH 1000
SET INCREMENT BY 2;
Now imagine that we insert two more rows thus:
INSERT INTO invoice_data
VALUES (DEFAULT, '2004-11-24', 'XXX', 123, 100, 10)
, (DEFAULT, '2004-11-25', 'YYY', 123, 100, 10);
Our mindless management will now see this data:
INVOICE# | SALE_DATE | CUSTOMER_ID | PRODUCT_ID | QUANTITY | PRICE |
---|---|---|---|---|---|
100 |
2001-11-22 |
ABC |
123 |
100 |
10.00 |
101 |
2002-11-22 |
DEF |
123 |
100 |
10.00 |
102 |
2003-11-22 |
GHI |
123 |
100 |
10.00 |
1000 |
2004-11-24 |
XXX |
123 |
100 |
10.00 |
1002 |
2004-11-25 |
YYY |
123 |
100 |
10.00 |
Restarting the identity column start number to a lower number, or to a higher number if the increment is a negative value, can result in the column getting duplicate values. This can also occur if the increment value is changed from positive to negative, or vice-versa. If no value is provided for the restart option, the sequence restarts at the previously defined start value.
15.8. Gaps in Identity Column Values
If an identity column is generated always, and no cache is used, and the increment value is 1, then there will usually be no gaps in the sequence of assigned values. But gaps can occur if an insert is subsequently rolled out instead of committed. In the following example, there will be no row in the table with customer number "1" after the rollback:
CREATE TABLE customers
( cust# INTEGER NOT NULL
GENERATED ALWAYS AS IDENTITY (NO CACHE)
, cname CHAR(10) NOT NULL
, ctype CHAR(03) NOT NULL
, PRIMARY KEY (cust#));
COMMIT;
SELECT cust#
FROM FINAL TABLE
(INSERT INTO customers
VALUES (DEFAULT, 'FRED', 'XXX'));
ROLLBACK;
ANSWER
CUST# |
---|
1 |
SELECT cust#
FROM FINAL TABLE
(INSERT INTO customers
VALUES (DEFAULT, 'FRED', 'XXX'));
COMMIT;
ANSWER
CUST# |
---|
2 |
15.9. Find Gaps in Values
The following query can be used to list the missing values in a table. It starts by getting the minimum and maximum values. It next generates every value in between. Finally, it checks the generated values against the source tables. Non-matches are selected.
WITH generate_values (min_val, max_val, num_val, cur_val) AS
(SELECT MIN(dat1)
, MAX(dat1)
, COUNT(*)
, MIN(dat1)
FROM test_data td1
UNION ALL
SELECT min_val
, max_val
, num_val
, cur_val + 1
FROM generate_values gv1
WHERE cur_val < max_val
)
SELECT *
FROM generate_values gv2
WHERE NOT EXISTS
(SELECT *
FROM test_data td2
WHERE td2.dat1 = cur_val)
ORDER BY cur_val;
INPUT
DAT1 |
---|
1 |
2 |
3 |
4 |
6 |
7 |
9 |
10 |
ANSWER
MIN_VAL | MAX_VAL | NUM_VAL | CUR_VAL |
---|---|---|---|
1 |
10 |
8 |
5 |
1 |
10 |
8 |
8 |
The above query may be inefficient if there is no suitable index on the DAT1 column. The next query gets around this problem by using an EXCEPT instead of a sub-query:
WITH generate_values (min_val, max_val, num_val, cur_val) AS
(SELECT MIN(dat1)
, MAX(dat1)
, COUNT(*)
, MIN(dat1)
FROM test_data td1
UNION ALL
SELECT min_val
, max_val
, num_val
, cur_val + 1
FROM generate_values gv1
WHERE cur_val < max_val)
SELECT cur_val
FROM generate_values gv2
EXCEPT ALL
SELECT dat1
FROM test_data td2
ORDER BY 1;
INPUT
DAT1 |
---|
1 |
2 |
3 |
4 |
6 |
7 |
9 |
10 |
ANSWER
CUR_VAL |
---|
5 |
8 |
The next query uses a totally different methodology. It assigns a rank to every value, and then looks for places where the rank and value get out of step:
WITH assign_ranks AS
(SELECT dat1
, DENSE_RANK() OVER(ORDER BY dat1) AS rank#
FROM test_data)
, locate_gaps AS
(SELECT dat1 - rank# AS diff
, min(dat1) AS min_val
, max(dat1) AS max_val
, ROW_NUMBER() OVER(ORDER BY dat1 - rank#) AS gap#
FROM assign_ranks ar1
GROUP BY dat1 - rank#)
SELECT lg1.gap# AS gap#
, lg1.max_val AS prev_val
, lg2.min_val AS next_val
, lg2.min_val - lg1.max_val AS diff
FROM locate_gaps lg1
, locate_gaps lg2
WHERE lg2.gap# = lg1.gap# + 1
ORDER BY lg1.gap#;
ANSWER
GAP# | PREV_VAL | NEXT_VAL | DIFF |
---|---|---|---|
1 |
4 |
6 |
2 |
2 |
7 |
9 |
2 |
15.9.1. IDENTITY_VAL_LOCAL Function
There are two ways to find out what values were generated when one inserted a row into a table with an identity column:
-
Embed the insert within a select statement (see Selecting identity column values inserted).
-
Call the IDENTITY_VAL_LOCAL function.
Certain rules apply to IDENTITY_VAL_LOCAL function usage:
-
The value returned from is a decimal (31.0) field.
-
The function returns null if the user has not done a single-row insert in the current unit of work. Therefore, the function has to be invoked before one does a commit. Having said this, in some versions of Db2 it seems to work fine after a commit.
-
If the user inserts multiple rows into table(s) having identity columns in the same unit of work, the result will be the value obtained from the last single-row insert. The result will be null if there was none.
-
Multiple-row inserts are ignored by the function. So if the user first inserts one row, and then separately inserts two rows (in a single SQL statement), the function will return the identity column value generated during the first insert.
-
The function cannot be called in a trigger or SQL function. To get the current identity column value in an insert trigger, use the trigger transition variable for the column. The value, and thus the transition variable, is defined before the trigger is begun.
-
If invoked inside an insert statement (i.e. as an input value), the value will be taken from the most recent (previous) single-row insert done in the same unit of work. The result will be null if there was none.
-
The value returned by the function is unpredictable if the prior single-row insert failed. It may be the value from the insert before, or it may be the value given to the failed insert.
-
The function is non-deterministic, which means that the result is determined at fetch time (i.e. not at open) when used in a cursor. So if one fetches a row from a cursor, and then does an insert, the next fetch may get a different value from the prior.
-
The value returned by the function may not equal the value in the table - if either a trigger or an update has changed the field since the value was generated. This can only occur if the identity column is defined as being "generated by default". An identity column that is "generated always" cannot be updated.
-
When multiple users are inserting into the same table concurrently, each will see their own most recent identity column value. They cannot see each other’s.
If the above sounds unduly complex, it is because it is. It is often much easier to simply get the values by embedding the insert inside a select:
SELECT MIN(cust#) AS minc
, MAX(cust#) AS maxc
, COUNT(*) AS rows
FROM FINAL TABLE
(INSERT INTO customers
VALUES (DEFAULT, 'FRED', 'xxx')
, (DEFAULT, 'DAVE', 'yyy')
, (DEFAULT, 'JOHN', 'zzz'));
ANSWER
MINC | MAXC | ROWS |
---|---|---|
3 |
5 |
3 |
Below are two examples of the function in use. Observe that the second invocation (done after the commit) returned a value, even though it is supposed to return null:
CREATE TABLE invoice_table
( invoice# INTEGER NOT NULL
GENERATED ALWAYS AS IDENTITY
, sale_date DATE NOT NULL
, customer_id CHAR(20) NOT NULL
, product_id INTEGER NOT NULL
, quantity INTEGER NOT NULL
, price DECIMAL(18,2) NOT NULL
, PRIMARY KEY (invoice#));
COMMIT;
INSERT INTO invoice_table
VALUES (DEFAULT, '2000-11-22', 'ABC', 123, 100, 10);
WITH temp (id) AS
(VALUES (IDENTITY_VAL_LOCAL()))
SELECT *
FROM temp; --> ANSWER: ID = 1
COMMIT;
WITH temp (id) AS
(VALUES (IDENTITY_VAL_LOCAL()))
SELECT *
FROM temp; --> ANSWER: ID = 1
In the next example, two separate inserts are done on the table defined above. The first inserts a single row, and so sets the function value to "2". The second is a multi-row insert, and so is ignored by the function:
INSERT INTO invoice_table
VALUES (DEFAULT, '2000-11-23', 'ABC', 123, 100, 10);
INSERT INTO invoice_table
VALUES (DEFAULT, '2000-11-24', 'ABC', 123, 100, 10)
, (DEFAULT, '2000-11-25', 'ABC', 123, 100, 10);
SELECT invoice# AS inv#
, sale_date
, IDENTITY_VAL_LOCAL() AS id
FROM invoice_table
ORDER BY 1;
COMMIT;
ANSWER
INV# | SALE_DATE | ID |
---|---|---|
1 |
11/22/2000 |
2 |
2 |
11/23/2000 |
2 |
3 |
11/24/2000 |
2 |
4 |
11/25/2000 |
2 |
One can also use the function to get the most recently inserted single row by the current user:
SELECT invoice# AS inv#
, sale_date
, IDENTITY_VAL_LOCAL() AS id
FROM invoice_table
WHERE id = IDENTITY_VAL_LOCAL();
ANSWER
INV# | SALE_DATE | ID |
---|---|---|
2 |
11/23/2000 |
2 |
15.10. Sequences
A sequence is almost the same as an identity column, except that it is an object that exists outside of any particular table.
CREATE SEQUENCE fred
AS DECIMAL(31)
START WITH 100
INCREMENT BY 2
NO MINVALUE
NO MAXVALUE
NO CYCLE
CACHE 20
ORDER;
-- SEQ# VALUES ASSIGNED
-- 100 102 104 106 etc.
The options and defaults for a sequence are exactly the same as those for an identity column (see Rules and Restrictions). Likewise, one can alter a sequence in much the same way as one would alter the status of an identity column:
ALTER SEQUENCE fred
RESTART WITH -55
INCREMENT BY -5
MINVALUE -1000
MAXVALUE +1000
NO CACHE
NO ORDER
CYCLE;
-- SEQ# VALUES ASSIGNED
-- -55 -60 -65 -70 etc.
The only sequence attribute that one cannot change with the ALTER command is the field type that is used to hold the current value.
15.10.1. Constant Sequence
If the increment is zero, the sequence will stay whatever value one started it with until it is altered. This can be useful if wants to have a constant that can be globally referenced:
CREATE SEQUENCE biggest_sale_to_date
AS INTEGER
START WITH 345678
INCREMENT BY 0;
-- SEQ# VALUES ASSIGNED
-- 345678, 345678, etc.
15.10.2. Getting the Sequence Value
There is no concept of a current sequence value. Instead one can either retrieve the next or the previous value (if there is one). And any reference to the next value will invariably cause the sequence to be incremented. The following example illustrates this:
CREATE SEQUENCE fred;
COMMIT;
WITH temp1 (n1) AS
(VALUES 1
UNION ALL
SELECT n1 + 1
FROM temp1
WHERE n1 < 5
)
SELECT NEXTVAL FOR fred AS seq#
FROM temp1;
ANSWER
SEQ# |
---|
1 |
2 |
3 |
4 |
5 |
15.10.3. NEXTVAL and PREVVAL - Usage Notes
-
One retrieves the next or previous value using a "NEXTVAL FOR sequence-name", or a "PREVVAL for sequence-name" call.
-
A NEXTVAL call generates and returns the next value in the sequence. Thus, each call will consume the returned value. This remains true even if the statement that did the retrieval subsequently fails or is rolled back.
-
A PREVVAL call returns the most recently generated value for the specified sequence for the current connection. Unlike when getting the next value, getting the prior value does not alter the state of the sequence, so multiple calls can retrieve the same value.
-
If no NEXTVAL reference (to the target sequence) has been made for the current connection, any attempt to get the PREVVAL will result in a SQL error.
15.10.4. NEXTVAL and PREVVAL - Usable Statements
-
SELECT INTO statement (within the select part), as long as there is no DISTINCT, GROUP BY, UNION, EXECPT, or INTERSECT.
-
INSERT statement - with restrictions.
-
UPDATE statement - with restrictions.
-
SET host variable statement.
15.10.5. NEXTVAL - Usable Statements
-
A trigger.
15.10.6. NEXTVAL and PREVVAL - Not Allowed In
-
DELETE statement.
-
Join condition of a full outer join.
-
Anywhere in a CREATE TABLE or CREATE VIEW statement.
15.10.7. NEXTVAL - Not Allowed In
-
CASE expression
-
Join condition of a join.
-
Parameter list of an aggregate function.
-
SELECT statement where there is an outer select that contains a DISTINCT, GROUP BY, UNION, EXCEPT, or INTERSECT.
-
Most sub-queries.
15.10.8. PREVVAL - Not Allowed In
-
A trigger.
There are many more usage restrictions, but you presumably get the picture. See the Db2 SQL Reference for the complete list.
15.11. Usage Examples
Below a sequence is defined, then various next and previous values are retrieved:
CREATE SEQUENCE fred;
COMMIT;
WITH temp1 (prv) AS
(VALUES (PREVVAL FOR fred))
SELECT *
FROM temp1; --> PRV : <error>
WITH temp1 (nxt) AS
(VALUES (NEXTVAL FOR fred))
SELECT *
FROM temp1; --> NXT: 1
WITH temp1 (prv) AS
(VALUES (PREVVAL FOR fred))
SELECT *
FROM temp1; --> PRV: 1
WITH temp1 (n1) AS
(VALUES 1
UNION ALL
SELECT n1 + 1
FROM temp1
WHERE n1 < 5
)
SELECT NEXTVAL FOR fred AS nxt
, PREVVAL FOR fred AS prv
FROM temp1;
NXT | PRV |
---|---|
2 |
1 |
3 |
1 |
4 |
1 |
5 |
1 |
6 |
1 |
One does not actually have to fetch a NEXTVAL result in order to increment the underlying sequence. In the next example, some of the rows processed are thrown away halfway thru the query, but their usage still affects the answer (of the subsequent query):
CREATE SEQUENCE fred;
COMMIT;
WITH temp1 AS
(SELECT id
, NEXTVAL FOR fred AS nxt
FROM staff
WHERE id < 100
)
SELECT *
FROM temp1
WHERE id = 50 + (nxt * 0);
ID | NXT |
---|---|
50 |
5 |
WITH temp1 (nxt, prv) AS
(VALUES (NEXTVAL FOR fred
, PREVVAL FOR fred))
SELECT *
FROM temp1;
NXT | PRV |
---|---|
10 |
9 |
The somewhat funky predicate at the end of the first query above prevents Db2 from stopping the nested-table-expression when it gets to "id = 50". If this were to occur, the last query above would get a next value of 6, and a previous value of 5. |
15.12. Multi-table Usage
Imagine that one wanted to maintain a unique sequence of values over multiple tables. One can do this by creating a before insert trigger on each table that replaces whatever value the user provides with the current one from a common sequence. Below is an example:
CREATE SEQUENCE cust#
START WITH 1
INCREMENT BY 1
NO MAXVALUE
NO CYCLE
ORDER;
CREATE TABLE us_customer
( cust# INTEGER NOT NULL
, cname CHAR(10) NOT NULL
, frst_sale DATE NOT NULL
, #sales INTEGER NOT NULL
, PRIMARY KEY (cust#));
CREATE TRIGGER us_cust_ins
NO CASCADE BEFORE INSERT ON us_customer
REFERENCING NEW AS nnn
FOR EACH ROW MODE Db2SQL
SET nnn.cust# = NEXTVAL FOR cust#;
CREATE TABLE intl_customer
( cust# INTEGER NOT NULL
, cname CHAR(10) NOT NULL
, frst_sale DATE NOT NULL
, #sales INTEGER NOT NULL
, PRIMARY KEY (cust#));
CREATE TRIGGER intl_cust_ins
NO CASCADE BEFORE INSERT ON intl_customer
REFERENCING NEW AS nnn
FOR EACH ROW MODE Db2SQL
SET nnn.cust# = NEXTVAL FOR cust#;
If we now insert some rows into the above tables, we shall find that customer numbers are assigned in the correct order, thus:
SELECT cust#
, cname
FROM FINAL TABLE
(INSERT INTO us_customer (cname, frst_sale, #sales)
VALUES ('FRED', '2002-10-22', 1)
, ('JOHN', '2002-10-23', 1));
SELECT cust#
, cname
FROM FINAL TABLE
(INSERT INTO intl_customer (cname, frst_sale, #sales)
VALUES ('SUE', '2002-11-12', 2)
, ('DEB', '2002-11-13', 2));
ANSWERS
CUST# | CNAME |
---|---|
1 |
FRED |
2 |
JOHN |
CUST# | CNAME |
---|---|
3 |
SUE |
4 |
DEB |
One of the advantages of a standalone sequence over a functionally similar identity column is that one can use a PREVVAL expression to get the most recent value assigned (to the user), even if the previous usage was during a multi-row insert. Thus, after doing the above inserts, we can run the following query:
WITH temp (prev) AS
(VALUES (PREVVAL FOR cust#))
SELECT *
FROM temp;
ANSWER
PREV |
---|
4 |
The following does the same as the above, but puts the result in a host variable:
VALUES PREVVAL FOR CUST# INTO :host-var
As with identity columns, the above result will not equal what is actually in the table(s) – if the most recent insert was subsequently rolled back.
15.13. Counting Deletes
In the next example, two sequences are created: One records the number of rows deleted from a table, while the other records the number of delete statements run against the same:
CREATE SEQUENCE delete_rows
START WITH 1
INCREMENT BY 1
NO MAXVALUE
NO CYCLE
ORDER;
CREATE SEQUENCE delete_stmts
START WITH 1
INCREMENT BY 1
NO MAXVALUE
NO CYCLE
ORDER;
CREATE TABLE customer
( cust# INTEGER NOT NULL
, cname CHAR(10) NOT NULL
, frst_sale DATE NOT NULL
, #sales INTEGER NOT NULL
, PRIMARY KEY (cust#));
CREATE TRIGGER cust_del_rows
AFTER DELETE ON customer
FOR EACH ROW MODE Db2SQL
WITH temp1 (n1) AS (VALUES(1))
SELECT NEXTVAL FOR delete_rows
FROM temp1;
CREATE TRIGGER cust_del_stmts
AFTER DELETE ON customer
FOR EACH STATEMENT MODE Db2SQL
WITH temp1 (n1) AS (VALUES(1))
SELECT NEXTVAL FOR delete_stmts
FROM temp1;
Be aware that the second trigger will be run, and thus will update the sequence, regardless of whether a row was found to delete or not.
15.14. Identity Columns vs. Sequences - a Comparison
First to compare the two types of sequences:
-
Only one identity column is allowed per table, whereas a single table can have multiple sequences and/or multiple references to the same sequence.
-
Identity column sequences cannot span multiple tables. Sequences can.
-
Sequences require triggers to automatically maintain column values (e.g. during inserts) in tables. Identity columns do not.
-
Sequences can be incremented during inserts, updates, deletes (via triggers), or selects, whereas identity columns only get incremented during inserts.
-
Sequences can be incremented (via triggers) once per row, or once per statement. Identity columns are always updated per row inserted.
-
Sequences can be dropped and created independent of any tables that they might be used to maintain values in. Identity columns are part of the table definition.
-
Identity columns are supported by the load utility. Trigger induced sequences are not.
For both types of sequence, one can get the current value by embedding the DML statement inside a select (e.g. see Selecting identity column values inserted). Alternatively, one can use the relevant expression to get the current status. These differ as follows:
-
The IDENTITY_VAL_LOCAL function returns null if no inserts to tables with identity columns have been done by the current user. In an equivalent situation, the PREVVAL expression gets a nasty SQL error.
-
The IDENTITY_VAL_LOCAL function ignores multi-row inserts (without telling you). In a similar situation, the PREVVAL expression returns the last value generated.
-
One cannot tell to which table an IDENTITY_VAL_LOCAL function result refers to. This can be a problem in one insert invokes another insert (via a trigger), which puts are row in another table with its own identity column. By contrast, in the PREVVAL function one explicitly identifies the sequence to be read.
-
There is no equivalent of the NEXTVAL expression for identity columns.
15.15. Roll Your Own
If one really, really, needs to have a sequence of values with no gaps, then one can do it using an insert trigger, but there are costs, in processing time, concurrency, and functionality. To illustrate, consider the following table:
CREATE TABLE sales_invoice
( invoice# INTEGER NOT NULL
, sale_date DATE NOT NULL
, customer_id CHAR(20) NOT NULL
, product_id INTEGER NOT NULL
, quantity INTEGER NOT NULL
, price DECIMAL(18, 2) NOT NULL
, PRIMARY KEY (invoice#));
The following trigger will be invoked before each row is inserted into the above table. It sets the new invoice# value to be the current highest invoice# value in the table, plus one:
CREATE TRIGGER sales_insert
NO CASCADE BEFORE
INSERT ON sales_invoice
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
SET nnn.invoice# =
(SELECT COALESCE(MAX(invoice#),0) + 1
FROM sales_invoice);
The good news about the above setup is that it will never result in gaps in the sequence of values. In particular, if a newly inserted row is rolled back after the insert is done, the next insert will simply use the same invoice# value. But there is also bad news: * Only one user can insert at a time, because the select (in the trigger) needs to see the highest invoice# in the table in order to complete. * Multiple rows cannot be inserted in a single SQL statement (i.e. a mass insert). The trigger is invoked before the rows are actually inserted, one row at a time, for all rows. Each row would see the same, already existing, high invoice#, so the whole insert would die due to a duplicate row violation.
There may be a tiny, tiny chance that if two users were to begin an insert at exactly the same time that they would both see the same high invoice# (in the before trigger), and so the last one to complete (i.e. to add a pointer to the unique invoice# index) would get a duplicate-row violation.
Below are some inserts to the above table. Ignore the values provided in the first field – they are replaced in the trigger. And observe that the third insert is rolled out:
INSERT INTO sales_invoice VALUES (0, '2001-06-22' ,'ABC', 123, 10, 1);
INSERT INTO sales_invoice VALUES (0, '2001-06-23' ,'DEF', 453, 10, 1);
COMMIT;
INSERT INTO sales_invoice VALUES (0, '2001-06-24' ,'XXX', 888, 10, 1);
ROLLBACK;
INSERT INTO sales_invoice VALUES (0, '2001-06-25' ,'YYY', 999, 10, 1);
COMMIT;
ANSWER
INVOICE# | SALE_DATE | CUSTOMER_ID | PRODUCT_ID | QUANTITY | PRICE |
---|---|---|---|---|---|
1 |
06/22/2001 |
ABC |
123 |
10 |
1.00 |
2 |
06/23/2001 |
DEF |
453 |
10 |
1.00 |
3 |
06/25/2001 |
YYY |
999 |
10 |
1.00 |
15.16. Support Multi-row Inserts
The next design is more powerful in that it supports multi-row inserts, and also more than one table if desired. It requires that there be a central location that holds the current high-value. In the example below, this value will be in a row in a special control table. Every insert into the related data table will, via triggers, first update, and then query, the row in the control table.
15.16.1. Control Table
The following table has one row per sequence of values being maintained:
CREATE TABLE control_table
( table_name CHAR(18) NOT NULL
, table_nmbr INTEGER NOT NULL
, PRIMARY KEY (table_name));
Now to populate the table with some initial sequence# values:
INSERT INTO control_table VALUES ('invoice_table', 0);
INSERT INTO control_table VALUES ('2nd_data_tble', 0);
INSERT INTO control_table VALUES ('3rd_data_tble', 0);
15.16.2. Data Table
Our sample data table has two fields of interest:
-
The UNQVAL column will be populated, using a trigger, with a GENERATE_UNIQUE function output value. This is done before the row is actually inserted. Once the insert has completed, we will no longer care about or refer to the contents of this field.
-
The INVOICE# column will be populated, using triggers, during the insert process with a unique ascending value. However, for part of the time during the insert the field will have a null value, which is why it is defined as being both non-unique and allowing nulls.
CREATE TABLE invoice_table
( unqval CHAR(13) FOR BIT DATA NOT NULL
, invoice# INTEGER NOT NULL
, sale_date DATE NOT NULL
, customer_id CHAR(20) NOT NULL
, product_id INTEGER NOT NULL
, quantity INTEGER NOT NULL
, price DECIMAL(18, 2) NOT NULL
, PRIMARY KEY(unqval));
Two insert triggers are required: The first acts before the insert is done, giving each new row a unique UNQVAL value:
CREATE TRIGGER invoice1
NO CASCADE BEFORE INSERT ON invoice_table
REFERENCING NEW AS nnn
FOR EACH ROW MODE Db2SQL
SET nnn.unqval = GENERATE_UNIQUE()
, nnn.invoice# = NULL;
The second trigger acts after the row is inserted. It first increments the control table by one, then updates invoice# in the current row with the same value. The UNQVAL field is used to locate the row to be changed in the second update:
CREATE TRIGGER invoice2
AFTER INSERT ON invoice_table
REFERENCING NEW AS nnn
FOR EACH ROW MODE Db2SQL
BEGIN ATOMIC
UPDATE control_table
SET table_nmbr = table_nmbr + 1
WHERE table_name = 'invoice_table';
UPDATE invoice_table
SET invoice# =
(SELECT table_nmbr
FROM control_table
WHERE table_name = 'invoice_table')
WHERE unqval = nnn.unqval
AND invoice# IS NULL;
END
The above two actions must be in a single trigger. If they are in two triggers, mass inserts will not work correctly because the first trigger (i.e. update) would be run (for all rows), followed by the second trigger (for all rows). In the end, every row inserted by the mass-insert would end up with the same invoice# value. |
A final update trigger is required to prevent updates to the invoice# column:
CREATE TRIGGER invoice3
NO CASCADE BEFORE UPDATE OF invoice# ON invoice_table
REFERENCING OLD AS ooo
NEW AS nnn
FOR EACH ROW MODE Db2SQL
WHEN (ooo.invoice# <> nnn.invoice#)
SIGNAL SQLSTATE '71001' ('no updates allowed - you twit');
15.17. Design Comments
Though the above design works, it has certain practical deficiencies:
-
The single row in the control table is a point of contention, because only one user can update it at a time. One must therefore commit often (perhaps more often than one would like to) in order to free up the locks on this row. Therefore, by implication, this design puts one is at the mercy of programmers.
-
The two extra updates add a considerable overhead to the cost of the insert.
-
The invoice number values generated by AFTER trigger cannot be obtained by selecting from an insert statement (see Insert Examples). In fact, selecting from the FINAL TABLE will result in a SQL error. One has to instead select from the NEW TABLE, which returns the new rows before the AFTER trigger was applied.
As with ordinary sequences, this design enables one to have multiple tables referring to a single row in the control table, and thus using a common sequence.
16. Temporary Tables
16.1. Introduction
How one defines a temporary table depends in part upon how often, and for how long, one intends to use it:
-
Within a query, single use.
-
Within a query, multiple uses.
-
For multiple queries in one unit of work.
-
For multiple queries, over multiple units of work, in one thread.
16.2. Single Use in Single Statement
If one intends to use a temporary table just once, it can be defined as a nested table expression. In the following example, we use a temporary table to sequence the matching rows in the STAFF table by descending salary. We then select the 2nd through 3rd rows:
SELECT id
, salary
FROM (SELECT s.*
, ROW_NUMBER() OVER(ORDER BY salary DESC) AS sorder
FROM staff s
WHERE id < 200
) AS xxx
WHERE sorder BETWEEN 2 AND 3
ORDER BY id;
ANSWER
ID | SALARY |
---|---|
50 |
20659.80 |
140 |
21150.00 |
A fullselect in parenthesis followed by a correlation name (see above) is also called a nested table expression. |
Here is another way to express the same:
WITH xxx (id, salary, sorder) AS
(SELECT ID
, salary
, ROW_NUMBER() OVER(ORDER BY salary DESC) AS sorder
FROM staff
WHERE id < 200
)
SELECT id
, salary
FROM xxx
WHERE sorder BETWEEN 2 AND 3
ORDER BY id;
ANSWER
ID | SALARY |
---|---|
50 |
20659.80 |
140 |
21150.00 |
16.3. Multiple Use in Single Statement
Imagine that one wanted to get the percentage contribution of the salary in some set of rows in the STAFF table - compared to the total salary for the same. The only way to do this is to access the matching rows twice; Once to get the total salary (i.e. just one row), and then again to join the total salary value to each individual salary - to work out the percentage.
Selecting the same set of rows twice in a single query is generally unwise because repeating the predicates increases the likelihood of typos being made. In the next example, the desired rows are first placed in a temporary table. Then the sum salary is calculated and placed in another temporary table. Finally, the two temporary tables are joined to get the percentage:
WITH rows_wanted AS
(SELECT *
FROM staff
WHERE id < 100
AND UCASE(name) LIKE '%T%')
, sum_salary AS
(SELECT SUM(salary) AS sum_sal
FROM rows_wanted)
SELECT id
, name
, salary
, sum_sal
, INT((salary * 100) / sum_sal) AS pct
FROM rows_wanted
, sum_salary
ORDER BY id;
ANSWER
ID | NAME | SALARY | SUM_SAL | PCT |
---|---|---|---|---|
70 |
Rothman |
16502.83 |
34504.58 |
47 |
90 |
Koonitz |
18001.75 |
34504.58 |
52 |
16.3.1. Multiple Use in Multiple Statements
To refer to a temporary table in multiple SQL statements in the same thread, one has to define a declared global temporary table. An example follows:
DECLARE GLOBAL TEMPORARY TABLE session.fred
( dept SMALLINT NOT NULL
, avg_salary DEC(7, 2)NOT NULL
, num_emps SMALLINT NOT NULL)
ON COMMIT PRESERVE ROWS;
COMMIT;
INSERT INTO session.fred
SELECT dept
, AVG(salary)
, COUNT(*)
FROM staff
WHERE id > 200
GROUP BY dept;
COMMIT;
SELECT COUNT(*) AS cnt
FROM session.fred;
DELETE FROM session.fred
WHERE dept > 80;
SELECT *
FROM session.fred;
ANSWER#1
CNT |
---|
4 |
ANSWER#2
DEPT | AVG_SALARY | NUM_EMPS |
---|---|---|
10 |
20168.08 |
3 |
51 |
15161.43 |
3 |
66 |
17215.24 |
5 |
Unlike an ordinary table, a declared global temporary table is not defined in the Db2 catalogue. Nor is it sharable by other users. It only exists for the duration of the thread (or less) and can only be seen by the person who created it. For more information, see Declared Global Temporary Tables.
16.3.2. Temporary Tables - in Statement
Three general syntaxes are used to define temporary tables in a query:
-
Use a WITH phrase at the top of the query to define a common table expression.
-
Define a fullselect in the FROM part of the query.
-
Define a fullselect in the SELECT part of the query.
The following three queries, which are logically equivalent, illustrate the above syntax styles. Observe that the first two queries are explicitly defined as left outer joins, while the last one is implicitly a left outer join:
WITH staff_dept AS
(SELECT dept AS dept#
, MAX(salary) AS max_sal
FROM staff
WHERE dept < 50
GROUP BY dept
)
SELECT id
, dept
, salary
, max_sal
FROM staff
LEFT OUTER JOIN staff_dept
ON dept = dept#
WHERE name LIKE 'S%'
ORDER BY id;
ANSWER
ID | DEPT | SALARY | MAX_SAL |
---|---|---|---|
10 |
20 |
18357.50 |
18357.50 |
190 |
20 |
14252.75 |
18357.50 |
200 |
42 |
11508.60 |
18352.80 |
220 |
51 |
17654.50 |
SELECT id
, dept
, salary
, max_sal
FROM staff
LEFT OUTER JOIN
(SELECT dept AS dept#
, MAX(salary) AS max_sal
FROM staff
WHERE dept < 50
GROUP BY dept) AS STAFF_dept
ON dept = dept#
WHERE name LIKE 'S%'
ORDER BY id;
ANSWER
ID | DEPT | SALARY | MAX_SAL |
---|---|---|---|
10 |
20 |
18357.50 |
18357.50 |
190 |
20 |
14252.75 |
18357.50 |
200 |
42 |
11508.60 |
18352.80 |
220 |
51 |
17654.50 |
SELECT id
, dept
, salary
, (SELECT MAX(salary)
FROM staff s2
WHERE s1.dept = s2.dept
AND s2.dept < 50
GROUP BY dept) AS max_sal
FROM staff s1
WHERE name LIKE 'S%'
ORDER BY id;
ANSWER
ID | DEPT | SALARY | MAX_SAL |
---|---|---|---|
10 |
20 |
18357.50 |
18357.50 |
190 |
20 |
14252.75 |
18357.50 |
200 |
42 |
11508.60 |
18352.80 |
220 |
51 |
17654.50 |
16.4. Common Table Expression
A common table expression is a named temporary table that is retained for the duration of a SQL statement. There can be many temporary tables in a single SQL statement. Each must have a unique name and be defined only once. All references to a temporary table (in a given SQL statement run) return the same result. This is unlike tables, views, or aliases, which are derived each time they are called. Also unlike tables, views, or aliases, temporary tables never contain indexes. Certain rules apply to common table expressions:
-
Column names must be specified if the expression is recursive, or if the query invoked returns duplicate column names.
-
The number of column names (if any) that are specified must match the number of columns returned.
-
If there is more than one common-table-expression, latter ones (only) can refer to the output from prior ones. Cyclic references are not allowed.
-
A common table expression with the same name as a real table (or view) will replace the real table for the purposes of the query. The temporary and real tables cannot be referred to in the same query.
-
Temporary table names must follow standard Db2 table naming standards.
-
Each temporary table name must be unique within a query.
-
Temporary tables cannot be used in sub-queries.
16.4.1. Select Examples
In this first query, we don’t have to list the field names (at the top) because every field already has a name (given in the SELECT):
WITH temp1 AS
(SELECT MAX(name) AS max_name
, MAX(dept) AS max_dept
FROM staff)
SELECT *
FROM temp1;
ANSWER
MAX_NAME | MAX_DEPT |
---|---|
Yamaguchi |
84 |
In this next example, the fields being selected are unnamed, so names have to be specified in the WITH statement:
WITH temp1 (max_name, max_dept) AS
(SELECT MAX(name)
, MAX(dept)
FROM staff)
SELECT *
FROM temp1;
ANSWER
MAX_NAME | MAX_DEPT |
---|---|
Yamaguchi |
84 |
A single query can have multiple common-table-expressions. In this next example we use two expressions to get the department with the highest average salary:
WITH temp1 AS
(SELECT dept
, AVG(salary) AS avg_sal
FROM staff
GROUP BY dept)
, temp2 AS
(SELECT MAX(avg_sal) AS max_avg
FROM temp1)
SELECT *
FROM temp2;
ANSWER
MAX_AVG |
---|
20865.8625 |
FYI, the exact same query can be written using nested table expressions thus:
SELECT *
FROM
(SELECT MAX(avg_sal) AS max_avg
FROM (SELECT dept
, AVG(salary) AS avg_sal
FROM staff
GROUP BY dept
) AS temp1
) AS temp2;
ANSWER
MAX_AVG |
---|
20865.8625 |
The next query first builds a temporary table, then derives a second temporary table from the first, and then joins the two temporary tables together. The two tables refer to the same set of rows, and so use the same predicates. But because the second table was derived from the first, these predicates only had to be written once. This greatly simplified the code:
WITH temp1 AS
(SELECT id
, name
, dept
, salary
FROM staff
WHERE id < 300
AND dept <> 55
AND name LIKE 'S%'
AND dept NOT IN
(SELECT deptnumb
FROM org
WHERE division = 'SOUTHERN'
OR location = 'HARTFORD'
)
)
, temp2 AS
(SELECT dept
, MAX(salary) AS max_sal
FROM temp1
GROUP BY dept)
SELECT t1.id
, t1.dept
, t1.salary
, t2.max_sal
FROM temp1 t1
, temp2 t2
WHERE t1.dept = t2.dept
ORDER BY t1.id;
ANSWER
ID | DEPT | SALARY | MAX_SAL |
---|---|---|---|
10 |
20 |
18357.50 |
18357.50 |
190 |
20 |
14252.75 |
18357.50 |
200 |
42 |
11508.60 |
11508.60 |
220 |
51 |
17654.50 |
17654.50 |
16.5. Insert Usage
A common table expression can be used to an insert-select-from statement to build all or part of the set of rows that are inserted:
INSERT INTO staff
WITH temp1 (max1) AS
(SELECT MAX(id) + 1
FROM staff)
SELECT max1, 'A', 1, 'B', 2, 3, 4
FROM temp1;
As it happens, the above query can be written equally well in the raw:
INSERT INTO staff
SELECT MAX(id) + 1, 'A', 1, 'B', 2, 3, 4
FROM staff;
16.6. Full-Select
A fullselect is an alternative way to define a temporary table. Instead of using a WITH clause at the top of the statement, the temporary table definition is embedded in the body of the SQL statement. Certain rules apply:
-
When used in a select statement, a fullselect can either be generated in the FROM part of the query - where it will return a temporary table, or in the SELECT part of the query where it will return a column of data.
-
When the result of a fullselect is a temporary table (i.e. in FROM part of a query), the table must be provided with a correlation name.
-
When the result of a fullselect is a column of data (i.e. in SELECT part of query), each reference to the temporary table must only return a single value.
16.7. Full-Select in FROM Phrase
The following query uses a nested table expression to get the average of an average - in this case the average departmental salary (an average in itself) per division:
SELECT division
, DEC(AVG(dept_avg),7,2) AS div_dept
, COUNT(*) AS #dpts
, SUM(#emps) AS #emps
FROM
(SELECT division
, dept
, AVG(salary) AS dept_avg
, COUNT(*) AS #emps
FROM staff
, org
WHERE dept = deptnumb
GROUP BY division
, dept) AS xxx
GROUP BY division;
ANSWER
DIVISION | DIV_DEPT | #DPTS | #EMPS |
---|---|---|---|
Corporate |
20865.86 |
1 |
4 |
Eastern |
15670.32 |
3 |
13 |
Midwest |
15905.21 |
2 |
9 |
Western |
16875.99 |
2 |
9 |
The next query illustrates how multiple fullselects can be nested inside each other:
SELECT id
FROM (SELECT *
FROM (SELECT id, years, salary
FROM (SELECT *
FROM (SELECT *
FROM staff
WHERE dept < 77) AS t1
WHERE id < 300) AS t2
WHERE job LIKE 'C%') AS t3
WHERE salary < 18000) AS t4
WHERE years < 5;
ANSWER
ID |
---|
170 |
180 |
230 |
A very common usage of a fullselect is to join a derived table to a real table. In the following example, the average salary for each department is joined to the individual staff row:
SELECT a.id
, a.dept
, a.salary
, DEC(b.avgsal,7,2) AS avg_dept
FROM staff a
LEFT OUTER JOIN (SELECT dept AS dept
, AVG(salary) AS avgsal
FROM staff
GROUP BY dept
HAVING AVG(salary) > 16000) AS b
ON a.dept = b.dept
WHERE a.id < 40
ORDER BY a.id;
ANSWER
ID | DEPT | SALARY | AVG_DEPT |
---|---|---|---|
10 |
20 |
18357.50 |
16071.52 |
20 |
20 |
78171.25 |
16071.52 |
30 |
38 |
77506.75 |
16.8. Table Function Usage
If the fullselect query has a reference to a row in a table that is outside of the fullselect, then it needs to be written as a TABLE function call. In the next example, the preceding "A" table is referenced in the fullselect, and so the TABLE function call is required:
SELECT a.id
, a.dept
, a.salary
, b.deptsal
FROM staff a
, TABLE (SELECT b.dept
, SUM(b.salary) AS deptsal
FROM staff b
WHERE b.dept = a.dept
GROUP BY b.dept) AS b
WHERE a.id < 40
ORDER BY a.id;
ANSWER
ID | DEPT | SALARY | DEPTSAL |
---|---|---|---|
10 |
20 |
18357.50 |
64286.10 |
20 |
20 |
78171.25 |
64286.10 |
30 |
38 |
77506.75 |
77285.55 |
Below is the same query written without the reference to the "A" table in the fullselect, and thus without a TABLE function call:
SELECT a.id
, a.dept
, a.salary
, b.deptsal
FROM staff a
, (SELECT b.dept
, SUM(b.salary) AS deptsal
FROM staff b
GROUP BY b.dept) AS b
WHERE a.id < 40
AND b.dept = a.dept
ORDER BY a.id;
ANSWER
ID | DEPT | SALARY | DEPTSAL |
---|---|---|---|
10 |
20 |
18357.50 |
64286.10 |
20 |
20 |
78171.25 |
64286.10 |
30 |
38 |
77506.75 |
77285.55 |
Any externally referenced table in a fullselect must be defined in the query syntax (starting at the first FROM statement) before the fullselect. Thus, in the first example above, if the "A" table had been listed after the "B" table, then the query would have been invalid.
16.9. Full-Select in SELECT Phrase
A fullselect that returns a single column and row can be used in the SELECT part of a query:
SELECT id
, salary
, (SELECT MAX(salary)
FROM staff) AS maxsal
FROM staff a
WHERE id < 60
ORDER BY id;
ANSWER
ID | SALARY | MAXSAL |
---|---|---|
10 |
18357.50 |
22959.20 |
20 |
78171.25 |
22959.20 |
30 |
77506.75 |
22959.20 |
40 |
18006.00 |
22959.20 |
50 |
20659.80 |
22959.20 |
A fullselect in the SELECT part of a statement must return only a single row, but it need not always be the same row. In the following example, the ID and SALARY of each employee is obtained - along with the max SALARY for the employee’s department.
SELECT id
, salary
, (SELECT MAX(salary)
FROM staff b
WHERE a.dept = b.dept) AS maxsal
FROM staff a
WHERE id < 60
ORDER BY id;
ANSWER
ID | SALARY | MAXSAL |
---|---|---|
10 |
18357.50 |
18357.50 |
20 |
78171.25 |
18357.50 |
30 |
77506.75 |
18006.00 |
40 |
18006.00 |
18006.00 |
50 |
20659.80 |
20659.80 |
SELECT id
, dept
, salary
, (SELECT MAX(salary)
FROM staff b
WHERE b.dept = a.dept)
, (SELECT MAX(salary)
FROM staff)
FROM staff a
WHERE id < 60
ORDER BY id;
ANSWER
ID | DEPT | SALARY | 4 | 5 |
---|---|---|---|---|
10 |
20 |
18357.50 |
18357.50 |
22959.20 |
20 |
20 |
78171.25 |
18357.50 |
22959.20 |
30 |
38 |
77506.75 |
18006.00 |
22959.20 |
40 |
38 |
18006.00 |
18006.00 |
22959.20 |
50 |
15 |
20659.80 |
20659.80 |
22959.20 |
16.10. INSERT Usage
The following query uses both an uncorrelated and correlated fullselect in the query that builds the set of rows to be inserted:
INSERT INTO staff
SELECT id + 1
, (SELECT MIN(name)
FROM staff)
, (SELECT dept
FROM staff s2
WHERE s2.id = s1.id - 100)
, 'A', 1, 2, 3
FROM staff s1
WHERE id = (SELECT MAX(id)
FROM staff);
16.11. UPDATE Usage
The following example uses an uncorrelated fullselect to assign a set of workers the average salary in the company - plus two thousand dollars.
UPDATE staff a
SET salary =
(SELECT AVG(salary)+ 2000
FROM staff)
WHERE id < 60;
ANSWER
ID | DEPT | SALARY BEFORE | SALARY AFTER |
---|---|---|---|
10 |
20 |
18357.50 |
18675.64 |
20 |
20 |
78171.25 |
18675.64 |
30 |
38 |
77506.75 |
18675.64 |
40 |
38 |
18006.00 |
18675.64 |
50 |
15 |
20659.80 |
18675.64 |
The next statement uses a correlated fullselect to assign a set of workers the average salary for their department - plus two thousand dollars. Observe that when there is more than one worker in the same department, that they all get the same new salary. This is because the fullselect is resolved before the first update was done, not after each.
UPDATE staff a
SET salary =
(SELECT AVG(salary) + 2000
FROM staff b
WHERE a.dept = b.dept)
WHERE id < 60;
ANSWER
ID | DEPT | SALARY BEFORE | SALARY AFTER |
---|---|---|---|
10 |
20 |
18357.50 |
18071.52 |
20 |
20 |
78171.25 |
18071.52 |
30 |
38 |
77506.75 |
17457.11 |
40 |
38 |
18006.00 |
17457.11 |
50 |
15 |
20659.80 |
17482.33 |
A fullselect is always resolved just once. If it is queried using a correlated expression, then the data returned each time may differ, but the table remains unchanged. |
The next update is the same as the prior, except that two fields are changed:
UPDATE staff a
SET (salary, years) =
(SELECT AVG(salary) + 2000
, MAX(years)
FROM staff b
WHERE a.dept = b.dept)
WHERE id < 60;
16.12. Declared Global Temporary Tables
If we want to temporarily retain some rows for processing by subsequent SQL statements, we can use a Declared Global Temporary Table. A temporary table only exists until the thread is terminated (or sooner). It is not defined in the Db2 catalogue, and neither its definition nor its contents are visible to other users. Multiple users can declare the same temporary table at the same time. Each will be independently working with their own copy.
Usage Notes
For a complete description of this feature, see the SQL reference. Below are some key points:
-
The temporary table name can be any valid Db2 table name. The table qualifier, if provided, must be SESSION. If the qualifier is not provided, it is assumed to be SESSION.
-
If the temporary table has been previously defined in this session, the WITH REPLACE clause can be used to override it. Alternatively, one can DROP the prior instance.
-
An index can be defined on a global temporary table. The qualifier (i.e. SESSION) must be explicitly provided.
-
Any column type can be used in the table, except for: BLOB, CLOB, DBCLOB, LONG VARCHAR, LONG VARGRAPHIC, DATALINK, reference, and structured data types.
-
One can choose to preserve or delete (the default) the rows in the table when a commit occurs. Deleting the rows does not drop the table.
-
Standard identity column definitions can be used if desired.
-
Changes are not logged.
16.12.1. Sample SQL
Below is an example of declaring a global temporary table by listing the columns:
DECLARE GLOBAL TEMPORARY TABLE session.fred
( dept SMALLINT NOT NULL
, avg_salary DEC(7, 2) NOT NULL
, num_emps SMALLINT NOT NULL)
ON COMMIT DELETE ROWS;
In the next example, the temporary table is defined to have exactly the same columns as the existing STAFF table:
DECLARE GLOBAL TEMPORARY TABLE session.fred
LIKE staff INCLUDING COLUMN DEFAULTS
WITH REPLACE
ON COMMIT PRESERVE ROWS;
In the next example, the temporary table is defined to have a set of columns that are returned by a particular select statement. The statement is not actually run at definition time, so any predicates provided are irrelevant:
DECLARE GLOBAL TEMPORARY TABLE session.fred AS
(SELECT dept
, MAX(id) AS max_id
, SUM(salary) AS sum_sal
FROM staff
WHERE name <> 'IDIOT'
GROUP BY dept)
DEFINITION ONLY
WITH REPLACE;
Indexes can be added to temporary tables in order to improve performance and/or to enforce uniqueness:
DECLARE GLOBAL TEMPORARY TABLE session.fred
LIKE staff
INCLUDING COLUMN DEFAULTS
WITH REPLACE ON COMMIT DELETE ROWS;
CREATE UNIQUE INDEX session.fredx ON Session.fred (id);
INSERT INTO session.fred
SELECT *
FROM staff
WHERE id < 200;
SELECT COUNT(*)
FROM session.fred; --> Returns 19
COMMIT;
SELECT COUNT(*)
FROM session.fred; --> Returns 0
A temporary table has to be dropped to reuse the same name:
DECLARE GLOBAL TEMPORARY TABLE session.fred
( dept SMALLINT NOT NULL
, avg_salary DEC(7, 2) NOT NULL
, num_emps SMALLINT NOT NULL)
ON COMMIT DELETE ROWS;
INSERT INTO session.fred
SELECT dept
, AVG(salary)
, COUNT(*)
FROM staff
GROUP BY dept;
SELECT COUNT(*)
FROM session.fred; --> Returns 8
DROP TABLE session.fred;
DECLARE GLOBAL TEMPORARY TABLE session.fred
(dept SMALLINT NOT NULL)
ON COMMIT DELETE ROWS;
SELECT COUNT(*)
FROM session.fred; --> Return 0
16.13. Tablespace
Before a user can create a declared global temporary table, a USER TEMPORARY tablespace that they have access to, has to be created. A typical definition follows:
CREATE USER TEMPORARY TABLESPACE FRED
MANAGED BY DATABASE
USING (FILE 'C:\Db2\TEMPFRED\FRED1' 1000
,FILE 'C:\Db2\TEMPFRED\FRED2' 1000
,FILE 'C:\Db2\TEMPFRED\FRED3' 1000);
GRANT USE OF TABLESPACE FRED TO PUBLIC;
16.14. Do NOT use to Hold Output
In general, do not use a Declared Global Temporary Table to hold job output data, especially if the table is defined ON COMMIT PRESERVE ROWS. If the job fails halfway through, the contents of the temporary table will be lost. If, prior to the failure, the job had updated and then committed Production data, it may be impossible to recreate the lost output because the committed rows cannot be updated twice.
17. Recursive SQL
Recursive SQL enables one to efficiently resolve all manner of complex logical structures that can be really tough to work with using other techniques. On the down side, it is a little tricky to understand at first and it is occasionally expensive. In this chapter we shall first show how recursive SQL works and then illustrate some of the really cute things that one use it for.
Use Recursion To
-
Create sample data.
-
Select the first "n" rows.
-
Generate a simple parser.
-
Resolve a Bill of Materials hierarchy.
-
Normalize and/or denormalize data structures.
When (Not) to Use Recursion
A good SQL statement is one that gets the correct answer, is easy to understand, and is efficient. Let us assume that a particular statement is correct. If the statement uses recursive SQL, it is never going to be categorized as easy to understand (though the reading gets much easier with experience). However, given the question being posed, it is possible that a recursive SQL statement is the simplest way to get the required answer. Recursive SQL statements are neither inherently efficient nor inefficient. Because they often involve a join, it is very important that suitable indexes be provided. Given appropriate indexes, it is quite probable that a recursive SQL statement is the most efficient way to resolve a particular business problem. It all depends upon the nature of the question: If every row processed by the query is required in the answer set (e.g. Find all people who work for Bob), then a recursive statement is likely to very efficient. If only a few of the rows processed by the query are actually needed (e.g. Find all airline flights from Boston to Dallas, then show only the five fastest) then the cost of resolving a large data hierarchy (or network), most of which is immediately discarded, can be very prohibitive. If one wants to get only a small subset of rows in a large data structure, it is very important that of the unwanted data is excluded as soon as possible in the processing sequence. Some of the queries illustrated in this chapter have some rather complicated code in them to do just this. Also, always be on the lookout for infinitely looping data structures.
Conclusion
Recursive SQL statements can be very efficient, if coded correctly, and if there are suitable indexes. When either of the above is not true, they can be very slow.
17.1. How Recursion Works
Below is a description of a very simple application. The table on the left contains a normalized representation of the hierarchical structure on the right. Each row in the table defines a relationship displayed in the hierarchy. The PKEY field identifies a parent key, the CKEY field has related child keys, and the NUM field has the number of times the child occurs within the related parent.
HIERARCHY
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
EEE |
44 |
DDD |
FFF |
5 |
FFF |
GGG |
5 |
AAA | +-----+-----+ | | | BBB CCC DDD | | +-+ +-+--+ | | | EEE FFF | | GGG
17.1.1. List Dependents of AAA
We want to use SQL to get a list of all the dependents of AAA. This list should include not only those items like CCC that are directly related, but also values such as GGG, which are indirectly related. The easiest way to answer this question (in SQL) is to use a recursive SQL statement that goes thus:
WITH parent (pkey, ckey) AS
(SELECT pkey
, ckey
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.pkey
, C.ckey
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey)
SELECT pkey
, ckey
FROM parent;
ANSWER
PKEY | CKEY | PROCESSING SEQUENCE |
---|---|---|
AAA |
BBB |
1st pass |
AAA |
CCC |
1st pass |
AAA |
DDD |
1st pass |
CCC |
EEE |
2nd pass |
DDD |
EEE |
3rd pass |
DDD |
FFF |
3rd pass |
FFF |
GGG |
4th pass |
The above statement is best described by decomposing it into its individual components, and then following of sequence of events that occur:
-
The WITH statement at the top defines a temporary table called PARENT. The upper part of the UNION ALL is only invoked once. It does an initial population of the PARENT table with the three rows that have an immediate parent key of AAA.
-
The lower part of the UNION ALL is run recursively until there are no more matches to the join. In the join, the current child value in the temporary PARENT table is joined to related parent values in the DATA table. Matching rows are placed at the front of the temporary PARENT table. This recursive processing will stop when all of the rows in the PARENT table have been joined to the DATA table.
-
The SELECT phrase at the bottom of the statement sends the contents of the PARENT table back to the user’s program.
-
Another way to look at the above process is to think of the temporary PARENT table as a stack of data. This stack is initially populated by the query in the top part of the UNION ALL.
-
Next, a cursor starts from the bottom of the stack and goes up. Each row obtained by the cursor is joined to the DATA table. Any matching rows obtained from the join are added to the top of the stack (i.e. in front of the cursor). When the cursor reaches the top of the stack, the statement is done. The following diagram illustrates this process:
+-----+-----+-----+ +-----+ +-----+-----+ +-----+ PKEY > | AAA | AAA | AAA | | CCC | | DDD | DDD | | FFF | +-----+-----+-----+ +-----+ +-----+-----+ +-----+ CKEY > | BBB | CCC | DDD | | EEE | | EEE | FFF | | GGG | +-----+-----+-----+ +-----+ +-----+-----+ +-----+ | | ^ ^ ^ | ^ +-----------------+ | | +--------+ | | | +-----------------------+----+
17.1.2. Notes & Restrictions
-
Recursive SQL requires that there be a UNION ALL phrase between the two main parts of the statement. The UNION ALL, unlike the UNION, allows for duplicate output rows, which is what often comes out of recursive processing.
-
If done right, recursive SQL is often fairly efficient. When it involves a join similar to the example shown above, it is important to make sure that this join is efficient. To this end, suitable indexes should be provided.
-
The output of a recursive SQL is a temporary table (usually). Therefore, all temporary table usage restrictions also apply to recursive SQL output. See the section titled "Common Table Expression" for details.
-
The output of one recursive expression can be used as input to another recursive expression in the same SQL statement. This can be very handy if one has multiple logical hierarchies to traverse (e.g. First find all of the states in the USA, then final all of the cities in each state).
-
Any recursive coding, in any language, can get into an infinite loop - either because of bad coding, or because the data being processed has a recursive value structure. To prevent your SQL running forever, see the section titled Halting Recursive Processing.
17.1.3. Sample Table DDL & DML
CREATE TABLE hierarchy
( pkey CHAR(03) NOT NULL
, ckey CHAR(03) NOT NULL
, num SMALLINT NOT NULL
, PRIMARY KEY(pkey, ckey)
, CONSTRAINT dt1 CHECK (pkey <> ckey)
, CONSTRAINT dt2 CHECK (num > 0));
COMMIT;
CREATE UNIQUE INDEX hier_x1
ON hierarchy (ckey, pkey);
COMMIT;
INSERT INTO hierarchy VALUES
('AAA','BBB', 1)
, ('AAA','CCC', 5)
, ('AAA','DDD',20)
, ('CCC','EEE',33)
, ('DDD','EEE',44)
, ('DDD','FFF', 5)
, ('FFF','GGG', 5);
COMMIT;
17.2. Introductory Recursion
This section will use recursive SQL statements to answer a series of simple business questions using the sample HIERARCHY table described on Sample Table DDL - Recursion. Be warned that things are going to get decidedly more complex as we proceed.
17.2.1. List all Children #1
Find all the children of AAA. Don’t worry about getting rid of duplicates, sorting the data, or any other of the finer details.
WITH parent (ckey) AS
(SELECT ckey
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.ckey
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey)
SELECT ckey
FROM parent;
ANSWER
CKEY |
---|
BBB |
CCC |
DDD |
EEE |
EEE |
FFF |
GGG |
HIERARCHY
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
EEE |
44 |
DDD |
FFF |
5 |
FFF |
GGG |
5 |
Much of the SQL shown in this section will loop forever if the target database has a recursive data structure. See Halting Recursive Processing for details on how to prevent this. |
The above SQL statement uses standard recursive processing. The first part of the UNION ALL seeds the temporary table PARENT. The second part recursively joins the temporary table to the source data table until there are no more matches. The final part of the query displays the result set. Imagine that the HIERARCHY table used above is very large and that we also want the above query to be as efficient as possible. In this case, two indexes are required; The first, on PKEY, enables the initial select to run efficiently. The second, on CKEY, makes the join in the recursive part of the query efficient. The second index is arguably more important than the first because the first is only used once, whereas the second index is used for each child of the toplevel parent.
17.2.2. List all Children #2
Find all the children of AAA, include in this list the value AAA itself. To satisfy the latter requirement we will change the first SELECT statement (in the recursive code) to select the parent itself instead of the list of immediate children. A DISTINCT is provided in order to ensure that only one line containing the name of the parent (i.e. "AAA") is placed into the temporary PARENT table. NOTE: _Before the introduction of recursive SQL processing, it often made sense to define the top-most level in a hierarchical data structure as being a parent-child of itself. For example, the HIERARCHY table might contain a row indicating that "AAA" is a child of "AAA". If the target table has data like this, add another predicate: C.PKEY <> C.CKEY
to the recursive part of the SQL statement to stop the query from looping forever.
WITH parent (ckey) AS
(SELECT DISTINCT pkey
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.ckey
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey)
SELECT ckey
FROM parent;
ANSWER
CKEY |
---|
AAA |
BBB |
CCC |
DDD |
EEE |
EEE |
FFF |
GGG |
HIERARCHY
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
EEE |
44 |
DDD |
FFF |
5 |
FFF |
GGG |
5 |
In most, but by no means all, business situations, the above SQL statement is more likely to be what the user really wanted than the SQL before. Ask before you code.
17.2.3. List Distinct Children
Get a distinct list of all the children of AAA. This query differs from the prior only in the use of the DISTINCT phrase in the final select.
WITH parent (ckey) AS
(SELECT DISTINCT pkey
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.ckey
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey)
SELECT DISTINCT ckey
FROM parent;
ANSWER
CKEY |
---|
AAA |
BBB |
CCC |
DDD |
EEE |
FFF |
GGG |
HIERARCHY
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
EEE |
44 |
DDD |
FFF |
5 |
FFF |
GGG |
5 |
The next thing that we want to do is build a distinct list of children of AAA that we can then use to join to other tables. To do this, we simply define two temporary tables. The first does the recursion and is called PARENT. The second, called DISTINCT_PARENT, takes the output from the first and removes duplicates.
WITH parent (ckey) AS
(SELECT DISTINCT pkey
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.ckey
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey)
, distinct_parent (ckey) AS
(SELECT DISTINCT ckey
FROM parent)
SELECT ckey
FROM distinct_parent;
ANSWER
CKEY |
---|
AAA |
BBB |
CCC |
DDD |
EEE |
FFF |
GGG |
HIERARCHY
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
EEE |
44 |
DDD |
FFF |
5 |
FFF |
GGG |
5 |
17.2.4. Show Item Level
Get a list of all the children of AAA. For each value returned, show its level in the logical hierarchy relative to AAA.
WITH parent (ckey, lvl) AS
(SELECT DISTINCT pkey
, 0
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.ckey
, P.lvl +1
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey)
SELECT ckey
, lvl
FROM parent;
ANSWER
CKEY | LVL |
---|---|
AAA |
0 |
BBB |
1 |
CCC |
1 |
DDD |
1 |
EEE |
2 |
EEE |
2 |
FFF |
2 |
GGG |
3 |
The above statement has a derived integer field called LVL. In the initial population of the temporary table this level value is set to zero. When subsequent levels are reached, this value in incremented by one.
17.2.5. Select Certain Levels
Get a list of all the children of AAA that are less than three levels below AAA.
WITH parent (ckey, lvl) AS
(SELECT DISTINCT pkey
, 0
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.ckey
, P.lvl +1
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey)
SELECT ckey, lvl
FROM parent
WHERE lvl < 3;
ANSWER
CKEY | LVL |
---|---|
AAA |
0 |
BBB |
1 |
CCC |
1 |
DDD |
1 |
EEE |
2 |
EEE |
2 |
FFF |
2 |
HIERARCHY
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
EEE |
44 |
DDD |
FFF |
5 |
FFF |
GGG |
5 |
The above statement has two main deficiencies:
-
It will run forever if the database contains an infinite loop.
-
It may be inefficient because it resolves the whole hierarchy before discarding those levels that are not required.
To get around both of these problems, we can move the level check up into the body of the recursive statement. This will stop the recursion from continuing as soon as we reach the target level. We will have to add "+ 1" to the check to make it logically equivalent:
WITH parent (ckey, lvl) AS
(SELECT DISTINCT pkey
, 0
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.ckey
, P.lvl +1
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey
AND P.lvl+1 < 3)
SELECT ckey, lvl
FROM parent;
ANSWER
CKEY | LVL |
---|---|
AAA |
0 |
BBB |
1 |
CCC |
1 |
DDD |
1 |
EEE |
2 |
EEE |
2 |
FFF |
2 |
The only difference between this statement and the one before is that the level check is now done in the recursive part of the statement. This new level-check predicate has a dual function: It gives us the answer that we want, and it stops the SQL from running forever if the database happens to contain an infinite loop (e.g. DDD was also a parent of AAA). One problem with this general statement design is that it can not be used to list only that data which pertains to a certain lower level (e.g. display only level 3 data). To answer this kind of question efficiently we can combine the above two queries, having appropriate predicates in both places (see next).
17.2.6. Select Explicit Level
Get a list of all the children of AAA that are exactly two levels below AAA.
WITH parent (ckey, lvl) AS
(SELECT DISTINCT pkey
, 0
FROM hierarchy
WHERE pkey = 'AAA'
UNION ALL
SELECT C.ckey
, P.lvl +1
FROM hierarchy C
, parent P
WHERE P.ckey = C.pkey
AND P.lvl+1 < 3)
SELECT ckey
, lvl
FROM parent
WHERE lvl = 2;
ANSWER
CKEY | LVL |
---|---|
EEE |
2 |
EEE |
2 |
FFF |
2 |
HIERARCHY
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
EEE |
44 |
DDD |
FFF |
5 |
FFF |
GGG |
5 |
In the recursive part of the above statement all of the levels up to and including that which is required are obtained. All undesired lower levels are then removed in the final select.
17.2.7. Trace a Path - Use Multiple Recursions
Multiple recursive joins can be included in a single query. The joins can run independently, or the output from one recursive join can be used as input to a subsequent. Such code enables one to do the following:
-
Expand multiple hierarchies in a single query. For example, one might first get a list of all departments (direct and indirect) in a particular organization, and then use the department list as a seed to find all employees (direct and indirect) in each department.
-
Go down, and then up, a given hierarchy in a single query. For example, one might want to find all of the children of AAA, and then all of the parents. The combined result is the list of objects that AAA is related to via a direct parent-child path.
-
Go down the same hierarchy twice, and then combine the results to find the matches, or the non-matches. This type of query might be used to, for example, see if two companies own shares in the same subsidiary.
-
The next example recursively searches the HIERARCHY table for all values that are either a child or a parent (direct or indirect) of the object DDD. The first part of the query gets the list of children, the second part gets the list of parents (but never the value DDD itself), and then the results are combined.
WITH children (kkey, lvl)
AS (SELECT ckey
, 1
FROM hierarchy
WHERE pkey = 'DDD'
UNION ALL
SELECT H.ckey
, C.lvl + 1
FROM hierarchy H
, children C
WHERE H.pkey = C.kkey)
, parents (kkey, lvl)
AS (SELECT pkey ,
-1
FROM hierarchy
WHERE ckey = 'DDD'
UNION ALL
SELECT H.pkey
, P.lvl - 1
FROM hierarchy H
, parents P
WHERE H.ckey = P.kkey)
SELECT kkey
, lvl
FROM children
UNION ALL
SELECT kkey
, lvl
FROM parents;
ANSWER
KKEY | LVL |
---|---|
AAA |
-1 |
EEE |
1 |
FFF |
1 |
GGG |
2 |
17.2.8. Extraneous Warning Message
Some recursive SQL statements generate the following warning when the Db2 parser has reason to suspect that the statement may run forever:
SQL0347W The recursive common table expression "GRAEME.TEMP1" may contain an infinite loop. SQLSTATE=01605
The text that accompanies this message provides detailed instructions on how to code recursive SQL so as to avoid getting into an infinite loop. The trouble is that even if you do exactly as told you may still get the silly message. To illustrate, the following two SQL statements are almost identical. Yet the first gets a warning and the second does not:
WITH temp1 (n1) AS
(SELECT id
FROM staff
WHERE id = 10
UNION ALL
SELECT n1 +10
FROM temp1
WHERE n1 < 50)
SELECT *
FROM temp1;
ANSWER
N1 |
---|
warn |
10 |
20 |
30 |
40 |
50 |
WITH temp1 (n1) AS
(SELECT INT(id)
FROM staff
WHERE id = 10
UNION ALL
SELECT n1 +10
FROM temp1
WHERE n1 < 50)
SELECT *
FROM temp1;
ANSWER
N1 |
---|
10 |
20 |
30 |
40 |
50 |
If you know what you are doing, ignore the message.
17.2.9. Logical Hierarchy Flavours
Before getting into some of the really nasty stuff, we best give a brief overview of the various kinds of logical hierarchy that exist in the real world and how each is best represented in a relational database. Some typical data hierarchy flavours are shown below. Note that the three on the left form one, mutually exclusive, set and the two on the right another. Therefore, it is possible for a particular hierarchy to be both divergent and unbalanced (or balanced), but not both divergent and convergent.
DIVERGENT CONVERGENT RECURSIVE BALANCED UNBALANCED ========= ========== ========= ======== ========== AAA AAA AAA<--+ AAA AAA | | | | | | +-+-+ +-+-+ +-+-+ | +-+-+ +-+-+ | | | | | | | | | | | BBB CCC BBB CCC BBB CCC>+ BBB CCC BBB CCC | | | | | | | +-+-+ +-+-+-+ +-+-+ | +---+ +-+-+ | | | | | | | | | | | DDD EEE DDD EEE DDD EEE DDD EEE FFF DDD EEE
17.2.10. Divergent Hierarchy
In this flavour of hierarchy, no object has more than one parent. Each object can have none, one, or more than one, dependent child objects. Physical objects (e.g. Geographic entities) tend to be represented in this type of hierarchy. This type of hierarchy will often incorporate the concept of different layers in the hierarchy referring to differing kinds of object - each with its own set of attributes. For example, a Geographic hierarchy might consist of countries, states, cities, and street addresses. A single table can be used to represent this kind of hierarchy in a fully normalized form. One field in the table will be the unique key, another will point to the related parent. Other fields in the table may pertain either to the object in question, or to the relationship between the object and its parent. For example, in the following table the PRICE field has the price of the object, and the NUM field has the number of times that the object occurs in the parent.
OBJECTS_RELATES
KEYO | PKEY | NUM | PRICE |
---|---|---|---|
AAA |
$10 |
||
BBB |
AAA |
1 |
$21 |
CCC |
AAA |
5 |
$23 |
DDD |
AAA |
20 |
$25 |
EEE |
DDD |
44 |
$33 |
FFF |
DDD |
5 |
$34 |
GGG |
FFF |
5 |
$44 |
Some database designers like to make the arbitrary judgment that every object has a parent, and in those cases where there is no "real" parent, the object considered to be a parent of itself. In the above table, this would mean that AAA would be defined as a parent of AAA. Please appreciate that this judgment call does not affect the objects that the database represents, but it can have a dramatic impact on SQL usage and performance. Prior to the introduction of recursive SQL, defining top level objects as being self-parenting was sometimes a good idea because it enabled one to resolve a hierarchy using a simple join without unions. This same process is now best done with recursive SQL. Furthermore, if objects in the database are defined as self-parenting, the recursive SQL will get into an infinite loop unless extra predicates are provided.
17.2.11. Convergent Hierarchy
NUMBER OF TABLES: A convergent hierarchy has many-to-many relationships that require two tables for normalized data storage. The other hierarchy types require but a single table.
In this flavour of hierarchy, each object can have none, one, or more than one, parent and/or dependent child objects.
Convergent hierarchies are often much more difficult to work with than similar divergent hierarchies. Logical entities, or man-made objects, (e.g. Company Divisions) often have this type of hierarchy. Two tables are required in order to represent this kind of hierarchy in a fully normalized form. One table describes the object, and the other describes the relationships between the objects.
OBJECTS
KEYO | PRICE |
---|---|
AAA |
$10 |
BBB |
$21 |
CCC |
$23 |
DDD |
$25 |
EEE |
$33 |
FFF |
$34 |
GGG |
$44 |
RELATIONSHIPS
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
EEE |
44 |
DDD |
FFF |
5 |
FFF |
GGG |
5 |
One has to be very careful when resolving a convergent hierarchy to get the answer that the user actually wanted. To illustrate, if we wanted to know how many children AAA has in the above structure the "correct" answer could be six, seven, or eight. To be precise, we would need to know if EEE should be counted twice and if AAA is considered to be a child of itself.
17.2.12. Recursive Hierarchy
Recursive data hierarchies will cause poorly written recursive SQL statements to run forever. See the section titled Halting Recursive Processing for details on how to prevent this, and how to check that a hierarchy is not recursive. |
In this flavour of hierarchy, each object can have none, one, or more than one parent. Also, each object can be a parent and/or a child of itself via another object, or via itself directly. In the business world, this type of hierarchy is almost always wrong. When it does exist, it is often because a standard convergent hierarchy has gone a bit haywire. This database design is exactly the same as the one for a convergent hierarchy. Two tables are (usually) required in order to represent the hierarchy in a fully normalized form. One table describes the object, and the other describes the relationships between the objects.
OBJECTS
KEYO | PRICE |
---|---|
AAA |
$10 |
BBB |
$21 |
CCC |
$23 |
DDD |
$25 |
EEE |
$33 |
FFF |
$34 |
GGG |
$44 |
RELATIONSHIPS
PKEY | CKEY | NUM |
---|---|---|
AAA |
BBB |
1 |
AAA |
CCC |
5 |
AAA |
DDD |
20 |
CCC |
EEE |
33 |
DDD |
AAA |
99 |
DDD |
FFF |
5 |
DDD |
EEE |
44 |
FFF |
GGG |
5 |
Prior to the introduction of recursive SQL, it took some non-trivial coding root out recursive data structures in convergent hierarchies. Now it is a no-brainer, see Halting Recursive Processing for details.
17.2.13. Balanced & Unbalanced Hierarchies
In some logical hierarchies the distance, in terms of the number of intervening levels, from the top parent entity to its lowest-level child entities is the same for all legs of the hierarchy. Such a hierarchy is considered to be balanced. An unbalanced hierarchy is one where the distance from a top-level parent to a lowest-level child is potentially different for each leg of the hierarchy.
AAA << Balanced hierarchy AAA | Unbalanced hierarchy >> | +-----+-----+ +---+----+ | | | | | | BBB CCC DDD | CCC DDD | | | | | | | | +-+-+ | +-+ +-+-+ | | | | | | | | EEE FFF GGG HHH FFF GGG HHH | | III
Balanced hierarchies often incorporate the concept of levels, where a level is a subset of the values in the hierarchy that are all of the same time and are also the same distance from the top level parent. For example, in the balanced hierarchy above each of the three levels shown might refer to a different category of object (e.g. country, state, city). By contrast, in the unbalanced hierarchy above is probable that the objects being represented are all of the same general category (e.g. companies that own other companies). Divergent hierarchies are the most likely to be balanced. Furthermore, balanced and/or divergent hierarchies are the kind that are most often used to do data summation at various intermediate levels. For example, a hierarchy of countries, states, and cities, is likely to be summarized at any level.
17.2.14. Data & Pointer Hierarchies
The difference between a data and a pointer hierarchy is not one of design, but of usage. In a pointer schema, the main application tables do not store a description of the logical hierarchy. Instead, they only store the base data. Separate to the main tables are one, or more, related tables that define which hierarchies each base data row belongs to.
Typically, in a pointer hierarchy, the main data tables are much larger and more active than the hierarchical tables. A banking application is a classic example of this usage pattern. There is often one table that contains core customer information and several related tables that enable one to do analysis by customer category. A data hierarchy is an altogether different beast. An example would be a set of tables that contain information on all that parts that make up an aircraft. In this kind of application the most important information in the database is often that which pertains to the relationships between objects. These tend to be very complicated often incorporating the attributes: quantity, direction, and version. Recursive processing of a data hierarchy will often require that one does a lot more than just find all dependent keys. For example, to find the gross weight of an aircraft from such a database one will have to work with both the quantity and weight of all dependent objects. Those objects that span sub-assembles (e.g. a bolt connecting to engine to the wing) must not be counted twice, missed out, nor assigned to the wrong sub-grouping. As always, such questions are essentially easy to answer, the trick is to get the right answer.
17.2.15. Halting Recursive Processing
One occasionally encounters recursive hierarchical data structures (i.e. where the parent item points to the child, which then points back to the parent). This section describes how to write recursive SQL statements that can process such structures without running forever. There are three general techniques that one can use: * Stop processing after reaching a certain number of levels. * Keep a record of where you have been, and if you ever come back, either fail or in some other way stop recursive processing. * Keep a record of where you have been, and if you ever come back, simply ignore that row and keep on resolving the rest of hierarchy.
Sample Table DDL & DML
The following table is a normalized representation of the recursive hierarchy on the right. Note that AAA and DDD are both a parent and a child of each other.
TROUBLE
PKEY | CKEY |
---|---|
AAA |
BBB |
AAA |
CCC |
AAA |
DDD |
CCC |
EEE |
DDD |
AAA |
DDD |
FFF |
DDD |
EEE |
FFF |
GGG |
AAA <------+ | | +-----+-----+ | | | | | BBB CCC DDD>-+ | | +-+ +-+--+ | | | EEE FFF | | GGG
Below is the DDL and DML that was used to create the above table.
CREATE TABLE trouble
( pkey CHAR(03) NOT NULL
, ckey CHAR(03) NOT NULL);
CREATE UNIQUE INDEX tble_x1 ON trouble (pkey, ckey);
CREATE UNIQUE INDEX tble_x2 ON trouble (ckey, pkey);
INSERT INTO trouble VALUES
('AAA','BBB')
, ('AAA','CCC')
, ('AAA','DDD')
, ('CCC','EEE')
, ('DDD','AAA')
, ('DDD','EEE')
, ('DDD','FFF')
, ('FFF','GGG');
17.2.16. Other Loop Types
In the above table, the beginning object (i.e. AAA) is part of the data loop. This type of loop can be detected using simpler SQL than what is given here. But a loop that does not include the beginning object (e.g. AAA points to BBB, which points to CCC, which points back to BBB) requires the somewhat complicated SQL that is used in this section.
Stop After "n" Levels
Find all the children of AAA. In order to avoid running forever, stop after four levels.
WITH parent (pkey, ckey, lvl) AS
(SELECT DISTINCT pkey
, pkey
, 0
FROM trouble
WHERE pkey = 'AAA'
UNION ALL
SELECT C.pkey
, C.ckey
, P.lvl + 1
FROM trouble C
, parent P
WHERE P.ckey = C.pkey
AND P.lvl + 1 < 4)
SELECT *
FROM parent;
ANSWER
PKEY | CKEY | LVL |
---|---|---|
AAA |
AAA |
0 |
AAA |
BBB |
1 |
AAA |
CCC |
1 |
AAA |
DDD |
1 |
CCC |
EEE |
2 |
DDD |
AAA |
2 |
DDD |
EEE |
2 |
DDD |
FFF |
2 |
AAA |
BBB |
3 |
AAA |
CCC |
3 |
AAA |
DDD |
3 |
FFF |
GGG |
3 |
TROUBLE
PKEY | CKEY |
---|---|
AAA |
BBB |
AAA |
CCC |
AAA |
DDD |
CCC |
EEE |
DDD |
AAA |
DDD |
FFF |
DDD |
EEE |
FFF |
GGG |
In order for the above statement to get the right answer, we need to know before beginning the maximum number of valid dependent levels (i.e. non-looping) there are in the hierarchy. This information is then incorporated into the recursive predicate (see: P.LVI + 1 < 4). If the number of levels is not known, and we guess wrong, we may not find all the children of AAA. For example, if we had stopped at "2" in the above query, we would not have found the child GGG. A more specific disadvantage of the above statement is that the list of children contains duplicates. These duplicates include those specific values that compose the infinite loop (i.e. AAA and DDD), and also any children of either of the above.
Stop When Loop Found
A far better way to stop recursive processing is to halt when, and only when, we determine that we have been to the target row previously. To do this, we need to maintain a record of where we have been, and then check this record against the current key value in each row joined to. Db2 does not come with an in-built function that can do this checking, so we shall define our own.
#Define Function
Below is the definition code for a user-defined Db2 function that is very similar to the standard LOCATE function. It searches for one string in another, block by block. For example, if one was looking for the string "ABC", this function would search the first three bytes, then the next three bytes, and so on. If a match is found, the function returns the relevant block number, else zero.
CREATE FUNCTION LOCATE_BLOCK(searchstr VARCHAR(30000)
, lookinstr VARCHAR(30000))
RETURNS INTEGER
BEGIN ATOMIC
DECLARE lookinlen, searchlen INT;
DECLARE locatevar, returnvar INT DEFAULT 0;
DECLARE beginlook
INT DEFAULT 1;
SET lookinlen = LENGTH(lookinstr);
SET searchlen = LENGTH(searchstr);
WHILE locatevar = 0 AND beginlook <= lookinlen
DO
SET locatevar = LOCATE(searchstr,SUBSTR(lookinstr
, beginlook
, searchlen));
SET beginlook = beginlook + searchlen;
SET returnvar = returnvar + 1;
END WHILE;
IF locatevar = 0 THEN
SET returnvar = 0;
END IF;
RETURN returnvar;
END
Below is an example of the function in use. Observe that the function did not find the string "th" in the name "Smith" because the two characters did not start in an position that was some multiple of the length of the test string:
SELECT id
, name
, LOCATE('th', name) AS l1
, LOCATE_BLOCK('th', name) AS l2
FROM staff
WHERE LOCATE('th', name) > 1;
ANSWER
ID | NAME | L1 | L2 |
---|---|---|---|
70 |
Rothman |
3 |
2 |
220 |
Smith |
4 |
0 |
The LOCATE_BLOCK function shown above is the minimalist version, without any error checking. If it were used in a Production environment, it would have checks for nulls, and for various invalid input values. |
#Use Function
Now all we need to do is build a string, as we do the recursion, that holds every key value that has previously been accessed. This can be done using simple concatenation:
WITH parent (pkey, ckey, lvl, path, loop) AS
(SELECT DISTINCT pkey
, pkey
, 0
, VARCHAR(pkey, 20)
, 0
FROM trouble
WHERE pkey = 'AAA'
UNION ALL
SELECT C.pkey
, C.ckey
, P.lvl + 1
, P.path || C.ckey
, LOCATE_BLOCK(C.ckey,P.path)
FROM trouble C
, parent P
WHERE P.ckey = C.pkey
AND P.lvl + 1 < 4)
SELECT *
FROM parent;
ANSWER
PKEY | CKEY | LVL | PATH | LOOP |
---|---|---|---|---|
AAA |
AAA |
0 |
AAA |
0 |
AAA |
BBB |
1 |
AAABBB |
0 |
AAA |
CCC |
1 |
AAACCC |
0 |
AAA |
DDD |
1 |
AAADDD |
0 |
CCC |
EEE |
2 |
AAACCCEEE |
0 |
DDD |
AAA |
2 |
AAADDDAAA |
1 |
DDD |
EEE |
2 |
AAADDDEEE |
0 |
DDD |
FFF |
2 |
AAADDDFFF |
0 |
AAA |
BBB |
3 |
AAADDDAAABBB |
0 |
AAA |
CCC |
3 |
AAADDDAAACCC |
0 |
AAA |
DDD |
3 |
AAADDDAAADDD |
2 |
FFF |
GGG |
3 |
AAADDDFFFGGG |
0 |
Now we can get rid of the level check, and instead use the LOCATE_BLOCK function to avoid loops in the data:
WITH parent (pkey, ckey, lvl, path) AS
(SELECT DISTINCT pkey
, pkey
, 0
, VARCHAR(pkey,20)
FROM trouble
WHERE pkey = 'AAA'
UNION ALL
SELECT C.pkey
, C.ckey
, P.lvl + 1
, P.path || C.ckey
FROM trouble C
, parent P
WHERE P.ckey = C.pkey
AND LOCATE_BLOCK(C.ckey,P.path) = 0)
SELECT *
FROM parent;
ANSWER
PKEY | CKEY | LVL | PATH |
---|---|---|---|
AAA |
AAA |
0 |
AAA |
AAA |
BBB |
1 |
AAABBB |
AAA |
CCC |
1 |
AAACCC |
AAA |
DDD |
1 |
AAADDD |
CCC |
EEE |
2 |
AAACCCEEE |
DDD |
EEE |
2 |
AAADDDEEE |
DDD |
FFF |
2 |
AAADDDFFF |
FFF |
GGG |
3 |
AAADDDFFFGGG |
The next query is the same as the previous, except that instead of excluding all loops from the answer-set, it marks them as such, and gets the first item, but goes no further.
WITH parent (pkey, ckey, lvl, path, loop) AS
(SELECT DISTINCT pkey
, pkey
, 0
, VARCHAR(pkey,20)
, 0
FROM trouble
WHERE pkey = 'AAA'
UNION ALL
SELECT C.pkey
, C.ckey
, P.lvl + 1
, P.path || C.ckey
, LOCATE_BLOCK(C.ckey,P.path) DDD AAA
FROM trouble C
, parent P
WHERE P.ckey = C.pkey
AND P.loop = 0)
SELECT *
FROM parent;
ANSWER
PKEY | CKEY | LVL | PATH | LOOP |
---|---|---|---|---|
AAA |
AAA |
0 |
AAA |
0 |
AAA |
BBB |
1 |
AAABBB |
0 |
AAA |
CCC |
1 |
AAACCC |
0 |
AAA |
DDD |
1 |
AAADDD |
0 |
CCC |
EEE |
2 |
AAACCCEEE |
0 |
DDD |
AAA |
2 |
AAADDDAAA |
1 |
DDD |
EEE |
2 |
AAADDDEEE |
0 |
DDD |
FFF |
2 |
AAADDDFFF |
0 |
FFF |
GGG |
3 |
AAADDDFFFGGG |
0 |
The next query tosses in another predicate (in the final select) to only list those rows that point back to a previously processed parent:
WITH parent (pkey, ckey, lvl, path, loop) AS
(SELECT DISTINCT pkey
, pkey
, 0
, VARCHAR(pkey,20)
, 0
FROM trouble
WHERE pkey = 'AAA'
UNION ALL
SELECT C.pkey
, C.ckey
, P.lvl + 1
, P.path || C.ckey
, LOCATE_BLOCK(C.ckey,P.path)
FROM trouble C
, parent P
WHERE P.ckey = C.pkey
AND P.loop = 0)
SELECT pkey
, ckey
FROM parent
WHERE loop > 0;
ANSWER
PKEY | CKEY |
---|---|
DDD |
AAA |
To delete the offending rows from the table, all one has to do is insert the above values into a temporary table, then delete those rows in the TROUBLE table that match. However, before one does this, one has decide which rows are the ones that should not be there. In the above query, we started processing at AAA, and then said that any row that points back to AAA, or to some child or AAA, is causing a loop. We thus identified the row from DDD to AAA as being a problem. But if we had started at the value DDD, we would have said instead that the row from AAA to DDD was the problem. The point to remember her is that the row you decide to delete is a consequence of the row that you decided to define as your starting point.
DECLARE GLOBAL TEMPORARY TABLE SESSION.del_list
( pkey CHAR(03) NOT NULL
, ckey CHAR(03) NOT NULL)
ON COMMIT PRESERVE ROWS;
INSERT INTO SESSION.del_list
WITH parent (pkey, ckey, lvl, path, loop) AS
(SELECT DISTINCT pkey
, pkey
, 0
, VARCHAR(pkey,20)
, 0
FROM trouble
WHERE pkey = 'AAA'
UNION ALL
SELECT C.pkey
, C.ckey
, P.lvl + 1
, P.path || C.ckey
, LOCATE_BLOCK(C.ckey,P.path)
FROM trouble C
, parent P
WHERE P.ckey = C.pkey
AND P.loop = 0)
SELECT pkey
, ckey
FROM parent
WHERE loop > 0;
DELETE FROM trouble
WHERE (pkey,ckey) IN
(SELECT pkey
, ckey
FROM SESSION.del_list);
Working with Other Key Types
The LOCATE_BLOCK solution shown above works fine, as long as the key in question is a fixed length character field. If it isn’t, it can be converted to one, depending on what it is:
-
Cast VARCHAR columns as type CHAR.
-
Convert other field types to character using the HEX function.
Keeping the Hierarchy Clean
Rather that go searching for loops, one can toss in a couple of triggers that will prevent the table from every getting data loops in the first place. There will be one trigger for inserts, and another for updates. Both will have the same general logic:
-
For each row inserted/updated, retain the new PKEY value.
-
Recursively scan the existing rows, starting with the new CKEY value.
-
Compare each existing CKEY value retrieved to the new PKEY value. If it matches, the changed row will cause a loop, so flag an error.
-
If no match is found, allow the change.
Here is the insert trigger:
CREATE TRIGGER TBL_INS
NO CASCADE BEFORE INSERT ON trouble
REFERENCING NEW AS NNN
FOR EACH ROW MODE Db2SQL
WITH temp (pkey, ckey) AS
(VALUES (NNN.pkey
, NNN.ckey)
UNION ALL
SELECT TTT.pkey
, CASE
WHEN TTT.ckey = TBL.pkey THEN RAISE_ERROR('70001','LOOP FOUND')
ELSE TBL.ckey
END
FROM trouble TBL
, temp TTT
WHERE TTT.ckey = TBL.pkey)
SELECT *
FROM temp;
TROUBLE
PKEY | CKEY |
---|---|
AAA |
BBB |
AAA |
CCC |
AAA |
DDD |
CCC |
EEE |
DDD |
AAA |
DDD |
FFF |
DDD |
EEE |
FFF |
GGG |
Here is the update trigger:
CREATE TRIGGER TBL_UPD
NO CASCADE BEFORE UPDATE OF pkey, ckey ON trouble
REFERENCING NEW AS NNN
FOR EACH ROW MODE Db2SQL
WITH temp (pkey, ckey) AS
(VALUES (NNN.pkey
, NNN.ckey)
UNION ALL
SELECT TTT.pkey
, CASE
WHEN TTT.ckey = TBL.pkey THEN RAISE_ERROR('70001','LOOP FOUND')
ELSE TBL.ckey
END
FROM trouble TBL
, temp TTT
WHERE TTT.ckey = TBL.pkey)
SELECT *
FROM temp;
Given the above preexisting TROUBLE data (absent the DDD to AAA row), the following statements would be rejected by the above triggers:
INSERT INTO trouble VALUES('GGG','AAA');
UPDATE trouble SET ckey = 'AAA' WHERE pkey = 'FFF';
UPDATE trouble SET pkey = 'GGG' WHERE ckey = 'DDD';
Observe that neither of the above triggers use the LOCATE_BLOCK function to find a loop. This is because these triggers are written assuming that the table is currently loop free. If this is not the case, they may run forever. The LOCATE_BLOCK function enables one to check every row processed, to see if one has been to that row before. In the above triggers, only the start position is checked for loops. So if there was a loop that did not encompass the start position, the LOCATE_BLOCK check would find it, but the code used in the triggers would not.
17.3. Clean Hierarchies and Efficient Joins
17.3.1. Introduction
One of the more difficult problems in any relational database system involves joining across multiple hierarchical data structures. The task is doubly difficult when one or more of the hierarchies involved is a data structure that has to be resolved using recursive processing. In this section, we will describe how one can use a mixture of tables and triggers to answer this kind of query very efficiently. A typical question might go as follows: Find all matching rows where the customer is in some geographic region, and the item sold is in some product category, and person who made the sale is in some company sub-structure. If each of these qualifications involves expanding a hierarchy of object relationships of indeterminate and/or nontrivial depth, then a simple join or standard data denormalization will not work. In Db2, one can answer this kind of question by using recursion to expand each of the data hierarchies. Then the query would join (sans indexes) the various temporary tables created by the recursive code to whatever other data tables needed to be accessed. Unfortunately, the performance will probably be lousy. Alternatively, one can often efficiently answer this general question using a set of suitably indexed summary tables that are an expanded representation of each data hierarchy. With these tables, the Db2 optimizer can much more efficiently join to other data tables, and so deliver suitable performance. In this section, we will show how to make these summary tables and, because it is a prerequisite, also show how to ensure that the related base tables do not have recursive data structures. Two solutions will be described: One that is simple and efficient, but which stops updates to key values. And another that imposes fewer constraints, but which is a bit more complicated.
17.3.2. Limited Update Solution
Below on the left is a hierarchy of data items. This is a typical unbalanced, non-recursive data hierarchy. In the center is a normalized representation of this hierarchy. The only thing that is perhaps a little unusual here is that an item at the top of a hierarchy (e.g. AAA) is deemed to be a parent of itself. On the right is an exploded representation of the same hierarchy.
AAA | BBB | +-----+ | | CCC EEE | DDD
HIERARCHY#1
KEYY | PKEY | DATA |
---|---|---|
AAA |
AAA |
SOME DATA |
BBB |
AAA |
MORE DATA |
CCC |
BBB |
MORE JUNK |
DDD |
CCC |
MORE JUNK |
EEE |
BBB |
JUNK DATA |
EXPLODED#1
PKEY | CKEY | LVL |
---|---|---|
AAA |
AAA |
0 |
AAA |
BBB |
1 |
AAA |
CCC |
2 |
AAA |
DDD |
3 |
AAA |
EEE |
2 |
BBB |
BBB |
0 |
BBB |
CCC |
1 |
BBB |
DDD |
2 |
BBB |
EEE |
1 |
CCC |
CCC |
0 |
CCC |
DDD |
1 |
DDD |
DDD |
0 |
EEE |
EEE |
0 |
Below is the CREATE code for the above normalized table and a dependent trigger:
CREATE TABLE hierarchy#1
( keyy CHAR(3) NOT NULL
, pkey CHAR(3) NOT NULL
, data VARCHAR(10)
, CONSTRAINT hierarchy11 PRIMARY KEY(keyy)
, CONSTRAINT hierarchy12 FOREIGN KEY(pkey)
REFERENCES hierarchy#1 (keyy) ON DELETE CASCADE);
CREATE TRIGGER HIR#1_UPD
NO CASCADE BEFORE UPDATE OF pkey ON hierarchy#1
REFERENCING NEW AS NNN
OLD AS OOO
FOR EACH ROW MODE Db2SQL
WHEN (NNN.pkey <> OOO.pkey)
SIGNAL SQLSTATE '70001' ('CAN NOT UPDATE pkey');
Note the following:
-
The KEYY column is the primary key, which ensures that each value must be unique, and that this field can not be updated.
-
The PKEY column is a foreign key of the KEYY column. This means that this field must always refer to a valid KEYY value. This value can either be in another row (if the new row is being inserted at the bottom of an existing hierarchy), or in the new row itself (if a new independent data hierarchy is being established).
-
The ON DELETE CASCADE referential integrity rule ensures that when a row is deleted, all dependent rows are also deleted.
-
The TRIGGER prevents any updates to the PKEY column. This is a BEFORE trigger, which means that it stops the update before it is applied to the database.
All of the above rules and restrictions act to prevent either an insert or an update for ever acting on any row that is not at the bottom of a hierarchy. Consequently, it is not possible for a hierarchy to ever exist that contains a loop of multiple data items.
17.3.3. Creating an Exploded Equivalent
Once we have ensured that the above table can never have recursive data structures, we can define a dependent table that holds an exploded version of the same hierarchy. Triggers will be used to keep the two tables in sync. Here is the CREATE code for the table:
CREATE TABLE exploded#1
( pkey CHAR(4) NOT NULL
, ckey CHAR(4) NOT NULL
, lvl SMALLINT NOT NULL
, PRIMARY KEY(pkey,ckey));
The following trigger deletes all dependent rows from the exploded table whenever a row is deleted from the hierarchy table:
CREATE TRIGGER EXP#1_DEL
AFTER DELETE ON hierarchy#1
REFERENCING OLD AS OOO
FOR EACH ROW MODE Db2SQL
DELETE FROM exploded#1
WHERE ckey = OOO.keyy;
The next trigger is run every time a row is inserted into the hierarchy table. It uses recursive code to scan the hierarchy table upwards, looking for all parents of the new row. The result set is then inserted into the exploded table:
CREATE TRIGGER EXP#1_INS
AFTER INSERT ON hierarchy#1
REFERENCING NEW AS NNN
FOR EACH ROW MODE Db2SQL
INSERT INTO exploded#1
WITH temp(pkey, ckey, lvl) AS
(VALUES (NNN.keyy
, NNN.keyy
, 0)
UNION ALL
SELECT N.pkey
, NNN.keyy
, T.lvl +1
FROM temp T
, hierarchy#1 N
WHERE N.keyy = T.pkey
AND N.keyy <> N.pkey)
SELECT *
FROM temp;
HIERARCHY#1
KEYY | PKEY | DATA |
---|---|---|
AAA |
AAA |
S… |
BBB |
AAA |
M… |
CCC |
BBB |
M… |
DDD |
CCC |
M… |
EEE |
BBB |
J… |
EXPLODED#1
PKEY | CKEY | LVL |
---|---|---|
AAA |
AAA |
0 |
AAA |
BBB |
1 |
AAA |
CCC |
2 |
AAA |
DDD |
3 |
AAA |
EEE |
2 |
BBB |
BBB |
0 |
BBB |
CCC |
1 |
BBB |
DDD |
2 |
BBB |
EEE |
1 |
CCC |
CCC |
0 |
CCC |
DDD |
1 |
DDD |
DDD |
0 |
EEE |
EEE |
0 |
There is no update trigger because updates are not allowed to the hierarchy table.
17.3.4. Querying the Exploded Table
Once supplied with suitable indexes, the exploded table can be queried like any other table. It will always return the current state of the data in the related hierarchy table.
SELECT *
FROM exploded#1
WHERE pkey = :host-var
ORDER BY pkey
, ckey
, lvl;
17.3.5. Full Update Solution
Not all applications want to limit updates to the data hierarchy as was done above. In particular, they may want the user to be able to move an object, and all its dependents, from one valid point (in a data hierarchy) to another. This means that we cannot prevent valid updates to the PKEY value. Below is the CREATE statement for a second hierarchy table. The only difference between this table and the previous one is that there is now an ON UPDATE RESTRICT clause. This prevents updates to PKEY that do not point to a valid KEYY value – either in another row, or in the row being updated:
CREATE TABLE hierarchy#2
( keyy CHAR(3) NOT NULL
, pkey CHAR(3) NOT NULL
, data VARCHAR(10)
, CONSTRAINT NO_loopS21 PRIMARY KEY(keyy)
, CONSTRAINT NO_loopS22 FOREIGN KEY(pkey)
REFERENCES hierarchy#2 (keyy) ON DELETE CASCADE
ON UPDATE RESTRICT);
The previous hierarchy table came with a trigger that prevented all updates to the PKEY field. This table comes instead with a trigger than checks to see that such updates do not result in a recursive data structure. It starts out at the changed row, then works upwards through the chain of PKEY values. If it ever comes back to the original row, it flags an error:
CREATE TRIGGER HIR#2_UPD
NO CASCADE BEFORE UPDATE OF pkey ON hierarchy#2
REFERENCING NEW AS NNN
OLD AS OOO
FOR EACH ROW MODE Db2SQL
WHEN (NNN.pkey <> OOO.pkey
AND NNN.pkey <> NNN.keyy)
WITH temp (keyy, pkey) AS
(VALUES (NNN.keyy
, NNN.pkey)
UNION ALL
SELECT LP2.keyy
, CASE
WHEN LP2.keyy = NNN.keyy THEN RAISE_ERROR('70001','LOOP FOUND')
ELSE LP2.pkey
END
FROM hierarchy#2 LP2
, temp TMP
WHERE TMP.pkey = LP2.keyy
AND TMP.keyy <> TMP.pkey)
SELECT *
FROM temp;
HIERARCHY#2
KEYY | PKEY | DATA |
---|---|---|
AAA |
AAA |
S… |
BBB |
AAA |
M… |
CCC |
BBB |
M… |
DDD |
CCC |
M… |
EEE |
BBB |
J… |
The above is a BEFORE trigger, which means that it gets run before the change is applied to the database. By contrast, the triggers that maintain the exploded table are all AFTER triggers. In general, one uses before triggers check for data validity, while after triggers are used to propagate changes. |
17.3.6. Creating an Exploded Equivalent
The following exploded table is exactly the same as the previous. It will be maintained in sync with changes to the related hierarchy table:
CREATE TABLE exploded#2
( pkey CHAR(4) NOT NULL
, ckey CHAR(4) NOT NULL
, lvl SMALLINT NOT NULL
, PRIMARY KEY(pkey,ckey));
Three triggers are required to maintain the exploded table in sync with the related hierarchy table. The first two, which handle deletes and inserts, are the same as what were used previously. The last, which handles updates, is new (and quite tricky). The following trigger deletes all dependent rows from the exploded table whenever a row is deleted from the hierarchy table:
CREATE TRIGGER EXP#2_DEL
AFTER DELETE ON hierarchy#2
REFERENCING OLD AS OOO
FOR EACH ROW MODE Db2SQL
DELETE FROM exploded#2
WHERE ckey = OOO.keyy;
The next trigger is run every time a row is inserted into the hierarchy table. It uses recursive code to scan the hierarchy table upwards, looking for all parents of the new row. The resultset is then inserted into the exploded table:
CREATE TRIGGER EXP#2_INS
AFTER INSERT ON hierarchy#2
REFERENCING NEW AS NNN
FOR EACH ROW MODE Db2SQL
INSERT INTO exploded#2
WITH temp(pkey, ckey, lvl) AS
(SELECT NNN.keyy
, NNN.keyy
, 0
FROM hierarchy#2
WHERE keyy = NNN.keyy
UNION ALL
SELECT N.pkey
, NNN.keyy
, T.lvl +1
FROM temp T
, hierarchy#2 N
WHERE N.keyy = T.pkey
AND N.keyy <> N.pkey)
SELECT *
FROM temp;
HIERARCHY#2
KEYY | PKEY | DATA |
---|---|---|
AAA |
AAA |
S… |
BBB |
AAA |
M… |
CCC |
BBB |
M… |
DDD |
CCC |
M… |
EEE |
BBB |
J… |
EXPLODED#2
PKEY | CKEY | LVL |
---|---|---|
AAA |
AAA |
0 |
AAA |
BBB |
1 |
AAA |
CCC |
2 |
AAA |
DDD |
3 |
AAA |
EEE |
2 |
BBB |
BBB |
0 |
BBB |
CCC |
1 |
BBB |
DDD |
2 |
BBB |
EEE |
1 |
CCC |
CCC |
0 |
CCC |
DDD |
1 |
DDD |
DDD |
0 |
EEE |
EEE |
0 |
The next trigger is run every time a PKEY value is updated in the hierarchy table. It deletes and then reinserts all rows pertaining to the updated object, and all it’s dependents. The code goes as follows: Delete all rows that point to children of the row being updated. The row being updated is also considered to be a child. In the following insert, first use recursion to get a list of all of the children of the row that has been updated. Then work out the relationships between all of these children and all of their parents. Insert this second result-set back into the exploded table.
CREATE TRIGGER EXP#2_UPD
AFTER UPDATE OF pkey ON hierarchy#2
REFERENCING OLD AS OOO
NEW AS NNN
FOR EACH ROW MODE Db2SQL
BEGIN ATOMIC
DELETE FROM exploded#2
WHERE ckey IN
(SELECT ckey
FROM exploded#2
WHERE pkey = OOO.keyy);
INSERT INTO exploded#2
WITH temp1(ckey) AS
(VALUES (NNN.keyy)
UNION ALL
SELECT N.keyy
FROM temp1 T
, hierarchy#2 N
WHERE N.pkey = T.ckey
AND N.pkey <> N.keyy)
, temp2(pkey, ckey, lvl) AS
(SELECT ckey
, ckey
, 0
FROM temp1
UNION ALL
SELECT N.pkey
, T.ckey
, T.lvl +1
FROM temp2 T
, hierarchy#2 N
WHERE N.keyy = T.pkey
AND N.keyy <> N.pkey)
SELECT *
FROM temp2;
END
The above trigger lacks a statement terminator because it contains atomic SQL, which means that the semi-colon can not be used. Choose anything you like. |
17.3.7. Querying the Exploded Table
Once supplied with suitable indexes, the exploded table can be queried like any other table. It will always return the current state of the data in the related hierarchy table.
SELECT *
FROM exploded#2
WHERE pkey = :host-var
ORDER BY pkey
, ckey
, lvl;
Below are some suggested indexes:
-
PKEY, CKEY (already defined as part of the primary key).
-
CKEY, PKEY (useful when joining to this table).
18. Triggers
A trigger initiates an action whenever a row, or set of rows, is changed. The change can be either an insert, update or delete.
The Db2 Application Development Guide: Programming Server Applications is an excellent source of information on using triggers. The SQL Reference has all the basics. |
18.1. Trigger Types
-
A BEFORE trigger is run before the row is changed. It is typically used to change the values being entered (e.g. set a field to the current date), or to flag an error. It cannot be used to initiate changes in other tables.
-
An AFTER trigger is run after the row is changed. It can do everything a before trigger can do, plus modify data in other tables or systems (e.g. it can insert a row into an audit table after an update).
-
An INSTEAD OF trigger is used in a view to do something instead of the action that the user intended (e.g. do an insert instead of an update). There can be only one instead of trigger per possible DML type on a given view.
See the chapter titled Retaining a Record for a sample application that uses INSTEAD OF triggers to record all changes to the data in a set of tables. |
18.2. Action Type
Each trigger applies to a single kind of DML action (i.e. insert, update, or delete). With the exception of instead of triggers, there can be as many triggers per action and per table as desired. An update trigger can be limited to changes to certain columns.
Object Type
-
A table can have both BEFORE and AFTER triggers. The former have to be defined FOR EACH ROW.
-
A view can have INSTEAD OF triggers (up to three - one per DML type).
Referencing
In the body of the trigger the object being changed can be referenced using a set of optional correlation names:
-
OLD refers to each individual row before the change (does not apply to an insert).
-
NEW refers to each individual row after the change (does not apply to a delete).
-
OLD_TABLE refers to the set of rows before the change (does not apply to an insert).
-
NEW_TABLE refers to the set of rows after the change (does to apply to a delete).
Application Scope
-
A trigger defined FOR EACH STATEMENT is invoked once per statement.
-
A trigger defined FOR EACH ROW is invoked once per individual row changed.
If one defines two FOR EACH ROW triggers, the first is applied for all rows before the second is run. To do two separate actions per row, one at a time, one has to define a single trigger that includes the two actions in a single compound SQL statement. |
When Check
One can optionally include some predicates so that the body of the trigger is only invoked when certain conditions are true.
Trigger Usage
A trigger can be invoked whenever one of the following occurs:
-
A row in a table is inserted, updated, or deleted.
-
An (implied) row in a view is inserted, updated, or deleted.
-
A referential integrity rule on a related table causes a cascading change (i.e. delete or set null) to the triggered table.
-
A trigger on an unrelated table or view is invoked - and that trigger changes rows in the triggered table.
If no rows are changed, a trigger defined FOR EACH ROW is not run, while a trigger defined FOR EACH STATEMENT is still run. To prevent the latter from doing anything when this happens, add a suitable WHEN check.
18.3. Trigger Examples
This section uses a set of simple sample tables to illustrate general trigger usage.
18.3.1. Sample Tables
CREATE TABLE cust_balance
( cust# INTEGER NOT NULL
GENERATED ALWAYS AS IDENTITY
, status CHAR(2) NOT NULL
, balance DECIMAL(18,2) NOT NULL
, num_trans INTEGER NOT NULL
, cur_ts TIMESTAMP NOT NULL
, PRIMARY KEY (cust#));
-- Every state of a row in the balance table will be recorded in the history table.
CREATE TABLE cust_history
( cust# INTEGER NOT NULL
, trans# INTEGER NOT NULL
, balance DECIMAL(18,2) NOT NULL
, bgn_ts TIMESTAMP NOT NULL
, end_ts TIMESTAMP NOT NULL
, PRIMARY KEY (cust#, bgn_ts));
-- Every valid change to the balance table will be recorded in the transaction table.
CREATE TABLE cust_trans
( min_cust# INTEGER
, max_cust# INTEGER
, rows_tot INTEGER NOT NULL
, change_val DECIMAL(18,2)
, change_type CHAR(1) NOT NULL
, cur_ts TIMESTAMP NOT NULL
, PRIMARY KEY (cur_ts));
18.4. Before Row Triggers - Set Values
The first trigger below overrides whatever the user enters during the insert, and before the row is inserted, sets both the cur-ts and number-of-trans columns to their correct values:
CREATE TRIGGER cust_bal_ins1
NO CASCADE BEFORE INSERT
ON cust_balance
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
SET nnn.cur_ts = CURRENT TIMESTAMP
, nnn.num_trans = 1;
The following trigger does the same before an update:
CREATE TRIGGER cust_bal_upd1
NO CASCADE BEFORE UPDATE
ON cust_balance
REFERENCING NEW AS nnn
OLD AS ooo
FOR EACH ROW
MODE Db2SQL
SET nnn.cur_ts = CURRENT TIMESTAMP
, nnn.num_trans = ooo.num_trans + 1;
18.5. Before Row Trigger - Signal Error
The next trigger will flag an error (and thus fail the update) if the customer balance is reduced by too large a value:
CREATE TRIGGER cust_bal_upd2
NO CASCADE BEFORE UPDATE OF balance
ON cust_balance
REFERENCING NEW AS nnn
OLD AS ooo
FOR EACH ROW
MODE Db2SQL
WHEN (ooo.balance - nnn.balance > 1000)
SIGNAL SQLSTATE VALUE '71001'
SET MESSAGE_TEXT = 'Cannot withdraw > 1000';
18.6. After Row Triggers - Record Data States
The three triggers in this section record the state of the data in the customer table. The first is invoked after each insert. It records the new data in the customer-history table:
CREATE TRIGGER cust_his_ins1
AFTER INSERT ON cust_balance
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
INSERT INTO cust_history
VALUES (nnn.cust#
, nnn.num_trans
, nnn.balance
, nnn.cur_ts
, '9999-12-31-24.00.00');
The next trigger is invoked after every update of a row in the customer table. It first runs an update (of the old history row), and then does an insert. Because this trigger uses a compound SQL statement, it cannot use the semi-colon as the statement delimiter:
CREATE TRIGGER cust_his_upd1
AFTER UPDATE ON cust_balance
REFERENCING OLD AS ooo
NEW AS nnn
FOR EACH ROW
MODE Db2SQL
BEGIN ATOMIC
UPDATE cust_history
SET end_ts = CURRENT TIMESTAMP
WHERE cust# = ooo.cust#
AND bgn_ts = ooo.cur_ts;
INSERT INTO cust_history
VALUES (nnn.cust#
, nnn.num_trans
, nnn.balance
, nnn.cur_ts
, '9999-12-31-24.00.00');
END
Notes
-
The above trigger relies on the fact that the customer-number cannot change (note: it is generated always) to link the two rows in the history table together. In other words, the old row will always have the same customer-number as the new row.
-
The above also trigger relies on the presence of the cust_bal_upd1 before trigger (see Trigger Examples) to set the nnn.cur_ts value to the current timestamp.
The final trigger records a delete by doing an update to the history table:
CREATE TRIGGER cust_his_del1
AFTER DELETE ON cust_balance
REFERENCING OLD AS ooo
FOR EACH ROW
MODE Db2SQL
UPDATE cust_history
SET end_ts = CURRENT TIMESTAMP
WHERE cust# = ooo.cust#
AND bgn_ts = ooo.cur_ts;
18.7. After Statement Triggers - Record Changes
The following three triggers record every type of change (i.e. insert, update, or delete) to any row, or set of rows (including an empty set) in the customer table. They all run an insert that records the type and number of rows changed:
CREATE TRIGGER trans_his_ins1
AFTER INSERT ON cust_balance
REFERENCING NEW_TABLE AS newtab
FOR EACH STATEMENT
MODE Db2SQL
INSERT INTO cust_trans
SELECT MIN(cust#)
, MAX(cust#)
, COUNT(*)
, SUM(balance)
, 'I'
, CURRENT TIMESTAMP
FROM newtab;
CREATE TRIGGER trans_his_upd1
AFTER UPDATE ON cust_balance
REFERENCING OLD_TABLE AS oldtab
NEW_TABLE AS newtab
FOR EACH STATEMENT
MODE Db2SQL
INSERT INTO cust_trans
SELECT MIN(nt.cust#)
, MAX(nt.cust#)
, COUNT(*)
, SUM(nt.balance - ot.balance)
, 'U'
, CURRENT TIMESTAMP
FROM oldtab ot
, newtab nt
WHERE ot.cust# = nt.cust#;
CREATE TRIGGER trans_his_del1
AFTER DELETE ON cust_balance
REFERENCING OLD_TABLE AS oldtab
FOR EACH STATEMENT
MODE Db2SQL
INSERT INTO cust_trans
SELECT MIN(cust#)
, MAX(cust#)
, COUNT(*)
, SUM(balance)
, 'D'
, CURRENT TIMESTAMP
FROM oldtab;
Notes
-
If the DML statement changes no rows, the OLD or NEW table referenced by the trigger will be empty, but still exist, and a SELECT COUNT(*) on the (empty) table will return a zero, which will then be inserted.
-
Any DML statements that failed (e.g. stopped by the before trigger), or that were subsequently rolled back, will not be recorded in the transaction table.
18.7.1. Examples of Usage
The following DML statements were run against the customer table:
INSERT INTO cust_balance (status, balance) VALUES ('C',123.45);
INSERT INTO cust_balance (status, balance) VALUES ('C',000.00);
INSERT INTO cust_balance (status, balance) VALUES ('D', -1.00);
UPDATE cust_balance
SET balance = balance + 123
WHERE cust# <= 2;
UPDATE cust_balance
SET balance = balance * -1
WHERE cust# = -1;
UPDATE cust_balance
SET balance = balance - 123
WHERE cust# = 1;
DELETE FROM cust_balance
WHERE cust# = 3;
18.8. Tables After DML
At the end of the above, the three tables had the following data:
CUST# | STATUS | BALANCE | NUM_TRANS | CUR_TS |
---|---|---|---|---|
1 |
C |
123.45 |
3 |
2005-05-31-19.58.46.096000 |
2 |
C |
123.00 |
2 |
2005-05-31-19.58.46.034000 |
CUST# | TRANS# | BALANCE | BGN_TS | END_TS |
---|---|---|---|---|
1 |
1 |
123.45 |
2005-05-31-19.58.45.971000 |
2005-05-31-19.58.46.034000 |
1 |
2 |
246.45 |
2005-05-31-19.58.46.034000 |
2005-05-31-19.58.46.096000 |
1 |
3 |
123.45 |
2005-05-31-19.58.46.096000 |
9999-12-31-24.00.00.000000 |
2 |
1 |
0.00 |
2005-05-31-19.58.45.987000 |
2005-05-31-19.58.46.034000 |
2 |
2 |
123.00 |
2005-05-31-19.58.46.034000 |
9999-12-31-24.00.00.000000 |
3 |
1 |
-1.00 |
2005-05-31-19.58.46.003000 |
2005-05-31-19.58.46.096003 |
MIN_CUST# | MAX_CUST# | ROWS | CHANGE_VAL | CHANGE_TYPE | CUR_TS |
---|---|---|---|---|---|
1 |
1 |
1 |
123.45 |
I |
2005-05-31-19.58.45.971000 |
2 |
2 |
1 |
0.00 |
I |
2005-05-31-19.58.45.987000 |
3 |
3 |
1 |
-1.00 |
I |
2005-05-31-19.58.46.003000 |
1 |
2 |
2 |
246.00 |
U |
2005-05-31-19.58.46.034000 |
- |
- |
0 |
- |
U |
2005-05-31-19.58.46.065000 |
1 |
1 |
1 |
-123.00 |
U |
2005-05-31-19.58.46.096000 |
3 |
3 |
1 |
1.00 |
D |
2005-05-31-19.58.46.096003 |
19. Protecting Your Data
There is no use having a database if the data in it is not valid. This chapter introduces some of the tools that exist in Db2 to enable one to ensure the validity of the data in your application.
Issues Covered
-
Enforcing field uniqueness.
-
Enforcing field value ranges.
-
Generating key and values.
-
Maintaining summary columns.
-
Enforcing relationships between and within tables.
-
Creating columns that have current timestamp of last change.
Issues Not Covered
-
Data access authorization.
-
Recovery and backup.
19.1. Sample Application
Consider the following two tables, which make up a very simple application:
CREATE TABLE customer_balance
( cust_id INTEGER
, cust_name VARCHAR(20)
, cust_sex CHAR(1)
, num_sales SMALLINT
, total_sales DECIMAL(12, 2)
, master_cust_id INTEGER
, cust_insert_ts TIMESTAMP
, cust_update_ts TIMESTAMP);
CREATE TABLE us_sales
( invoice# INTEGER
, cust_id INTEGER
, sale_value DECIMAL(18, 2)
, sale_insert_ts TIMESTAMP
, sale_update_ts TIMESTAMP);
19.2. Customer Balance Table
We want Db2 to enforce the following business rules:
-
CUST_ID will be a unique positive integer value, always ascending, never reused, and automatically generated by Db2. This field cannot be updated by a user.
-
CUST_NAME has the customer name. It can be anything, but not blank.
-
CUST_SEX must be either "M" or "F".
-
NUM_SALES will have a count of the sales (for the customer), as recorded in the related US-sales table. The value will be automatically maintained by Db2. It cannot be updated directly by a user.
-
TOTAL_SALES will have the sum sales (in US dollars) for the customer. The value will be automatically updated by Db2. It cannot be updated directly by a user.
-
MASTER_CUST_ID will have, if there exists, the customer-ID of the customer that this customer is a dependent of. If there is no master customer, the value is null. If the master customer is deleted, this row will also be deleted (if possible).
-
CUST_INSERT_TS has the timestamp when the row was inserted. The value is automatically generated by Db2. Any attempt to change will induce an error.
-
CUST_UPDATE_TS has the timestamp when the row, or a dependent US_SALES row, was last updated by a user. The value is automatically generated by Db2. Any attempt to change directly will induce an error.
-
A row can only be deleted when there are no corresponding rows in the US-sales table (i.e. for the same customer).
US Sales Table
We want Db2 to enforce the following business rules:
-
INVOICE#: will be a unique ascending integer value. The uniqueness will apply to the US-sales table, plus any international sales tables (i.e. to more than one table).
-
CUST_ID is the customer ID, as recorded in the customer-balance table. No row can be inserted into the US-sales table except that there is a corresponding row in the customerbalance table. Once inserted, this value cannot be updated.
-
SALE_VALUE is the value of the sale, in US dollars. When a row is inserted, this value is added to the related total-sales value in the customer-balance table. If the value is subsequently updated, the total-sales value is maintained in sync.
-
SALE_INSERT_TS has the timestamp when the row was inserted. The value is automatically generated by Db2. Any attempt to change will induce an error.
-
SALE_UPDATE_TS has the timestamp when the row was last updated. The value is automatically generated by Db2. Any attempt to change will induce an error.
-
Deleting a row from the US-sales table has no impact on the customer-balance table (i.e. the total-sales is not decremented). But a row can only be deleted from the latter when there are no more related rows in the US-sales table.
19.3. Enforcement Tools
To enforce the above business rules, we are going to have to use:
-
Unique indexes.
-
Secondary non-unique indexes (needed for performance).
-
Primary and foreign key definitions.
-
User-defined distinct data types.
-
Nulls-allowed and not-null columns.
-
Column value constraint rules.
-
Before and after triggers.
-
Generated row change timestamps.
-
Distinct Data Types
Two of the fields are to contain US dollars, the implication being the data in these columns should not be combined with columns that contain Euros, or Japanese Yen, or my shoe size. To this end, we will define a distinct data type for US dollars:
CREATE DISTINCT TYPE us_dollars
AS decimal(18, 2) WITH COMPARISONS;
See Distinct Types for a more detailed discussion of this topic.
19.4. Customer-Balance Table
Now that we have defined the data type, we can create our first table:
CREATE TABLE customer_balance
( cust_id INTEGER NOT NULL
GENERATED ALWAYS AS IDENTITY
( START WITH 1
, INCREMENT BY 1
, NO CYCLE
, NO CACHE)
, cust_name VARCHAR(20) NOT NULL
, cust_sex CHAR(1) NOT NULL
, num_sales SMALLINT NOT NULL
, total_sales us_dollars NOT NULL
, master_cust_id INTEGER
, cust_insert_ts TIMESTAMP NOT NULL
, cust_update_ts TIMESTAMP NOT NULL
, PRIMARY KEY (cust_id)
, CONSTRAINT c1 CHECK (cust_name <> '')
, CONSTRAINT c2 CHECK (cust_sex = 'F' OR cust_sex = 'M')
, CONSTRAINT c3 FOREIGN KEY (master_cust_id)
REFERENCES customer_balance (cust_id)
ON DELETE CASCADE);
The following business rules are enforced above:
-
The customer-ID is defined as an identity column (see Identity Columns and Sequences), which means that the value is automatically generated by Db2 using the rules given. The field cannot be updated by the user.
-
The customer-ID is defined as the primary key, which automatically generates a unique index on the field, and also enables us to reference the field using a referential integrity rule. Being a primary key prevents updates, but we had already prevented them because the field is an identity column.
-
The total-sales column uses the type us-dollars.
-
Constraints C1 and C2 enforce two data validation rules.
-
Constraint C3 relates the current row to a master customer, if one exists. Furthermore, if the master customer is deleted, this row is also deleted.
-
All of the columns, except for the master-customer-id, are defined as NOT NULL, which means that a value must be provided. We still have several more business rules to enforce - relating to automatically updating fields and/or preventing user updates. These will be enforced using triggers.
19.5. US-Sales Table
Now for the related US-sales table:
CREATE TABLE us_sales
( invoice# INTEGER NOT NULL
, cust_id INTEGER NOT NULL
, sale_value us_dollars NOT NULL
, sale_insert_ts TIMESTAMP NOT NULL
, sale_update_ts TIMESTAMP NOT NULL
GENERATED ALWAYS
FOR EACH ROW ON UPDATE
AS ROW CHANGE TIMESTAMP
, PRIMARY KEY (invoice#)
, CONSTRAINT u1 CHECK (sale_value > us_dollars(0))
, CONSTRAINT u2 FOREIGN KEY (cust_id)
REFERENCES customer_balance
ON DELETE RESTRICT);
COMMIT;
CREATE INDEX us_sales_cust ON us_sales (cust_id);
The following business rules are enforced above:
-
The invoice# is defined as the primary key, which automatically generates a unique index on the field, and also prevents updates.
-
The sale-value uses the type us-dollars.
-
Constraint U1 checks that the sale-value is always greater than zero.
-
Constraint U2 checks that the customer-ID exists in the customer-balance table, and also prevents rows from being deleted from the latter if there is a related row in this table.
-
All of the columns are defined as NOT NULL, so a value must be provided for each.
-
A secondary non-unique index is defined on customer-ID, so that deletes to the customerbalance table (which require checking this table for related customer-ID rows) are as efficient as possible.
-
The CUST_UPDATE_TS column is generated always (by Db2) and gets a unique value that is the current timestamp.
19.5.1. Generated Always Timestamp Columns
A TIMESTAMP column that is defined as GENERATED ALWAYS will get a value that is unique for all rows in the table. This value will usually be the CURRENT TIMESTAMP of the last insert or update of the row. However, if more than row was inserted or updated in a single stmt, the secondary rows (updated) will get a value that is equal to the CURRENT TIMESTAMP special register, plus "n" microseconds, where "n" goes up in steps of 1. One consequence of the above logic is that some rows changed will get a timestamp value that is ahead of the CURRENT TIMESTAMP special register. This can cause problems if one is relying on this value to find all rows that were changed before the start of the query. To illustrate, imagine that one inserted multiple rows (in a single insert) into the US_SALES table, and then immediately ran the following query:
SELECT *
FROM us_sales
WHERE sale_update_ts <= CURRENT TIMESTAMP;
In some environments (e.g. Windows) the CURRENT TIMESTAMP special register value may be the same from one stmt to the next. If this happens, the above query will find the first row just inserted, but not any subsequent rows, because their SALE_UPDATE_TS value will be greater than the CURRENT TIMESTAMP special register.
Certain restrictions apply:
-
Only one TIMESTAMP column can be defined GENERATED ALWAYS per table. The column must be defined NOT NULL.
-
The TIMESTAMP column is updated, even if no other value in the row changes. So if one does an update that sets SALE_VALUE = SALE_VALUE + 0, the SALE_UPDATE_TS column will be updated on all matching rows.
The ROW CHANGE TIMESTAMP special register can be used get the last time that the row was updated, even when one does not know the name of the column that holds this data:
SELECT ROW CHANGE TIMESTAMP FOR us_sales
FROM us_sales
WHERE invoice# = 5;
The (unique) TIMESTAMP value obtained above can be used to validate that the target row has not been updated when a subsequent UPDATE is done:
UPDATE us_sales
SET sale_value = DECIMAL(sale_value) + 1
WHERE invoice# = 5
AND ROW CHANGE TIMESTAMP for us_sales = '2007-11-10-01.02.03';
19.5.2. Triggers
Triggers can sometimes be quite complex little programs. If coded incorrectly, they can do an amazing amount of damage. As such, it pays to learn quite a lot before using them. Below are some very brief notes, but please refer to the official Db2 documentation for a more detailed description. See also Triggers for a brief chapter on triggers. Individual triggers are defined on a table, and for a particular type of DML statement:
-
Insert
-
Update
-
Delete
A trigger can be invoked once per:
-
Row changed.
-
Statement run.
A trigger can be invoked:
-
Before the change is made.
-
After the change is made.
Before triggers change input values before they are entered into the table and/or flag an error. After triggers do things after the row is changed. They may make more changes (to the target table, or to other tables), induce an error, or invoke an external program. SQL statements that select the changes made by DML (see Insert Examples) cannot see the changes made by an after trigger if those changes impact the rows just changed. The action of one "after" trigger can invoke other triggers, which may then invoke other triggers, and so on. Before triggers cannot do this because they can only act upon the input values of the DML statement that invoked them. When there are multiple triggers for a single table/action, each trigger is run for all rows before the next trigger is invoked - even if defined "for each row". Triggers are invoked in the order that they were created.
19.5.3. Customer-Balance - Insert Trigger
For each row inserted into the Customer-Balance table we need to do the following:
-
Set the num-sales to zero.
-
Set the total-sales to zero.
-
Set the update-timestamp to the current timestamp.
-
Set the insert-timestamp to the current timestamp.
All of this can be done using a simple before trigger:
CREATE TRIGGER cust_balance_ins1
NO CASCADE BEFORE INSERT
ON customer_balance
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
SET nnn.num_sales = 0
, nnn.total_sales = 0
, nnn.cust_insert_ts = CURRENT TIMESTAMP
, nnn.cust_update_ts = CURRENT TIMESTAMP;
19.5.4. Customer-Balance - Update Triggers
For each row updated in the Customer-Balance table we need to do:
-
Set the update-timestamp to the current timestamp.
-
Prevent updates to the insert-timestamp, or sales fields.
-
We can use the following trigger to maintain the update-timestamp:
CREATE TRIGGER cust_balance_upd1
NO CASCADE BEFORE UPDATE OF cust_update_ts
ON customer_balance
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
SET nnn.cust_update_ts = CURRENT TIMESTAMP;
We can prevent updates to the insert-timestamp with the following trigger:
CREATE TRIGGER cust_balance_upd2
NO CASCADE BEFORE UPDATE OF cust_insert_ts
ON customer_balance
FOR EACH ROW
MODE Db2SQL
SIGNAL SQLSTATE VALUE '71001'
SET MESSAGE_TEXT = 'Cannot update CUST insert-ts';
We don’t want users to update the two sales counters directly. But the two fields do have to be updated (by a trigger) whenever there is a change to the us-sales table. The solution is to have a trigger that prevents updates if there is no corresponding row in the us-sales table where the update-timestamp is greater than or equal to the current timestamp:
CREATE TRIGGER cust_balance_upd3
NO CASCADE BEFORE UPDATE OF num_sales
, total_sales
ON customer_balance
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
WHEN (CURRENT TIMESTAMP > (SELECT MAX(sss.sale_update_ts)
FROM us_sales sss
WHERE nnn.cust_id = sss.cust_id))
SIGNAL SQLSTATE VALUE '71001'
SET MESSAGE_TEXT = 'Feilds only updated via US-Sales';
19.5.5. US-Sales - Insert Triggers
For each row inserted into the US-sales table we need to do the following:
-
Determine the invoice-number, which is unique over multiple tables.
-
Set the update-timestamp to the current timestamp.
-
Set the insert-timestamp to the current timestamp.
-
Add the sale-value to the existing total-sales in the customer-balance table.
-
Increment the num-sales counter in the customer-balance table.
The invoice-number is supposed to be unique over several tables, so we cannot generate it using an identity column. Instead, we have to call the following external sequence:
CREATE SEQUENCE us_sales_seq AS INTEGER
START WITH 1
INCREMENT BY 1
NO CYCLE
NO CACHE
ORDER;
Once we have the above, the following trigger will take of the first three items:
CREATE TRIGGER us_sales_ins1
NO CASCADE BEFORE INSERT
ON us_sales
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
SET nnn.invoice# = NEXTVAL FOR us_sales_seq
, nnn.sale_insert_ts = CURRENT TIMESTAMP;
We need to use an "after" trigger to maintain the two related values in the Customer-Balance table. This will invoke an update to change the target row:
CREATE TRIGGER sales_to_cust_ins1
AFTER INSERT ON us_sales
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
UPDATE customer_balance ccc
SET ccc.num_sales = ccc.num_sales + 1
, ccc.total_sales = DECIMAL(ccc.total_sales) + DECIMAL(nnn.sale_value)
WHERE ccc.cust_id = nnn.cust_id;
19.5.6. US-Sales - Update Triggers
For each row updated in the US-sales table we need to do the following:
-
Prevent the customer-ID or insert-timestamp from being updated.
-
Propagate the change to the sale-value to the total-sales in the customer-balance table.
The next trigger prevents updates to the Customer-ID and insert-timestamp:
CREATE TRIGGER us_sales_upd2
NO CASCADE BEFORE UPDATE OF cust_id
, sale_insert_ts
ON us_sales
FOR EACH ROW
MODE Db2SQL
SIGNAL SQLSTATE VALUE '71001'
SET MESSAGE_TEXT = 'Can only update sale_value';
We need to use an "after" trigger to maintain sales values in the Customer-Balance table:
CREATE TRIGGER sales_to_cust_upd1
AFTER UPDATE OF sale_value
ON us_sales
REFERENCING NEW AS nnn
OLD AS ooo
FOR EACH ROW
MODE Db2SQL
UPDATE customer_balance ccc
SET ccc.total_sales = DECIMAL(ccc.total_sales)
- DECIMAL(ooo.sale_value)
+ DECIMAL(nnn.sale_value)
WHERE ccc.cust_id = nnn.cust_id;
19.6. Conclusion
The above application will now have logically consistent data. There is, of course, nothing to prevent an authorized user from deleting all rows, but whatever rows are in the two tables will obey the business rules that we specified at the start.
Tools Used
-
Primary key - to enforce uniqueness, prevent updates, enable referential integrity.
-
Unique index - to enforce uniqueness.
-
Non-unique index - for performance during referential integrity check.
-
Sequence object - to automatically generate key values for multiple tables.
-
Identity column - to automatically generate key values for 1 table.
-
Not-null columns - to prevent use of null values.
-
Column constraints - to enforce basic domain-range rules.
-
Distinct types - to prevent one type of data from being combined with another type.
-
Referential integrity - to enforce relationships between rows/tables, and to enable cascading deletes when needed.
-
Before triggers - to prevent unwanted changes and set certain values.
-
After triggers - to propagate valid changes.
-
Automatically generated timestamp value that is always the current timestamp or (in the case of a multi-row update), the current timestamp plus a few microseconds.
20. Retaining a Record
This chapter was written back in the versions before V9.7. It is helpful showing how to use triggers for a specific use case but if you want to implement such a scenario in Db2 you should use the "Time Travel" feature. A chapter about the "Time Travel" will be written in one of the new versions of the book. |
This chapter will describe a rather complex table/view/trigger schema that will enable us to offer several features that are often asked for:
-
Record every change to the data in an application (auditing).
-
Show the state of the data, as it was, at any point in the past (historical analysis).
-
Follow the sequence of changes to any item (e.g. customer) in the database.
-
Do "what if" analysis by creating virtual copies of the real world, and then changing them as desired, without affecting the real-world data.
Some sample code to illustrate the above concepts will be described below.
20.1. Schema Design
Recording Changes
Below is a very simple table that records relevant customer data:
CREATE TABLE customer
( cust# INTEGER NOT NULL
, cust_name CHAR(10)
, cust_mgr CHAR(10)
, PRIMARY KEY(cust#));
One can insert, update, and delete rows in the above table. The latter two actions destroy data, and so are incompatible with using this table to see all (prior) states of the data. One way to record all states of the above table is to create a related customer-history table, and then to use triggers to copy all changes in the main table to the history table. Below is one example of such a history table:
CREATE TABLE customer_his
( cust# INTEGER NOT NULL
, cust_name CHAR(10)
, cust_mgr CHAR(10)
, cur_ts TIMESTAMP NOT NULL
, cur_actn CHAR(1) NOT NULL
, cur_user VARCHAR(10) NOT NULL
, prv_cust# INTEGER
, prv_ts TIMESTAMP
, PRIMARY KEY(cust#, cur_ts));
CREATE UNIQUE INDEX customer_his_x1
ON customer_his(cust#, prv_ts, cur_ts);
The secondary index shown above will make the following view processing, which looks for a row that replaces the current, much more efficient. |
20.2. Table Design
The history table has the same fields as the original Customer table, plus the following:
-
CUR-TS: The current timestamp of the change.
-
CUR-ACTN: The type of change (i.e. insert, update, or delete).
-
CUR-USER: The user who made the change (for auditing purposes).
-
PRV-CUST#: The previous customer number. This field enables one follow the sequence of changes for a given customer. The value is null if the action is an insert.
-
PRV-TS: The timestamp of the last time the row was changed (null for inserts).
Observe that this history table does not have an end-timestamp. Rather, each row points back to the one that it (optionally) replaces. One advantage of such a schema is that there can be a many-to-one relationship between any given row, and the row, or rows, that replace it. When we add versions into the mix, this will become important.
20.3. Triggers
Below is the relevant insert trigger. It replicates the new customer row in the history table, along with the new fields. Observe that the two "previous" fields are set to null:
CREATE TRIGGER customer_ins
AFTER INSERT ON customer
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
INSERT INTO customer_his
VALUES (nnn.cust#
, nnn.cust_name
, nnn.cust_mgr
, CURRENT TIMESTAMP
, 'I'
, USER
, NULL
, NULL);
Below is the update trigger. Because the customer table does not have a record of when it was last changed, we have to get this value from the history table - using a sub-query to find the most recent row:
CREATE TRIGGER customer_upd
AFTER UPDATE ON customer
REFERENCING NEW AS nnn
OLD AS ooo
FOR EACH ROW
MODE Db2SQL
INSERT INTO customer_his
VALUES (nnn.cust#
, nnn.cust_name
, nnn.cust_mgr
, CURRENT TIMESTAMP
, 'U'
, USER
, ooo.cust#
, (SELECT MAX(cur_ts)
FROM customer_his hhh
WHERE ooo.cust# = hhh.cust#));
Below is the delete trigger. It is similar to the update trigger, except that the action is different and we are under no obligation to copy over the old non-key-data columns - but we can if we wish:
CREATE TRIGGER customer_del
AFTER DELETE ON customer
REFERENCING OLD AS ooo
FOR EACH ROW
MODE Db2SQL
INSERT INTO customer_his
VALUES (ooo.cust#
, NULL
, NULL
, CURRENT TIMESTAMP
, 'D'
, USER
, ooo.cust#
, (SELECT MAX(cur_ts)
FROM customer_his hhh
WHERE ooo.cust# = hhh.cust#));
20.4. Views
We are now going to define a view that will let the user query the customer-history table – as if it were the ordinary customer table, but to look at the data as it was at any point in the past. To enable us to hide all the nasty SQL that is required to do this, we are going to ask that the user first enter a row into a profile table that has two columns:
-
The user’s Db2 USER value.
-
The point in time at which the user wants to see the customer data.
Here is the profile table definition:
CREATE TABLE profile
( user_id VARCHAR(10) NOT NULL
, bgn_ts TIMESTAMP NOT NULL DEFAULT '9999-12-31-24.00.00'
, PRIMARY KEY(user_id));
Below is a view that displays the customer data, as it was at the point in time represented by the timestamp in the profile table. The view shows all customer-history rows, as long as:
-
The action was not a delete.
-
The current-timestamp is ⇐ the profile timestamp.
-
There does not exist any row that "replaces" the current row (and that row has a current timestamp that is ⇐ to the profile timestamp).
Now for the code:
CREATE VIEW customer_vw AS
SELECT hhh.*
, ppp.bgn_ts
FROM customer_his hhh
, profile ppp
WHERE ppp.user_id = USER
AND hhh.cur_ts <= ppp.bgn_ts
AND hhh.cur_actn <> 'D'
AND NOT EXISTS
(SELECT *
FROM customer_his nnn
WHERE nnn.prv_cust# = hhh.cust#
AND nnn.prv_ts = hhh.cur_ts
AND nnn.cur_ts <= ppp.bgn_ts);
The above sample schema shows just one table, but it can easily be extended to support every table is a very large application. One could even write some scripts to make the creation of the history tables, triggers, and views, all but automatic.
20.5. Limitations
The above schema has the following limitations:
-
Every data table has to have a unique key.
-
The cost of every insert, update, and delete, is essentially doubled.
-
Data items that are updated very frequently (e.g. customer daily balance) may perform poorly when queried because many rows will have to be processed in order to find the one that has not been replaced.
-
The view uses the USER special register, which may not be unique per actual user.
20.6. Multiple Versions of the World
The next design is similar to the previous, but we are also going to allow users to both see and change the world - as it was in the past, and as it is now, without affecting the real-world data. These extra features require a much more complex design:
-
We cannot use a base table and a related history table, as we did above. Instead we have just the latter, and use both views and INSTEAD OF triggers to make the users think that they are really seeing and/or changing the former.
-
We need a version table - to record when the data in each version (i.e. virtual copy of the real world) separates from the real world data.
-
Data integrity features, like referential integrity rules, have to be hand-coded in triggers, rather that written using standard Db2 code.
20.6.1. Version Table
The following table has one row per version created:
CREATE TABLE version
( vrsn INTEGER NOT NULL
, vrsn_bgn_ts TIMESTAMP NOT NULL
, CONSTRAINT version1 CHECK(vrsn >= 0)
, CONSTRAINT version2 CHECK(vrsn < 1000000000)
, PRIMARY KEY(vrsn));
The following rules apply to the above:
-
Each version has a unique number. Up to one billion can be created.
-
Each version must have a begin-timestamp, which records at what point in time it separates from the real world. This value must be ⇐ the current time.
-
Rows cannot be updated or deleted in this table - only inserted. This rule is necessary to ensure that we can always trace all changes - in every version.
-
The real-world is deemed to have a version number of zero, and a begin-timestamp value of high-values.
20.6.2. Profile Table
The following profile table has one row per user (i.e. USER special register) that reads from or changes the data tables. It records what version the user is currently using (note: the version timestamp data is maintained using triggers):
CREATE TABLE profile
( user_id VARCHAR(10) NOT NULL
, vrsn INTEGER NOT NULL
, vrsn_bgn_ts TIMESTAMP NOT NULL
, CONSTRAINT profile1 FOREIGN KEY(vrsn)
REFERENCES version(vrsn)
ON DELETE RESTRICT
, PRIMARY KEY(user_id));
20.6.3. Customer (data) Table
Below is a typical data table. This one holds customer data:
CREATE TABLE customer_his
( cust# INTEGER NOT NULL
, cust_name CHAR(10) NOT NULL
, cust_mgr CHAR(10)
, cur_ts TIMESTAMP NOT NULL
, cur_vrsn INTEGER NOT NULL
, cur_actn CHAR(1) NOT NULL
, cur_user VARCHAR(10) NOT NULL
, prv_cust# INTEGER
, prv_ts TIMESTAMP
, prv_vrsn INTEGER
, CONSTRAINT customer1 FOREIGN KEY(cur_vrsn)
REFERENCES version(vrsn)
ON DELETE RESTRICT
, CONSTRAINT customer2 CHECK(cur_actn IN ('I','U','D'))
, PRIMARY KEY(cust#,cur_vrsn,cur_ts));
CREATE INDEX customer_x2
ON customer_his(prv_cust#
, prv_ts
, prv_vrsn);
Note the following:
-
The first three fields are the only ones that the user will see.
-
The users will never update this table directly. They will make changes to a view of the table, which will then invoke INSTEAD OF triggers.
-
The foreign key check (on version) can be removed - if it is forbidden to ever delete any version. This check stops the removal of versions that have changed data.
-
The constraint on CUR_ACTN is just a double-check - to make sure that the triggers that will maintain this table do not have an error. It can be removed, if desired.
-
The secondary index will make the following view more efficient.
The above table has the following hidden fields:
-
CUR-TS: The current timestamp of the change.
-
CUR-VRSN: The version in which change occurred. Zero implies reality.
-
CUR-ACTN: The type of change (i.e. insert, update, or delete).
-
CUR-USER: The user who made the change (for auditing purposes).
-
PRV-CUST#: The previous customer number. This field enables one follow the sequence of changes for a given customer. The value is null if the action is an insert.
-
PRV-TS: The timestamp of the last time the row was changed (null for inserts).
-
PRV-VRNS: The version of the row being replaced (null for inserts).
20.6.4. Views
The following view displays the current state of the data in the above customer table – based on the version that the user is currently using:
CREATE VIEW customer_vw AS
SELECT *
FROM customer_his hhh
, profile ppp
WHERE ppp.user_id = USER
AND hhh.cur_actn <> 'D'
AND ( (ppp.vrsn = 0 AND hhh.cur_vrsn = 0)
OR (ppp.vrsn > 0 AND hhh.cur_vrsn = 0 AND hhh.cur_ts < ppp.vrsn_bgn_ts)
OR (ppp.vrsn > 0 AND hhh.cur_vrsn = ppp.vrsn)
)
AND NOT EXISTS
(SELECT *
FROM customer_his nnn
WHERE nnn.prv_cust# = hhh.cust#
AND nnn.prv_ts = hhh.cur_ts
AND nnn.prv_vrsn = hhh.cur_vrsn
AND ( (ppp.vrsn = 0 AND nnn.cur_vrsn = 0)
OR (ppp.vrsn > 0 AND nnn.cur_vrsn = 0 AND nnn.cur_ts < ppp.vrsn_bgn_ts)
OR (ppp.vrsn > 0 AND nnn.cur_vrsn = ppp.vrsn)
)
);
The above view shows all customer rows, as long as:
-
The action was not a delete.
-
The version is either zero (i.e. reality), or the user’s current version.
-
If the version is reality, then the current timestamp is < the version begin-timestamp (as duplicated in the profile table).
-
There does not exist any row that "replaces" the current row (and that row has a current timestamp that is ⇐ to the profile (version) timestamp).
To make things easier for the users, we will create another view that sits on top of the above view. This one only shows the business fields:
CREATE VIEW customer AS
SELECT cust#
, cust_name
, cust_mgr
FROM customer_vw;
All inserts, updates, and deletes, are done against the above view, which then propagates down to the first view, whereupon they are trapped by INSTEAD OF triggers. The changes are then applied (via the triggers) to the underlying tables.
20.6.5. Insert Trigger
The following INSTEAD OF trigger traps all inserts to the first view above, and then applies the insert to the underlying table - with suitable modifications:
CREATE TRIGGER customer_ins
INSTEAD OF INSERT ON customer_vw
REFERENCING NEW AS nnn
FOR EACH ROW
MODE Db2SQL
INSERT INTO customer_his
VALUES(nnn.cust#
, nnn.cust_name
, nnn.cust_mgr
, CURRENT TIMESTAMP
, (SELECT vrsn
FROM profile
WHERE user_id = USER)
, CASE
WHEN 0 < (SELECT COUNT(*)
FROM customer
WHERE cust# = nnn.cust#)
THEN RAISE_ERROR('71001','ERROR: Duplicate cust#')
ELSE 'I'
END
, USER
, NULL
, NULL
, NULL);
Observe the following:
-
The basic customer data is passed straight through.
-
The current timestamp is obtained from Db2.
-
The current version is obtained from the user’s profile-table row.
-
A check is done to see if the customer number is unique. One cannot use indexes to enforce such rules in this schema, so one has to code accordingly.
-
The previous fields are all set to null.
20.6.6. Update Trigger
The following INSTEAD OF trigger traps all updates to the first view above, and turns them into an insert to the underlying table - with suitable modifications:
CREATE TRIGGER customer_upd
INSTEAD OF UPDATE ON customer_vw
REFERENCING NEW AS nnn
OLD AS ooo
FOR EACH ROW
MODE Db2SQL
INSERT INTO customer_his
VALUES (nnn.cust#
, nnn.cust_name
, nnn.cust_mgr
, CURRENT TIMESTAMP
, ooo.vrsn
, CASE
WHEN nnn.cust# <> ooo.cust#
THEN RAISE_ERROR('72001','ERROR: Cannot change cust#')
ELSE 'U'
END
, ooo.user_id
, ooo.cust#
, ooo.cur_ts
, ooo.cur_vrsn);
In this particular trigger, updates to the customer number (i.e. business key column) are not allowed. This rule is not necessary, it simply illustrates how one would write such code if one so desired.
20.6.7. Delete Trigger
The following INSTEAD OF trigger traps all deletes to the first view above, and turns them into an insert to the underlying table - with suitable modifications:
CREATE TRIGGER customer_del
INSTEAD OF DELETE ON customer_vw
REFERENCING OLD AS ooo
FOR EACH ROW
MODE Db2SQL
INSERT INTO customer_his
VALUES (ooo.cust#
, ooo.cust_name
, ooo.cust_mgr
, CURRENT TIMESTAMP
, ooo.vrsn
, 'D'
, ooo.user_id
, ooo.cust#
, ooo.cur_ts
, ooo.cur_vrsn);
20.7. Summary
The only thing that the user need see in the above schema in the simplified (second) view that lists the business data columns. They would insert, update, and delete the rows in this view as if they were working on a real table. Under the covers, the relevant INSTEAD OF trigger would convert whatever they did into a suitable insert to the underlying table.
This schema supports the following:
-
To do "what if" analysis, all one need do is insert a new row into the version table – with a begin timestamp that is the current time. This insert creates a virtual copy of every table in the application, which one can then update as desired.
-
To do historical analysis, one simply creates a version with a begin-timestamp that is at some point in the past. Up to one billion versions are currently supported.
-
To switch between versions, all one need do is update one’s row in the profile table.
-
One can use recursive SQL (not shown here) to follow the sequence of changes to any particular item, in any particular version.
This schema has the following limitations:
-
Every data table has to have a unique (business) key.
-
Data items that are updated very frequently (e.g. customer daily balance) may perform poorly when queried because many rows will have to be processed in order to find the one that has not been replaced.
-
The views use the USER special register, which may not be unique per actual user.
-
Data integrity features, like referential integrity rules, cascading deletes, and unique key checks, have to be hand-coded in the INSTEAD OF triggers.
-
Getting the triggers right is quite hard. If the target application has many tables, it might be worthwhile to first create a suitable data-dictionary, and then write a script that generates as much of the code as possible.
21. Using SQL to Make SQL
This chapter describes how to use SQL to make SQL. For example, one might want to make DDL statements to create views on a set of tables.
21.1. Export Command
The following query will generate a set of queries that will count the rows in each of the selected Db2 catalogue views:
SELECT 'SELECT COUNT(*) FROM '
CONCAT RTRIM(tabschema)
CONCAT '.'
CONCAT tabname
CONCAT ';'
FROM syscat.tables
WHERE tabschema = 'SYSCAT'
AND tabname LIKE 'N%'
ORDER BY tabschema
, tabname;
ANSWER
SELECT COUNT(*) FROM SYSCAT.NAMEMAPPINGS; SELECT COUNT(*) FROM SYSCAT.NODEGROUPDEF; SELECT COUNT(*) FROM SYSCAT.NODEGROUPS;
If we wrap the above inside an EXPORT statement, and define no character delimiter, we will be able to create a file with the above answer - and nothing else. This could in turn be run as if were some SQL statement that we had written:
EXPORT TO C:\FRED.TXT OF DEL
MODIFIED BY NOCHARDEL
SELECT 'SELECT COUNT(*) FROM '
CONCAT RTRIM(tabschema)
CONCAT '.'
CONCAT tabname
CONCAT ';'
FROM syscat.tables
WHERE tabschema = 'SYSCAT'
AND tabname LIKE 'N%'
ORDER BY tabschema
, tabname;
21.2. Export Command Notes
The key EXPORT options used above are:
-
The file name is "C\FRED.TXT".
-
The data is sent to a delimited (i.e. DEL) file.
-
The delimited output file uses no character delimiter (i.e. NOCHARDEL).
The remainder of this chapter will assume that we are using the EXPORT command, and will describe various ways to generate more elaborate SQL statements.
21.3. SQL to Make SQL
The next query is the same as the prior two, except that we have added the table name to each row of output:
SELECT 'SELECT '''
CONCAT tabname
CONCAT ''', COUNT(*) FROM '
CONCAT RTRIM(tabschema)
CONCAT '.'
CONCAT tabname
CONCAT ';'
FROM syscat.tables
WHERE tabschema = 'SYSCAT'
AND tabname LIKE 'N%'
ORDER BY tabschema
, tabname;
ANSWER
SELECT 'NAMEMAPPINGS', COUNT(*) FROM SYSCAT.NAMEMAPPINGS; SELECT 'NODEGROUPDEF', COUNT(*) FROM SYSCAT.NODEGROUPDEF; SELECT 'NODEGROUPS', COUNT(*) FROM SYSCAT.NODEGROUPS;
We can make more readable output by joining the result set to four numbered rows, and then breaking the generated query down into four lines:
WITH temp1 (num) AS
(VALUES (1), (2), (3), (4))
SELECT CASE num
WHEN 1 THEN 'SELECT ''' || tabname || ''' AS tname'
WHEN 2 THEN ' , COUNT(*)' || ' AS #rows'
WHEN 3 THEN 'FROM ' || RTRIM(tabschema) || '.' || tabname || ';'
WHEN 4 THEN ''
END
FROM syscat.tables
, temp1
WHERE tabschema = 'SYSCAT',
AND tabname LIKE 'N%'
ORDER BY tabschema
, tabname
, num;
ANSWER
SELECT 'NAMEMAPPINGS' AS tname
, COUNT(*) AS #rows
FROM SYSCAT.NAMEMAPPINGS;
SELECT 'NODEGROUPDEF' AS tname
, COUNT(*) AS #rows
FROM SYSCAT.NODEGROUPDEF;
SELECT 'NODEGROUPS' AS tname
, COUNT(*) AS #rows
FROM SYSCAT.NODEGROUPS;
So far we have generated separate SQL statements for each table that matches. But imagine that instead we wanted to create a single statement that processed all tables. For example, we might want to know the sum of the rows in all of the matching tables. There are two ways to do this, but neither of them are very good:
-
We can generate a single large query that touches all of the matching tables. A query can be up to 2MB long, so we could reliably use this technique as long as we had less than about 5,000 tables to process.
-
We can declare a global temporary table, then generate insert statements (one per matching table) that insert a count of the rows in the table. After running the inserts, we can sum the counts in the temporary table.
The next example generates a single query that counts all of the rows in the matching tables:
WITH temp1 (num) AS
(VALUES (1), (2), (3), (4))
SELECT CASE num
WHEN 1 THEN 'SELECT SUM(C1)'
when 2 then 'FROM ('
WHEN 3 THEN 'SELECT COUNT(*) AS C1 FROM '
CONCAT RTRIM(tabschema)
CONCAT '.'
CONCAT tabname
CONCAT CASE dd
WHEN 1 THEN ''
ELSE ' UNION ALL'
END
WHEN 4 THEN ') AS xxx;'
END
FROM (SELECT tab.*
, ROW_NUMBER() OVER(ORDER BY tabschema ASC
, tabname ASC) AS aa
, ROW_NUMBER() OVER(ORDER BY tabschema DESC
, tabname DESC) AS dd
FROM syscat.tables tab
WHERE tabschema = 'SYSCAT'
AND tabname LIKE 'N%') AS xxx
, emp1
WHERE (num <= 2 AND aa = 1)
OR (num = 3)
OR (num = 4 AND dd = 1)
ORDER BY tabschema ASC
, tabname ASC
, num ASC;
ANSWER
SELECT SUM(C1)
FROM (SELECT COUNT(*) AS C1 FROM SYSCAT.NAMEMAPPINGS UNION ALL
SELECT COUNT(*) AS C1 FROM SYSCAT.NODEGROUPDEF UNION ALL
SELECT COUNT(*) AS C1 FROM SYSCAT.NODEGROUPS
) AS xxx;
The above query works as follows:
-
A temporary table (i.e. temp1) is generated with one column and four rows.
-
A nested table expression (i.e. xxx) is created with the set of matching rows (tables).
-
Within the nested table expression the ROW_NUMBER function is used to define two new columns. The first will have the value 1 for the first matching row, and the second will have the value 1 for the last matching row.
-
The xxx and temp1 tables are joined. Two new rows (i.e. num ⇐ 2) are added to the front, and one new row (i.e. num = 4) is added to the back.
-
The first two new rows (i.e. num = 1 and 2) are used to make the first part of the generated query.
-
The last new row (i.e. num = 4) is used to make the tail end of the generated query.
-
All other rows (i.e. num = 3) are used to create the core of the generated query.
In the above query no SQL is generated if no rows (tables) match. Alternatively, we might want to generate a query that returns zero if no rows match.
22. Running SQL Within SQL
This chapter describes how to generate and run SQL statements within SQL statements.
22.1. Introduction
Consider the following query:
SELECT empno
, lastname
, workdept
, salary
FROM employee
WHERE empno = '000250';
The above query exhibits all the usual strengths and weaknesses of the SQL language. It is easy to understand, simple to write, and assuming suitable indexes, efficient to run. But the query is annoyingly rigid in the sense that the both the internal query logic (i.e. which rows to fetch from what tables), and the set of columns to be returned, are fixed in the query syntax. Reasonably intelligent programs accessing suitably well-structured data might want to run queries like the following:
SELECT all-columns
FROM all-relevant-tables
WHERE all-predicates-are-true
It would of course be possible to compose the required query in the program and then run it. But there are some situations where it would be nice if we could also generate and then run the above pseudo-query inside the SQL language itself. This can be done, if there are two simple enhancements to the language:
-
The ability to generate and run SQL within SQL.
-
A way to make the query output-column-independent.
22.2. Generate SQL within SQL
To test the first concept above I wrote some very simple user-defined scalar functions (see Db2 SQL Functions) that enable one to generate and run SQL within SQL. In these functions the first row/column value fetched is returned. To illustrate, consider the following pseudoquery:
SELECT COUNT(*)
FROM all-relevant-tables
WHERE empno = '000250';
In the above pseudo-query we want to count all matching rows in all matching tables where the EMPNO is a given value. If we use the Db2 catalogue tables as a source dictionary, and we call a user-defined scalar function that can run SQL within SQL (see Db2 SQL Functions for the function definition), we can write the following query:
SELECT CHAR(tabname,15) AS tabname
, get_INTEGER(
' SELECT COUNT(*)' ||
' FROM ' || tabschema || '.' || tabname ||
' WHERE ' || colname || ' = ''000250'''
) AS num_rows
FROM syscat.columns
WHERE tabschema = USER
AND colname = 'EMPNO'
AND typename = 'CHARACTER'
ORDER BY tabname;
ANSWER
TABNAME | NUM_ROWS |
---|---|
EMP_PHOTO |
0 |
VEMP |
1 |
VEMPDPT1 |
1 |
VEMPPROJACT |
9 |
VSTAFAC2 |
9 |
22.3. Make Query Column-Independent
The second issue to address was how to make the SQL language output-column-independent. This capability is needed in order to support the following type of pseudo-query:
SELECT all-columns
FROM all-relevant-tables
WHERE empno = '000250';
The above query cannot be written in SQL because the set of columns to be returned can not be determined until the set of matching tables are identified. To get around this constraint, I wrote a very simple Db2 table function in Java (see Transpose Function) that accepts any valid query as input, runs it, and then returns all of the rows and columns fetched. But before returning anything, the function transposes each row/column instance into a single row – with a set of fixed columns returned that describe each row/column data instance. The function is used below to run the above pseudo-query:
WITH temp1 AS
(SELECT tabname
, VARCHAR(
' SELECT *' ||
' FROM ' || tabschema || '.' || tabname ||
' WHERE ' || colname || ' = ''000250'''
) AS SQL_text
FROM syscat.columns
WHERE tabschema = USER
AND colname = 'EMPNO'
AND typename = 'CHARACTER'
)
SELECT CHAR(t1.tabname,10) AS tabname
, t2.row_number AS row#
, t2.col_num AS col#
, CHAR(t2.col_name,15) AS colname
, CHAR(t2.col_type,15) AS coltype
, CHAR(t2.col_value,20) AS colvalue
FROM temp1 t1
, TABLE(tab_transpose(sql_text)) AS t2
ORDER BY t1.tabname
, t2.row_number
, t2.col_num;
Below are the first three "rows" of the answer:
TABNAME | ROW# | COL# | COLNAME | COLTYPE | COLVALUE |
---|---|---|---|---|---|
EMPLOYEE |
1 |
1 |
EMPNO |
CHAR |
000250 |
EMPLOYEE |
1 |
2 |
FIRSTNME |
VARCHAR |
DANIEL |
EMPLOYEE |
1 |
3 |
MIDINIT |
CHAR |
S |
EMPLOYEE |
1 |
4 |
LASTNAME |
VARCHAR |
SMITH |
EMPLOYEE |
1 |
5 |
WORKDEPT |
CHAR |
D21 |
EMPLOYEE |
1 |
6 |
PHONENO |
CHAR |
0961 |
EMPLOYEE |
1 |
7 |
HIREDATE |
DATE |
1999-10-30 |
EMPLOYEE |
1 |
8 |
JOB |
CHAR |
CLERK |
EMPLOYEE |
1 |
9 |
EDLEVEL |
SMALLINT |
15 |
EMPLOYEE |
1 |
10 |
SEX |
CHAR |
M |
EMPLOYEE |
1 |
11 |
BIRTHDATE |
DATE |
1969-11-12 |
EMPLOYEE |
1 |
12 |
SALARY |
DECIMAL |
49180.00 |
EMPLOYEE |
1 |
13 |
BONUS |
DECIMAL |
400.00 |
EMPLOYEE |
1 |
14 |
COMM |
DECIMAL |
1534.00 |
EMPPROJACT |
1 |
1 |
EMPNO |
CHAR |
000250 |
EMPPROJACT |
1 |
2 |
PROJNO |
CHAR |
AD3112 |
EMPPROJACT |
1 |
3 |
ACTNO |
SMALLINT |
60 |
EMPPROJACT |
1 |
4 |
EMPTIME |
DECIMAL |
1.00 |
EMPPROJACT |
1 |
5 |
EMSTDATE |
DATE |
2002-01-01 |
EMPPROJACT |
1 |
6 |
EMENDATE |
DATE |
2002-02-01 |
EMPPROJACT |
2 |
1 |
EMPNO |
CHAR |
000250 |
EMPPROJACT |
2 |
2 |
PROJNO |
CHAR |
AD3112 |
EMPPROJACT |
2 |
3 |
ACTNO |
SMALLINT |
60 |
EMPPROJACT |
2 |
4 |
EMPTIME |
DECIMAL |
0.50 |
EMPPROJACT |
2 |
5 |
EMSTDATE |
DATE |
2002-02-01 |
EMPPROJACT |
2 |
6 |
EMENDATE |
DATE |
2002-03-15 |
22.4. Business Uses
At this point, I’ve got an interesting technical solution looking for a valid business problem. Some possible uses follow:
22.4.1. Frictionless Query
Imagine a relational database application where the table definitions are constantly changing. The programs using the data are able adapt accordingly, in which case the intermediate SQL queries have to also be equally adaptable. The application could maintain a data dictionary that was updated in sync with the table changes. Each query would reference the dictionary at the start of its processing, and then build the main body of the query (i.e. that which obtains the desired application data) as needed. I did some simple experiments using this concept. It worked, but I could see no overwhelming reason why one would use it, as opposed to building the query external to Db2, and then running it.
22.4.2. Adaptive Query
One could write a query where the internal query logic changed – depending on what data was encountered along the way. I tested this concept, and found that it works, but one still needs to define the general processing logic of the query somewhere. It was often easier to code a series of optional joins (in the query) to get the same result.
22.4.3. Meta-Data to Real-Data Join
A meta-data to real-data join can only be done using the SQL enhancements described above.
Some examples of such a join include:
-
List all tables containing a row where EMPID = '123'.
-
List all rows (in any table) that duplicate a given row.
-
Confirm that two "sets of tables" have identical data.
-
Scan all plan-tables looking for specific access paths.
-
Find the largest application table that has no index.
These types of query are relatively rare, but they certainly do exist, and they are legitimate business queries.
22.4.4. Meta Data Dictionaries
In the above examples the Db2 catalogue was used as the source of meta-data that describes the relationships between the tables accessed by the query. This works up to a point, but the Db2 catalogue is not really designed for this task. Thus it would probably be better to use a purpose-built meta-data dictionary. Whenever application tables were changed, the meta-data dictionary would be updated accordingly - or might in fact be the source of the change. SQL queries generated using the meta-data dictionary would automatically adjust as the table changes were implemented.
22.5. Db2 SQL Functions
This section describes how to join meta-data to real data in a single query. In other words, a query will begin by selecting a list of tables from the Db2 catalogue. It will then access each table in the list. Such a query cannot be written using ordinary SQL, because the set of tables to be accessed is not known to the statement. But it can be written if the query references a very simple user-defined scalar function and related stored procedure. To illustrate, the following query will select a list of tables, and for each matching table get a count of the rows in the same:
SELECT CHAR(tabschema,8) AS schema
, CHAR(tabname,20) AS tabname
, return_INTEGER(
'SELECT COUNT(*) ' ||
'FROM ' || tabschema || '.' || tabname
) AS #rows
FROM syscat.tables
WHERE tabschema = 'SYSCAT'
AND tabname LIKE 'RO%'
ORDER BY tabschema
, tabname
FOR FETCH ONLY
WITH UR;
ANSWER
SCHEMA | TABNAME | #ROWS |
---|---|---|
SYSCAT |
ROUTINEAUTH |
168 |
SYSCAT |
ROUTINEDEP |
41 |
SYSCAT |
ROUTINEPARMS |
2035 |
SYSCAT |
ROUTINES |
314 |
22.6. Function and Stored Procedure Used
The above query calls a user-defined scalar function called return_INTEGER that accepts as input any valid single-column query and returns (you guessed it) an integer value that is the first row fetched by the query. The function is actually nothing more than a stub:
CREATE FUNCTION return_INTEGER (in_stmt VARCHAR(4000))
RETURNS INTEGER
LANGUAGE SQL
READS SQL DATA
NO EXTERNAL ACTION
BEGIN ATOMIC
DECLARE out_val INTEGER;
CALL return_INTEGER(in_stmt, out_val);
RETURN out_val;
END
The real work is done by a stored procedure that is called by the function:
CREATE PROCEDURE return_INTEGER (IN in_stmt VARCHAR(4000)
, OUT out_val INTEGER)
LANGUAGE SQL
READS SQL DATA
NO EXTERNAL ACTION
BEGIN
DECLARE c1 CURSOR FOR s1;
PREPARE s1 FROM in_stmt;
OPEN c1;
FETCH c1 INTO out_val;
CLOSE c1;
RETURN;
END
The combined function and stored-procedure logic goes as follow:
-
Main query calls function - sends query text.
-
Function calls stored-procedure - sends query text.
-
Stored-procedure prepares, opens, fetches first row, and then closes query.
-
Stored procedure returns result of first fetch back to the function
-
Function returns the result back to the main query.
22.7. Different Data Types
One needs to have a function and related stored-procedure for each column type that can be returned. Below is a DECIMAL example:
CREATE PROCEDURE return_DECIMAL (IN in_stmt VARCHAR(4000)
, OUT out_val DECIMAL(31, 6))
LANGUAGE SQL
READS SQL DATA
NO EXTERNAL ACTION
BEGIN
DECLARE c1 CURSOR FOR s1;
PREPARE s1 FROM in_stmt;
OPEN c1;
FETCH c1 INTO out_val;
CLOSE c1;
RETURN;
END
CREATE FUNCTION return_DECIMAL (in_stmt VARCHAR(4000))
RETURNS DECIMAL(31, 6)
LANGUAGE SQL
READS SQL DATA
NO EXTERNAL ACTION
BEGIN ATOMIC
DECLARE out_val DECIMAL(31,6);
CALL return_DECIMAL(in_stmt,out_val);
RETURN out_val;
END
22.8. Usage Examples
The query below lists those tables that have never had RUNSTATS run (i.e. the stats-time is null), and that currently have more than 1,000 rows:
SELECT CHAR(tabschema,8) AS schema
, CHAR(tabname,20) AS tabname
, #rows
FROM (SELECT tabschema
, tabname
, return_INTEGER(
' SELECT COUNT(*)' ||
' FROM ' || tabschema || '.' || tabname ||
' FOR FETCH ONLY WITH UR'
) AS #rows
FROM syscat.tables tab
WHERE tabschema LIKE 'SYS%'
AND type = 'T'
AND stats_time IS NULL) AS xxx
WHERE #rows > 1000
ORDER BY #rows DESC
FOR FETCH ONLY
WITH UR;
ANSWER
SCHEMA | TABNAME | #ROWS |
---|---|---|
SYSIBM |
SYSCOLUMNS |
3518 |
SYSIBM |
SYSROUTINEPARMS |
2035 |
22.9. Efficient Queries
The query shown above would typically process lots of rows, but this need not be the case. The next example lists all tables with a department column and at least one row for the 'A00' department. Only a single matching row is fetched from each table, so as long as there is a suitable index on the department column, the query should fly:
SELECT CHAR(tab.tabname,15) AS tabname
, CHAR(col.colname,10) AS colname
, CHAR(COALESCE(return_VARCHAR(
' SELECT ''Y''' ||
' FROM ' || tab.tabschema || '.' || tab.tabname ||
' WHERE ' || col.colname || ' = ''A00''' ||
' FETCH FIRST 1 ROWS ONLY ' ||
' OPTIMIZE FOR 1 ROW ' ||
' WITH UR'
), 'N'), 1) AS has_dept
FROM syscat.columns col
, syscat.tables tab
WHERE col.tabschema = USER
AND col.colname IN ('DEPTNO', 'WORKDEPT')
AND col.tabschema = tab.tabschema
AND col.tabname = tab.tabname
AND tab.type = 'T'
FOR FETCH ONLY
WITH UR;
ANSWER
TABNAME | COLNAME | HAS_DEPT |
---|---|---|
DEPARTMENT |
DEPTNO |
Y |
EMPLOYEE |
WORKDEPT |
Y |
PROJECT |
DEPTNO |
N |
The next query is the same as the previous, except that it only searches those matching tables that have a suitable index on the department field:
SELECT CHAR(tab.tabname,15) AS tabname
, CHAR(col.colname,10) AS colname
, CHAR(COALESCE(return_VARCHAR(
' SELECT ''Y''' ||
' FROM ' || tab.tabschema || '.' || tab.tabname ||
' WHERE ' || col.colname || ' = ''A00''' ||
' FETCH FIRST 1 ROWS ONLY ' ||
' OPTIMIZE FOR 1 ROW ' ||
' WITH UR'
), 'N'), 1) AS has_dept
FROM syscat.columns col
, syscat.tables tab
WHERE col.tabschema = USER
AND col.colname IN ('DEPTNO', 'WORKDEPT')
AND col.tabschema = tab.tabschema
AND col.tabname = tab.tabname
AND tab.type = 'T'
AND col.colname IN (SELECT SUBSTR(idx.colnames, 2, LENGTH(col.colname))
FROM syscat.indexes idx
WHERE tab.tabschema = idx.tabschema
AND tab.tabname = idx.tabname)
FOR FETCH ONLY
WITH UR;
ANSWER
TABNAME | COLNAME | HAS_DEPT |
---|---|---|
DEPARTMENT |
DEPTNO |
Y |
Using logic very similar to the above, one can efficiently ask questions like: "list all tables in the application that have references to customer-number 1234 in indexed fields". Even if the query has to process hundreds of tables, each with billions of rows, it should return an answer in less than ten seconds. In the above examples we knew what columns we wanted to process, but not the tables. But for some questions we don’t even need to know the column name. For example, we could scan all indexed DATE columns in an application - looking for date values that are more than five years old. Once again, such a query should run in seconds.
22.10. Java Functions
We can do the same as the above by calling a user-defined-function that invokes a java program, but we can also do much more. This section will cover the basics.
22.10.1. Scalar Functions
The following code creates a user-defined scalar function that sends a query to a java program, and gets back the first row/column fetched when the query is run:
CREATE FUNCTION get_Integer(VARCHAR(4000))
RETURNS INTEGER
LANGUAGE JAVA
EXTERNAL NAME 'Graeme2!get_Integer'
PARAMETER STYLE Db2GENERAL
NO EXTERNAL ACTION
NOT DETERMINISTIC
READS SQL DATA
FENCED;
Below is the corresponding java code:
import java.lang.*;
import COM.ibm.db2.app.*;
import java.sql.*;
import java.math.*;
import java.io.*;
public class Graeme2 extends UDF {
public void get_Integer(String inStmt
, int outValue) throws Exception {
try {
Connection con = DriverManager.getConnection ("jdbc:default:connection");
PreparedStatement stmt = con.prepareStatement(inStmt);
ResultSet rs = stmt.executeQuery();
if (rs.next() == true && rs.getString(1) != null) {
set(2, rs.getInt(1));
}
rs.close();
stmt.close();
con.close();
}
catch (SQLException sqle) {
setSQLstate("38999");
setSQLmessage("SQLCODE = " + sqle.getSQLState());
return;
}
}
}
Java Logic
-
Establish connection.
-
Prepare the SQL statement (i.e. input string).
-
Execute the SQL statement (i.e. open cursor).
-
If a row is found, and the value (of the first column) is not null, return value.
-
Close cursor.
-
Return.
22.10.2. Usage Example
SELECT workdept AS dept
, empno
, salary
, get_Integer(
' SELECT count(*)' ||
' FROM employee' ||
' where workdept = ''' || workdept || ''' ') AS #rows
FROM employee
WHERE salary < 35500
ORDER BY workdept
, empno;
ANSWER
DEPT | EMPNO | SALARY | #ROWS |
---|---|---|---|
E11 |
000290 |
35340.00 |
7 |
E21 |
200330 |
35370.00 |
6 |
E21 |
200340 |
31840.00 |
6 |
22.10.3. Tabular Functions
So far, all we have done in this chapter is get single values from tables. Now we will retrieve sets of rows from tables. To do this we need to define a tabular function:
CREATE FUNCTION tab_Varchar (VARCHAR(4000))
RETURNS TABLE (row_number INTEGER
, row_value VARCHAR(254))
LANGUAGE JAVA
EXTERNAL NAME 'Graeme2!tab_Varchar'
PARAMETER STYLE DB2GENERAL
NO EXTERNAL ACTION
NOT DETERMINISTIC
DISALLOW PARALLEL
READS SQL DATA
FINAL CALL
FENCED;
Below is the corresponding java code. Observe that two columns are returned – a row-number and the value fetched:
import java.lang.*;
import COM.ibm.db2.app.*;
import java.sql.*;
import java.math.*;
import java.io.*;
public class Graeme2 extends UDF {
Connection con;
Statement stmt;
ResultSet rs;
int rowNum;
public void tab_Varchar(String inStmt, int outNumber, String outValue) throws Exception {
switch (getCallType())
{
case SQLUDF_TF_FIRST:
break;
case SQLUDF_TF_OPEN:
rowNum = 1;
try {
con = DriverManager.getConnection("jdbc:default:connection");
stmt = con.createStatement();
rs = stmt.executeQuery(inStmt);
}
catch(SQLException sqle) {
setSQLstate("38999");
setSQLmessage("SQLCODE = " + sqle.getSQLState());
return;
}
break;
case SQLUDF_TF_FETCH:
if (rs.next() == true) {
set(2, rowNum);
if (rs.getString(1) != null) {
set(3, rs.getString(1));
}
rowNum++;
}
else {
setSQLstate ("02000");
}
break;
case SQLUDF_TF_CLOSE:
rs.close();
stmt.close();
con.close();
break;
case SQLUDF_TF_FINAL:
break;
}
}
}
22.10.4. Java Logic
Java programs that send data to Db2 table functions use a particular type of CASE logic to return the output data. In particular, a row is returned at the end of every FETCH process.
OPEN:
-
Establish connection.
-
Prepare the SQL statement (i.e. input string).
-
Execute the SQL statement (i.e. open cursor).
-
Set row-number variable to one.
FETCH:
-
If row exists, set row-number output value.
-
If value fetched is not null, set output value.
-
Increment row-number variable.
CLOSE:
-
Close cursor.
-
Return.
22.10.5. Usage Example
The following query lists all EMPNO values that exist in more than four tables:
WITH make_queries AS
(SELECT tab.tabschema
, tab.tabname
, ' SELECT EMPNO ' ||
' FROM ' || tab.tabschema || '.' || tab.tabname
AS sql_text
FROM syscat.tables tab
, syscat.columns col
WHERE tab.tabschema = USER
AND tab.type = 'T'
AND col.tabschema = tab.tabschema
AND col.tabname = tab.tabname
AND col.colname = 'EMPNO'
AND col.typename = 'CHARACTER'
AND col.length = 6)
, run_queries AS
(SELECT qqq.*
, ttt.*
FROM make_queries qqq,
TABLE(tab_Varchar(sql_text)) AS ttt)
SELECT CHAR(row_value,10) AS empno
, COUNT(*) AS #rows
, COUNT(DISTINCT tabschema || tabname) AS #tabs
, CHAR(MIN(tabname), 18) AS min_tab
, CHAR(MAX(tabname), 18) AS max_tab
FROM run_queries
GROUP BY row_value
HAVING COUNT(DISTINCT tabschema || tabname) > 3
ORDER BY row_value
FOR FETCH ONLY WITH UR;
ANSWER
EMPNO | #ROWS | #TABS | MIN_TAB | MAX_TAB |
---|---|---|---|---|
000130 |
7 |
4 |
EMP_PHOTO |
EMPPROJACT |
000140 |
10 |
4 |
EMP_PHOTO |
EMPPROJACT |
000150 |
7 |
4 |
EMP_PHOTO |
EMPPROJACT |
000190 |
7 |
4 |
EMP_PHOTO |
EMPPROJACT |
22.10.6. Transpose Function
Below is some pseudo-code for a really cool query:
SELECT all columns
FROM unknown tables
WHERE any unknown columns = '%ABC%'
In the above query we want to retrieve an unknown number of unknown types of columns (i.e. all columns in each matching row) from an unknown set of tables where any unknown column in the row equals 'ABC'. Needless to say, the various (unknown) tables will have differing types and numbers of columns. The above query is remarkably easy to write in SQL (see Transpose Function) and reasonably efficient to run, if we invoke a cute little java program that transposes columns into rows. The act of transposition means that each row/column instance retrieved becomes a separate row. So the following result:
SELECT * FROM empprojact WHERE empno = '000150';
ANSWER
EMPNO | PROJNO | ACTNO | EMPTIME | EMSTDATE | EMENDATE |
---|---|---|---|---|---|
000150 |
MA2112 |
60 |
1.00 |
01/01/2002 |
07/15/2002 |
000150 |
MA2112 |
180 |
1.00 |
07/15/2002 |
02/01/2003 |
Becomes this result:
SELECT SMALLINT(row_number) AS row#
, col_num AS col#
, CHAR(col_name,13) AS col_name
, CHAR(col_type,10) AS col_type
, col_length AS col_len
, SMALLINT(LENGTH(col_value)) AS val_len
, SUBSTR(col_value,1,20) AS col_value
FROM TABLE(tab_Transpose(
' SELECT*' ||
' FROM empprojact' ||
' WHERE empno = ''000150''')) AS ttt
ORDER BY 1,2;
ANSWER
ROW# | COL# | COL_NAME | COL_TYPE | COL_LEN | VAL_LEN | COL_VALUE |
---|---|---|---|---|---|---|
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
1 |
2 |
PROJNO |
CHAR |
6 |
6 |
MA2112 |
1 |
3 |
ACTNO |
SMALLINT |
6 |
2 |
60 |
1 |
4 |
EMPTIME |
DECIMAL |
7 |
4 |
1.00 |
1 |
5 |
EMSTDATE |
DATE |
10 |
10 |
2002-01-01 |
1 |
6 |
EMENDATE |
DATE |
10 |
10 |
2002-07-15 |
2 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
2 |
2 |
PROJNO |
CHAR |
6 |
6 |
MA2112 |
2 |
3 |
ACTNO |
SMALLINT |
6 |
3 |
180 |
2 |
4 |
EMPTIME |
DECIMAL |
7 |
4 |
1.00 |
2 |
5 |
EMSTDATE |
DATE |
10 |
10 |
2002-07-15 |
2 |
6 |
EMENDATE |
DATE |
10 |
10 |
2003-02-01 |
The user-defined transpose function invoked above accepts a query as input. It executes the query then returns the query result as one row per row/column instance found. The function output table has the following columns:
-
ROW_NUMBER: The number of the row fetched.
-
NUM_COLS: The number of columns fetched per row.
-
COL_NUM: The column-number for the current row. This value, in combination with the prior row-number value, identifies a unique output row.
-
COL_NAME: The name of the data column - as given in the query. If there is no name, the value is the column number.
-
COL_TYPE: The Db2 column-type for the value.
-
COL_LENGTH: The Db2 column-length (note: not data item length) for the value.
-
COL_VALUE: The row/column instance value itself. If the data column is too long, or of an unsupported type (e.g. CLOB, DBCLOB, or XML), null is returned.
The transpose function always returns the same set of columns, regardless of which table is being accessed. So we can use it to write a query where we don’t know which tables we want to select from. In the next example, we select all columns from all rows in all tables where the EMPNO column has a certain value:
WITH make_queries AS
(SELECT tab.tabschema
, tab.tabname,
' SELECT *' ||
' FROM ' || tab.tabname ||
' WHERE empno = ''000150'''
AS sql_text
FROM syscat.tables tab
, syscat.columns col
WHERE tab.tabschema = USER
AND tab.type = 'T'
AND col.tabschema = tab.tabschema
AND col.tabname = tab.tabname
AND col.colname = 'EMPNO'
AND col.typename = 'CHARACTER'
AND col.length = 6)
, run_queries AS
(SELECT qqq.*
, ttt.*
FROM make_queries qqq
, TABLE(tab_Transpose(sql_text)) AS ttt)
SELECT SUBSTR(tabname,1,11) AS tab_name
, SMALLINT(row_number) AS row#
, col_num AS col#
, CHAR(col_name,13) AS col_name
, CHAR(col_type,10) AS col_type
, col_length AS col_len
, SMALLINT(LENGTH(col_value)) AS val_len
, SUBSTR(col_value,1,20) AS col_value
FROM run_queries
ORDER BY 1,2,3;
When we run the above, we get the following answer:
TAB_NAME | ROW# | COL# | COL_NAME | COL_TYPE | COL_LEN | VAL_LEN | COL_VALUE |
---|---|---|---|---|---|---|---|
EMP_PHOTO |
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMP_PHOTO |
1 |
2 |
PHOTO_FORMAT |
VARCHAR |
10 |
6 |
bitmap |
EMP_PHOTO |
1 |
3 |
PICTURE |
BLOB |
204800 |
- |
- |
EMP_PHOTO |
1 |
4 |
EMP_ROWID |
CHAR |
40 |
40 |
|
EMP_PHOTO |
2 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMP_PHOTO |
2 |
2 |
PHOTO_FORMAT |
VARCHAR |
10 |
3 |
gif |
EMP_PHOTO |
2 |
3 |
PICTURE |
BLOB |
204800 |
- |
- |
EMP_PHOTO |
2 |
4 |
EMP_ROWID |
CHAR |
40 |
40 |
|
EMP_RESUME |
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMP_RESUME |
1 |
2 |
RESUME_FORMAT |
VARCHAR |
10 |
5 |
ascii |
EMP_RESUME |
1 |
3 |
RESUME |
CLOB |
5120 |
- |
- |
EMP_RESUME |
1 |
4 |
EMP_ROWID |
CHAR |
40 |
40 |
|
EMP_RESUME |
2 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMP_RESUME |
2 |
2 |
RESUME_FORMAT |
VARCHAR |
10 |
4 |
html |
EMP_RESUME |
2 |
3 |
RESUME |
CLOB |
5120 |
- |
|
- |
EMP_RESUME |
2 |
4 |
EMP_ROWID |
CHAR |
40 |
40 |
EMPLOYEE |
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
|
000150 |
EMPLOYEE |
1 |
2 |
FIRSTNME |
VARCHAR |
12 |
5 |
BRUCE |
EMPLOYEE |
1 |
3 |
MIDINIT |
CHAR |
1 |
1 |
EMPLOYEE |
1 |
4 |
LASTNAME |
VARCHAR |
15 |
7 |
|
ADAMSON |
EMPLOYEE |
1 |
5 |
WORKDEPT |
CHAR |
3 |
3 |
D11 |
EMPLOYEE |
1 |
6 |
PHONENO |
CHAR |
4 |
4 |
4510 |
EMPLOYEE |
1 |
7 |
HIREDATE |
DATE |
10 |
10 |
2002-02-12 |
EMPLOYEE |
1 |
8 |
JOB |
CHAR |
8 |
8 |
DESIGNER |
EMPLOYEE |
1 |
9 |
EDLEVEL |
SMALLINT |
6 |
2 |
16 |
EMPLOYEE |
1 |
10 |
SEX |
CHAR |
1 |
1 |
M |
EMPLOYEE |
1 |
11 |
BIRTHDATE |
DATE |
10 |
10 |
1977-05-17 |
EMPLOYEE |
1 |
12 |
SALARY |
DECIMAL |
11 |
8 |
55280.00 |
EMPLOYEE |
1 |
13 |
BONUS |
DECIMAL |
11 |
6 |
500.00 |
EMPLOYEE |
1 |
14 |
COMM |
DECIMAL |
11 |
7 |
2022.00 |
EMPPROJACT |
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMPPROJACT |
1 |
2 |
PROJNO |
CHAR |
6 |
6 |
MA2112 |
EMPPROJACT |
1 |
3 |
ACTNO |
SMALLINT |
6 |
2 |
60 |
EMPPROJACT |
1 |
4 |
EMPTIME |
DECIMAL |
7 |
4 |
1.00 |
EMPPROJACT |
1 |
5 |
EMSTDATE |
DATE |
10 |
10 |
2002-01-01 |
EMPPROJACT |
1 |
6 |
EMENDATE |
DATE |
10 |
10 |
2002-07-15 |
EMPPROJACT |
2 |
1 |
EMPNO |
CHAR |
6 |
|
6 |
000150 |
EMPPROJACT |
2 |
2 |
PROJNO |
CHAR |
6 |
6 |
MA2112 |
EMPPROJACT |
2 |
3 |
ACTNO |
SMALLINT |
6 |
3 |
180 |
EMPPROJACT |
2 |
4 |
EMPTIME |
DECIMAL |
7 |
4 |
1.00 |
EMPPROJACT |
2 |
5 |
EMSTDATE |
DATE |
10 |
10 |
2002-07-15 |
EMPPROJACT |
2 |
6 |
EMENDATE |
DATE |
10 |
We are obviously on a roll, so now we will write the pseudo-query that we began this chapter with (see Transpose Function). We will fetch every row/column instance in all matching tables where any qualifying column in the row is a particular value.
22.11. Query Logic
-
Define the search parameters.
-
Get the list of matching tables and columns to search.
-
Recursively work through the list of columns to search (for each table), building a search query with multiple EQUAL predicates – one per searchable column.
-
Run the generated queries (i.e. the final line of generated query for each table).
-
Select the output.
Now for the query:
WITH search_values (search_type,search_length,search_value) AS
(VALUES ('CHARACTER',6,'000150'))
, list_columns AS
(SELECT val.search_value
, tab.tabschema
, tab.tabname
, col.colname
, ROW_NUMBER() OVER(PARTITION BY val.search_value
, tab.tabschema
, tab.tabname
ORDER BY col.colname ASC) AS col_a
, ROW_NUMBER() OVER(PARTITION BY val.search_value
, tab.tabschema
, tab.tabname
ORDER BY col.colname DESC) AS col_d
FROM search_values val
, syscat.tables tab
, syscat.columns col
WHERE tab.tabschema = USER
AND tab.type = 'T'
AND tab.tabschema = col.tabschema
AND tab.tabname = col.tabname
AND col.typename = val.search_type
AND col.length = val.search_length)
, make_queries (search_value, tabschema, tabname, colname, col_a, col_d, sql_text) AS
(SELECT tb1.*
, VARCHAR(' SELECT *' ||
' FROM ' || tabname ||
' WHERE ' || colname || ' = ''' || search_value || ''''
, 4000)
FROM list_columns tb1
WHERE col_a = 1
UNION ALL
SELECT tb2.*
, mqy.sql_text || ' OR ' || tb2.colname || ' = ''' || tb2.search_value || ''''
FROM list_columns tb2
, make_queries mqy
WHERE tb2.search_value = mqy.search_value
AND tb2.tabschema = mqy.tabschema
AND tb2.tabname = mqy.tabname
AND tb2.col_a = mqy.col_a + 1)
, run_queries AS
(SELECT qqq.*
, ttt.*
FROM make_queries qqq
, TABLE(tab_Transpose_4K(sql_text)) AS ttt
WHERE col_d = 1)
SELECT SUBSTR(tabname,1,11) AS tab_name
, SMALLINT(row_number) AS row#
, col_num AS col#
, CHAR(col_name,13) AS col_name
, CHAR(col_type,10) AS col_type
, col_length AS col_len
, SMALLINT(LENGTH(col_value)) AS val_len
, SUBSTR(col_value,1,20) AS col_value
FROM run_queries
ORDER BY 1,2,3;
Below is the answer (with a few values truncated to fit):
TAB_NAME | ROW# | COL# | COL_NAME | COL_TYPE | COL_LEN | VAL_LEN | COL_VALUE |
---|---|---|---|---|---|---|---|
EMP_PHOTO |
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMP_PHOTO |
1 |
2 |
PHOTO_FORMAT |
VARCHAR |
10 |
6 |
bitmap |
EMP_PHOTO |
1 |
3 |
PICTURE |
BLOB |
204800 |
- |
- |
EMP_PHOTO |
1 |
4 |
EMP_ROWID |
CHAR |
40 |
40 |
|
EMP_PHOTO |
2 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMP_PHOTO |
2 |
2 |
PHOTO_FORMAT |
VARCHAR |
10 |
3 |
gif |
EMP_PHOTO |
2 |
3 |
PICTURE |
BLOB |
204800 |
- |
- |
EMP_PHOTO |
2 |
4 |
EMP_ROWID |
CHAR |
40 |
40 |
|
EMP_RESUME |
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMP_RESUME |
1 |
2 |
RESUME_FORMAT |
VARCHAR |
10 |
5 |
ascii |
EMP_RESUME |
1 |
3 |
RESUME |
CLOB |
5120 |
- |
- |
EMP_RESUME |
1 |
4 |
EMP_ROWID |
CHAR |
40 |
40 |
|
EMP_RESUME |
2 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMP_RESUME |
2 |
2 |
RESUME_FORMAT |
VARCHAR |
10 |
4 |
html |
EMP_RESUME |
2 |
3 |
RESUME |
CLOB |
5120 |
- |
- |
EMP_RESUME |
2 |
4 |
EMP_ROWID |
CHAR |
40 |
40 |
|
EMPLOYEE |
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMPLOYEE |
1 |
2 |
FIRSTNME |
VARCHAR |
12 |
5 |
BRUCE |
EMPLOYEE |
1 |
3 |
MIDINIT |
CHAR |
1 |
1 |
|
EMPLOYEE |
1 |
4 |
LASTNAME |
VARCHAR |
15 |
7 |
ADAMSON |
EMPLOYEE |
1 |
5 |
WORKDEPT |
CHAR |
3 |
3 |
D11 |
EMPLOYEE |
1 |
6 |
PHONENO |
CHAR |
4 |
4 |
4510 |
EMPLOYEE |
1 |
7 |
HIREDATE |
DATE |
10 |
10 |
2002-02-12 |
EMPLOYEE |
1 |
8 |
JOB |
CHAR |
8 |
8 |
DESIGNER |
EMPLOYEE |
1 |
9 |
EDLEVEL |
SMALLINT |
6 |
2 |
16 |
EMPLOYEE |
1 |
10 |
SEX |
CHAR |
1 |
1 |
M |
EMPLOYEE |
1 |
11 |
BIRTHDATE |
DATE |
10 |
10 |
1977-05-17 |
EMPLOYEE |
1 |
12 |
SALARY |
DECIMAL |
11 |
8 |
55280.00 |
EMPLOYEE |
1 |
13 |
BONUS |
DECIMAL |
11 |
6 |
500.00 |
EMPLOYEE |
1 |
14 |
COMM |
DECIMAL |
11 |
7 |
2022.00 |
EMPPROJACT |
1 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMPPROJACT |
1 |
2 |
PROJNO |
CHAR |
6 |
6 |
MA2112 |
EMPPROJACT |
1 |
3 |
ACTNO |
SMALLINT |
6 |
2 |
60 |
EMPPROJACT |
1 |
4 |
EMPTIME |
DECIMAL |
7 |
4 |
1.00 |
EMPPROJACT |
1 |
5 |
EMSTDATE |
DATE |
10 |
10 |
2002-01-01 |
EMPPROJACT |
1 |
6 |
EMENDATE |
DATE |
10 |
10 |
2002-07-15 |
EMPPROJACT |
2 |
1 |
EMPNO |
CHAR |
6 |
6 |
000150 |
EMPPROJACT |
2 |
2 |
PROJNO |
CHAR |
6 |
6 |
MA2112 |
EMPPROJACT |
2 |
3 |
ACTNO |
SMALLINT |
6 |
3 |
|
180 |
EMPPROJACT |
2 |
4 |
EMPTIME |
DECIMAL |
7 |
4 |
1.00 |
EMPPROJACT |
2 |
5 |
EMSTDATE |
DATE |
10 |
10 |
2002-07-15 |
EMPPROJACT |
2 |
6 |
EMENDATE |
DATE |
10 |
10 |
2003-02-01 |
PROJECT |
1 |
1 |
PROJNO |
CHAR |
6 |
6 |
MA2112 |
PROJECT |
1 |
2 |
PROJNAME |
VARCHAR |
24 |
16 |
W L ROBOT |
PROJECT |
1 |
3 |
DEPTNO |
CHAR |
3 |
3 |
D11 |
PROJECT |
1 |
4 |
RESPEMP |
CHAR |
6 |
6 |
000150 |
PROJECT |
1 |
5 |
PRSTAFF |
DECIMAL |
7 |
4 |
3.00 |
PROJECT |
1 |
6 |
PRSTDATE |
DATE |
10 |
10 |
2002-01-01 |
PROJECT |
1 |
7 |
PRENDATE |
DATE |
10 |
10 |
1982-12-01 |
PROJECT |
1 |
8 |
MAJPROJ |
CHAR |
6 |
6 |
Below are the queries that were generated and run to get the above answer:
SELECT * FROM ACT WHERE ACTKWD = '000150'
SELECT * FROM DEPARTMENT WHERE MGRNO = '000150'
SELECT * FROM EMP_PHOTO WHERE EMPNO = '000150'
SELECT * FROM EMP_RESUME WHERE EMPNO = '000150'
SELECT * FROM EMPLOYEE WHERE EMPNO = '000150'
SELECT * FROM EXPLAIN_OPERATOR WHERE OPERATOR_TYPE = '000150'
SELECT * FROM PROJACT WHERE PROJNO = '000150'
SELECT * FROM EMPPROJACT WHERE EMPNO = '000150' OR PROJNO = '000150'
SELECT * FROM PROJECT WHERE MAJPROJ = '000150' OR PROJNO = '000150' OR
RESPEMP = '000150'
22.11.1. Function Definition
The Db2 user-defined tabular function that does the transposing is defined thus:
CREATE FUNCTION tab_Transpose (VARCHAR(4000))
RETURNS TABLE (row_number INTEGER
, num_cols SMALLINT
, col_num SMALLINT
, col_name VARCHAR(128)
, col_type VARCHAR(128)
, col_length INTEGER
, col_value VARCHAR(254))
LANGUAGE JAVA
EXTERNAL NAME 'Graeme2!tab_Transpose'
PARAMETER STYLE Db2GENERAL
NO EXTERNAL ACTION
NOT DETERMINISTIC
DISALLOW PARALLEL
READS SQL DATA
FINAL CALL
FENCED;
Java Code
import java.lang.*;
import COM.ibm.db2.app.*;
import java.sql.*;
import java.math.*;
import java.io.*;
public class Graeme2 extends UDF {
Connection con;
Statement stmt;
ResultSet rs;
ResultSetMetaData rsmtadta;
int rowNum;
int i;
int outLength;
short colNum;
int colCount;
String[] colName = new String[1100];
String[] colType = new String[1100];
int[] colSize = new int[1100];
public void writeRow() throws Exception {
set(2, rowNum);
set(3, (short) colCount);
set(4, colNum);
set(5, colName[colNum]);
set(6, colType[colNum]);
set(7, colSize[colNum]);
if (colType[colNum].equals("XML") ||
colType[colNum].equals("BLOB") ||
colType[colNum].equals("CLOB") ||
colType[colNum].equals("DBLOB") ||
colType[colNum].equals("GRAPHIC") ||
colType[colNum].equals("VARGRAPHIC") ||
colSize[colNum] > outLength) {
// DON'T DISPLAY THIS VALUE
return;
}
else if (rs.getString(colNum) != null) {
// DISPLAY THIS COLUMN VALUE
set(8, rs.getString(colNum));
}
}
public void tab_Transpose(String inStmt
, int rowNumber
, short numColumns
, short outColNumber
, String outColName
, String outColtype
, int outColSize
, String outColValue) throws Exception {
switch (getCallType()) {
case SQLUDF_TF_FIRST:
break;
case SQLUDF_TF_OPEN:
try {
con = DriverManager.getConnection("jdbc:default:connection");
stmt = con.createStatement();
rs = stmt.executeQuery(inStmt);
// GET COLUMN NAMES
rsmtadta = rs.getMetaData();
colCount = rsmtadta.getColumnCount();
for (i=1; i <= colCount; i++) {
colName[i] = rsmtadta.getColumnName(i);
colType[i] = rsmtadta.getColumnTypeName(i);
colSize[i] = rsmtadta.getColumnDisplaySize(i);
}
rowNum = 1;
colNum = 1;
outLength = 254;
}
catch(SQLException sqle) {
setSQLstate("38999");
setSQLmessage("SQLCODE = " + sqle.getSQLState());
return;
}
break;
case SQLUDF_TF_FETCH:
if (colNum == 1 && rs.next() == true) {
writeRow();
colNum++;
if (colNum > colCount) {
colNum = 1;
rowNum++;
}
}
else if (colNum > 1 && colNum <= colCount) {
writeRow();
colNum++;
if (colNum > colCount) {
colNum = 1;
rowNum++;
}
}
else {
setSQLstate ("02000");
}
break;
case SQLUDF_TF_CLOSE:
rs.close();
stmt.close();
con.close();
break;
case SQLUDF_TF_FINAL:
break;
}
}
}
22.12. Java Logic
OPEN (run once):
-
Establish connection.
-
Prepare the SQL statement (i.e. input string).
-
Execute the SQL statement (i.e. open cursor).
-
Get meta-data for each column returned by query.
-
Set row-number and column-number variables to one.
-
Set the maximum output length accepted to 254.
FETCH (run for each row/column instance):
-
If row exists and column-number is 1, fetch row.
-
For value is not null and of valid Db2 type, return row.
-
Increment row-number and column-number variables.
CLOSE (run once):
-
Close the cursor.
-
Return.
22.13. Update Real Data using Meta-Data
Db2 does not allow one to do DML or DDL using a scalar function, but one can do something similar by calling a table function. Thus if the table function defined below is joined to in a query, the following happens:
-
User query joins to table function - sends DML or DDL statement to be executed.
-
Table function calls stored procedure - sends statement to be executed.
-
Stored procedure executes statement.
-
Stored procedure returns SQLCODE of statement to the table function.
-
Table function joins back to the user query a single-row table with two columns: The SQLCODE and the original input statement.
Now for the code:
CREATE PROCEDURE execute_immediate (IN in_stmt VARCHAR(1000)
, OUT out_sqlcode INTEGER)
LANGUAGE SQL
MODIFIES SQL DATA
BEGIN
DECLARE sqlcode INTEGER;
DECLARE EXIT HANDLER FOR sqlexception
SET out_sqlcode = sqlcode;
EXECUTE IMMEDIATE in_stmt;
SET out_sqlcode = sqlcode;
RETURN;
END!
CREATE FUNCTION execute_immediate (in_stmt VARCHAR(1000))
RETURNS TABLE (sqltext VARCHAR(1000)
, sqlcode INTEGER)
LANGUAGE SQL
MODIFIES SQL DATA
BEGIN ATOMIC
DECLARE out_sqlcode INTEGER;
CALL execute_immediate(in_stmt, out_sqlcode);
RETURN VALUES (in_stmt, out_sqlcode);
END!
This example uses an "!" as the stmt delimiter. |
This code is extremely dangerous! Use with care. As we shall see, it is very easy for the above code to do some quite unexpected. |
22.13.1. Usage Examples
The following query gets a list of materialized query tables for a given table-schema that need to be refreshed, and then refreshes the table:
WITH temp1 AS
(SELECT tabschema
, tabname
FROM syscat.tables
WHERE tabschema = 'FRED'
AND type = 'S'
AND status = 'C'
AND tabname LIKE '%DEPT%')
SELECT CHAR(tab.tabname, 20) AS tabname
, stm.sqlcode AS sqlcode
, CHAR(stm.sqltext,100) AS sqltext
FROM temp1 AS tab
, TABLE(execute_immediate(
'REFRESH TABLE ' ||RTRIM(tab.tabschema) || '.' || tab.tabname))AS stm
ORDER BY tab.tabname
WITH UR;
I had two matching tables that needed to be refreshed, so I got the following answer:
TABNAME | SQLCODE | SQLTEXT |
---|---|---|
STAFF_DEPT1 |
0 |
REFRESH TABLE FRED.STAFF_DEPT1 |
STAFF_DEPT2 |
0 |
REFRESH TABLE FRED.STAFF_DEPT2 |
Observe above that the set of matching tables to be refreshed was defined in a common-tableexpression, and then joined to the table function. It is very important that one always code thus, because in an ordinary join it is possible for the table function to be called before all of the predicates have been applied. To illustrate this concept, the next query is supposed to make a copy of two matching tables. The answer indicates that it did just this. But what it actually did was make copies of many more tables - because the table function was called before all of the predicates on SYSCAT.TABLES were applied. The other tables that were created don’t show up in the query output, because they were filtered out later in the query processing:
SELECT CHAR(tab.tabname, 20) AS tabname
, stm.sqlcode AS sqlcode
, CHAR(stm.sqltext,100) AS sqltext
FROM syscat.tables AS tab
, TABLE(execute_immediate(
' CREATE TABLE ' || RTRIM(tab.tabschema) || '.' || tab.tabname || '_C1' ||
' LIKE ' || RTRIM(tab.tabschema) || '.' || tab.tabname))AS stm
WHERE tab.tabschema = USER
AND tab.tabname LIKE 'S%'
ORDER BY tab.tabname
FOR FETCH ONLY
WITH UR;
ANSWER
TABNAME | SQLCODE | SQLTEXT |
---|---|---|
SALES |
0 |
CREATE TABLE FRED.SALES_C1 LIKE FRED.SALES |
STAFF |
0 |
CREATE TABLE FRED.STAFF_C1 LIKE FRED.STAFF |
The above is bad enough, but I once managed to do much worse. In a variation of the above code, the query created a copy, of a copy, of a copy, etc. The table function kept finding the table just created, and making a copy of it - until the TABNAME reached the length limit. The correct way to create a copy of a set of tables is shown below. In this query, the list of tables to be copied is identified in a common table expression before the table function is called:
WITH temp1 AS
(SELECT tabschema
, tabname
FROM syscat.tables
WHERE tabschema = USER
AND tabname LIKE 'S%')
SELECT CHAR(tab.tabname, 20) AS tabname
, stm.sqlcode AS sqlcode
, CHAR(stm.sqltext,100) AS sqltext
FROM temp1 tab
, TABLE(execute_immediate(
' CREATE TABLE ' ||RTRIM(tab.tabschema) || '.' || tab.tabname || '_C1' ||
' LIKE ' || RTRIM(tab.tabschema) || '.' || tab.tabname))AS stm
ORDER BY tab.tabname
FOR FETCH ONLY
WITH UR;
ANSWER
TABNAME | SQLCODE | SQLTEXT |
---|---|---|
SALES |
0 |
CREATE TABLE FRED.SALES_C1 LIKE FRED.SALES |
STAFF |
0 |
CREATE TABLE FRED.STAFF_C1 LIKE FRED.STAFF |
The next example is similar to the previous, except that it creates a copy, and then populates the new table with the contents of the original table:
WITH temp0 AS
(SELECT RTRIM(tabschema) AS schema
, tabname AS old_tabname
, tabname || '_C2' AS new_tabname
FROM syscat.tables
WHERE tabschema = USER
AND tabname LIKE 'S%')
, temp1 AS
(SELECT tab.*
, stm.sqlcode AS sqlcode1
, CHAR(stm.sqltext,200) AS sqltext1
FROM temp0 AS tab
, TABLE(execute_immediate(
' CREATE TABLE ' || schema || '.' || new_tabname ||
' LIKE ' || schema || '.' || old_tabname))AS stm)
, temp2 AS
(SELECT tab.*
, stm.sqlcode AS sqlcode2
, CHAR(stm.sqltext,200) AS sqltext2
FROM temp1 AS tab
, TABLE(execute_immediate(
' INSERT INTO ' || schema || '.' || new_tabname ||
' SELECT * FROM ' || schema || '.' || old_tabname))AS stm)
SELECT CHAR(old_tabname,20) AS tabname
, sqlcode1
, sqlcode2
FROM temp2
ORDER BY old_tabname
FOR FETCH ONLY
WITH UR;
ANSWER
TABNAME | SQLCODE1 | SQLCODE2 |
---|---|---|
SALES |
0 |
0 |
STAFF |
0 |
0 |
22.13.2. Query Processing Sequence
In order to explain the above, we need to understand in what sequence the various parts of a query are executed in order to avoid semantic ambiguity:
-
FROM clause
-
JOIN ON clause
-
WHERE clause
-
GROUP BY and aggregate
-
HAVING clause
-
SELECT list
-
ORDER BY clause
-
FETCH FIRST
Observe above that the FROM clause is resolved before any WHERE predicates are applied. This is why the query in Create copies of tables - wrong did the wrong thing.
23. Fun with SQL
In this chapter will shall cover some of the fun things that one can and, perhaps, should not do, using Db2 SQL. Read on at your own risk.
23.1. Creating Sample Data
If every application worked exactly as intended from the first, we would never have any need for test databases. Unfortunately, one often needs to builds test systems in order to both tune the application SQL, and to do capacity planning. In this section we shall illustrate how very large volumes of extremely complex test data can be created using relatively simple SQL statements.
Good Sample Data is
-
Reproducible.
-
Easy to make.
-
Similar to Production:
-
Same data volumes (if needed).
-
Same data distribution characteristics.
-
23.1.1. Data Generation
Create the set of integers between zero and one hundred. In this statement we shall use recursive coding to expand a single value into many more.
WITH temp1 (col1) AS
(VALUES 0
UNION ALL
SELECT col1 + 1
FROM temp1
WHERE col1 + 1 < 100)
SELECT *
FROM temp1;
ANSWER
COL1 |
---|
1 |
2 |
3 |
etc |
Instead of coding a recursion join every time, we use the table function described on Generating Numbers to create the required rows. Assuming that the function exists, one would write the following:
SELECT *
FROM TABLE(NumList(100)) AS xxx;
23.1.2. Make Reproducible Random Data
So far, all we have done is create sets of fixed values. These are usually not suitable for testing purposes because they are too consistent. To mess things up a bit we need to use the RAND function, which generates random numbers in the range of zero to one inclusive. In the next example we will get a (reproducible) list of five random numeric values:
WITH temp1 (s1, r1) AS
(VALUES (0, RAND(1))
UNION ALL
SELECT s1+1
, RAND()
FROM temp1
WHERE s1+1 < 5)
SELECT SMALLINT(s1) AS seq#
, DECIMAL(r1,5,3) AS ran1
FROM temp1;
ANSWER
SEQ# | RAN1 |
---|---|
0 |
0.001 |
1 |
0.563 |
2 |
0.193 |
3 |
0.808 |
4 |
0.585 |
The initial invocation of the RAND function above is seeded with the value 1. Subsequent invocations of the same function (in the recursive part of the statement) use the initial value to generate a reproducible set of pseudo-random numbers.
23.1.3. Using the GENERATE_UNIQUE function
With a bit of data manipulation, the GENERATE_UNIQUE function can be used (instead of the RAND function) to make suitably random test data. The are advantages and disadvantages to using both functions:
-
The GENERATE_UNIQUE function makes data that is always unique. The RAND function only outputs one of 32,000 distinct values.
-
The RAND function can make reproducible random data, while the GENERATE_UNIQUE function can not.
See the description of the GENERATE_UNIQUE function (see GENERATE_UNIQUE) for an example of how to use it to make random data.
23.1.4. Make Random Data - Different Ranges
There are several ways to mess around with the output from the RAND function: We can use simple arithmetic to alter the range of numbers generated (e.g. convert from 0 to 10 to 0 to 10,000). We can alter the format (e.g. from FLOAT to DECIMAL). Lastly, we can make fewer, or more, distinct random values (e.g. from 32K distinct values down to just 10). All of this is done below:
WITH temp1 (s1, r1) AS
(VALUES (0
, RAND(2))
UNION ALL
SELECT s1+1
, RAND()
FROM temp1
WHERE s1+1 < 5)
SELECT SMALLINT(s1) AS seq#
, SMALLINT(r1*10000) AS ran2
, DECIMAL(r1,6,4) AS ran1
, SMALLINT(r1*10) AS ran3
FROM temp1;
ANSWER
SEQ# | RAN2 | RAN1 | RAN3 |
---|---|---|---|
0 |
13 |
0.0013 |
0 |
1 |
8916 |
0.8916 |
8 |
2 |
7384 |
0.7384 |
7 |
3 |
5430 |
0.5430 |
5 |
4 |
8998 |
0.8998 |
8 |
23.1.5. Make Random Data - Varying Distribution
In the real world, there is a tendency for certain data values to show up much more frequently than others. Likewise, separate fields in a table usually have independent semi-random data distribution patterns. In the next statement we create three independently random fields. The first has the usual 32K distinct values evenly distributed in the range of zero to one. The second and third have random numbers that are skewed towards the low end of the range, and have many more distinct values:
WITH temp1 (s1) AS
(VALUES (0)
UNION ALL
SELECT s1 + 1
FROM temp1
WHERE s1 + 1 < 5)
SELECT SMALLINT(s1) AS s#
, INTEGER((RAND(1)) * 1E6) AS ran1
, INTEGER((RAND() * RAND()) * 1E6) AS ran2
, INTEGER((RAND() * RAND()* RAND()) * 1E6) AS ran3
FROM temp1;
ANSWER
S# | RAN1 | RAN2 | RAN3 |
---|---|---|---|
0 |
1251 |
365370 |
114753 |
1 |
350291 |
280730 |
88106 |
2 |
710501 |
149549 |
550422 |
3 |
147312 |
33311 |
2339 |
4 |
8911 |
556 |
73091 |
23.1.6. Make Random Data - Different Flavours
The RAND function generates random numbers. To get random character data one has to convert the RAND output into a character. There are several ways to do this. The first method shown below uses the CHR function to convert a number in the range: 65 to 90 into the ASCII equivalent: "A" to "Z". The second method uses the CHAR function to translate a number into the character equivalent.
WITH temp1 (s1, r1) AS
(VALUES (0
, RAND(2))
UNION ALL
SELECT s1+1
, RAND()
FROM temp1
WHERE s1+1 < 5)
SELECT SMALLINT(s1) AS seq#
, SMALLINT(r1*26+65) AS ran2
, CHR(SMALLINT(r1*26+65)) AS ran3
, CHAR(SMALLINT(r1*26)+65) AS ran4
FROM temp1;
ANSWER
SEQ# | RAN2 | RAN3 | RAN4 |
---|---|---|---|
0 |
65 |
A |
65 |
1 |
88 |
X |
88 |
2 |
84 |
T |
84 |
3 |
79 |
O |
79 |
4 |
88 |
X |
88 |
23.1.7. Make Test Table & Data
So far, all we have done in this chapter is use SQL to select sets of rows. Now we shall create a Production-like table for performance testing purposes. We will then insert 10,000 rows of suitably lifelike test data into the table. The DDL, with constraints and index definitions, follows. The important things to note are:
-
The EMP# and the SOCSEC# must both be unique.
-
The JOB_FTN, FST_NAME, and LST_NAME fields must all be non-blank.
-
The SOCSEC# must have a special format.
-
The DATE_BN must be greater than 1900.
Several other fields must be within certain numeric ranges.
CREATE TABLE personnel
( emp# INTEGER NOT NULL
, socsec# CHAR(11) NOT NULL
, job_ftn CHAR(4) NOT NULL
, dept SMALLINT NOT NULL
, salary DECIMAL(7, 2) NOT NULL
, date_bn DATE NOT NULL WITH DEFAULT
, fst_name VARCHAR(20)
, lst_name VARCHAR(20)
, CONSTRAINT pex1 PRIMARY KEY (emp#)
, CONSTRAINT pe01 CHECK (emp# > 0)
, CONSTRAINT pe02 CHECK (LOCATE(' ', socsec#) = 0)
, CONSTRAINT pe03 CHECK (LOCATE('-', socsec#,1) = 4)
, CONSTRAINT pe04 CHECK (LOCATE('-', socsec#,5) = 7)
, CONSTRAINT pe05 CHECK (job_ftn <> '')
, CONSTRAINT pe06 CHECK (dept BETWEEN 1 AND 99)
, CONSTRAINT pe07 CHECK (salary BETWEEN 0 AND 99999)
, CONSTRAINT pe08 CHECK (fst_name <> '')
, CONSTRAINT pe09 CHECK (lst_name <> '')
, CONSTRAINT pe10 CHECK (date_bn >= '1900-01-01' ));
CREATE UNIQUE INDEX PEX2 ON PERSONNEL (SOCSEC#);
CREATE UNIQUE INDEX PEX3 ON PERSONNEL (DEPT, EMP#);
Now we shall populate the table. The SQL shall be described in detail latter. For the moment, note the four RAND fields. These contain, independently generated, random numbers which are used to populate the other data fields.
INSERT INTO personnel
WITH temp1 (s1, r1, r2, r3, r4) AS
(VALUES (0
, RAND(2)
, RAND() + (RAND() /1E5)
, RAND() * RAND()
, RAND() * RAND() * RAND())
UNION ALL
SELECT s1 + 1
, RAND()
, RAND() + (RAND() / 1E5)
, RAND() * RAND()
, RAND() * RAND() * RAND()
FROM temp1
WHERE s1 < 10000)
SELECT 100000 + s1
, SUBSTR(DIGITS(INT(r2*988+10)), 8) || '-' ||
SUBSTR(DIGITS(INT(r1*88+10)),9) || '-' ||
TRANSLATE(SUBSTR(DIGITS(s1), 7), '9873450126', '0123456789')
, CASE
WHEN INT(r4*9) > 7 THEN 'MGR'
WHEN INT(r4*9) > 5 THEN 'SUPR'
WHEN INT(r4*9) > 3 THEN 'PGMR'
WHEN INT(R4*9) > 1 THEN 'SEC'
ELSE 'WKR'
END
, INT(r3*98+1)
, DECIMAL(r4*99999, 7, 2)
, DATE('1930-01-01') + INT(50-(r4*50)) YEARS + INT(r4*11) MONTHS + INT(r4*27) DAYS
, CHR(INT(r1*26+65))|| CHR(INT(r2*26+97))|| CHR(INT(r3*26+97)) ||
CHR(INT(r4*26+97))|| CHR(INT(r3*10+97))|| CHR(INT(r3*11+97))
, CHR(INT(r2*26+65))|| TRANSLATE(CHAR(INT(r2*1E7)), 'aaeeiibmty', '0123456789')
FROM temp1;
Some sample data follows:
EMP# | SOCSEC# | JOB_ | DEPT | SALARY | DATE_BN | F_NME | L_NME |
---|---|---|---|---|---|---|---|
100000 |
484-10-9999 |
WKR |
47 |
13.63 |
1979-01-01 |
Ammaef |
Mimytmbi |
100001 |
449-38-9998 |
SEC |
53 |
35758.87 |
1962-04-10 |
Ilojff |
Liiiemea |
100002 |
979-90-9997 |
WKR |
1 |
8155.23 |
1975-01-03 |
Xzacaa |
Zytaebma |
100003 |
580-50-9993 |
WKR |
31 |
16643.50 |
1971-02-05 |
Lpiedd |
Pimmeeat |
100004 |
264-87-9994 |
WKR |
21 |
962.87 |
1979-01-01 |
Wgfacc |
Geimteei |
100005 |
661-84-9995 |
WKR |
19 |
4648.38 |
1977-01-02 |
Wrebbc |
Rbiybeet |
100006 |
554-53-9990 |
WKR |
8 |
375.42 |
1979-01-01 |
Mobaaa |
Oiiaiaia |
100007 |
482-23-9991 |
SEC |
36 |
23170.09 |
1968-03-07 |
Emjgdd |
Mimtmamb |
100008 |
536-41-9992 |
WKR |
6 |
10514.11 |
1974-02-03 |
Jnbcaa |
Nieebayt |
In order to illustrate some of the tricks that one can use when creating such data, each field above was calculated using a different schema:
-
The EMP# is a simple ascending number.
-
The SOCSEC# field presented three problems: It had to be unique, it had to be random with respect to the current employee number, and it is a character field with special layout constraints (see the DDL on Production-like test table DDL).
-
To make it random, the first five digits were defined using two of the temporary random number fields. To try and ensure that it was unique, the last four digits contain part of the employee number with some digit-flipping done to hide things. Also, the first random number used is the one with lots of unique values. The special formatting that this field required is addressed by making everything in pieces and then concatenating.
-
The JOB FUNCTION is determined using the fourth (highly skewed) random number. This ensures that we get many more workers than managers.
-
The DEPT is derived from another, somewhat skewed, random number with a range of values from one to ninety nine.
-
The SALARY is derived using the same, highly skewed, random number that was used for the job function calculation. This ensures that theses two fields have related values.
-
The BIRTH DATE is a random date value somewhere between 1930 and 1981.
-
The FIRST NAME is derived using seven independent invocation of the CHR function, each of which is going to give a somewhat different result.
-
The LAST NAME is (mostly) made by using the TRANSLATE function to convert a large random number into a corresponding character value. The output is skewed towards some of the vowels and the lower-range characters during the translation.
23.1.8. Time-Series Processing
The following table holds data for a typical time-series application. Observe is that each row has both a beginning and ending date, and that there are three cases where there is a gap between the end-date of one row and the begin-date of the next (with the same key).
CREATE TABLE time_series
( KYY CHAR(03) NOT NULL
, bgn_dt DATE NOT NULL
, end_dt DATE NOT NULL
, CONSTRAINT tsc1 CHECK (kyy <> '')
, CONSTRAINT tsc2 CHECK (bgn_dt <= end_dt));
COMMIT;
INSERT INTO TIME_series values
('AAA','1995-10-01','1995-10-04')
, ('AAA','1995-10-06','1995-10-06')
, ('AAA','1995-10-07','1995-10-07')
, ('AAA','1995-10-15','1995-10-19')
, ('BBB','1995-10-01','1995-10-01')
, ('BBB','1995-10-03','1995-10-03');
Find Overlapping Rows
We want to find any cases where the begin-to-end date range of one row overlaps another with the same KYY value. The following diagram illustrates our task. The bold line at the top represents the begin and end date for a row. This row is overlapped (in time) by the six lower rows, but the nature of the overlap differs in each case.
time >>>------------------------------------------------------------------>>> <<<------------------------------>>> <<<------------------------------>>> <<<--->>> <<<--->>> <<<------------->>> <<<---------->>> <<<---------------------------------------------------------->>>>
The general types of overlap are:
-
The related row has identical date ranges.
-
The related row begins before the start-date and ends after the same.
-
The row begins and ends between the start and finish dates.
When writing SQL to check overlapping data ranges, make sure that all possible types of overlap (see diagram above) are tested. Some SQL statements work with some flavors of overlap, but not with others. |
The relevant SQL follows. When reading it, think of the "A" table as being the bold line above and "B" table as being the four overlapping rows shown as single lines.
SELECT kyy
, bgn_dt
, end_dt
FROM time_series a
WHERE EXISTS
(SELECT *
FROM time_series b
WHERE a.kyy = b.kyy
AND a.bgn_dt <> b.bgn_dt
AND (a.bgn_dt BETWEEN b.bgn_dt AND b.end_dt
OR
b.bgn_dt BETWEEN a.bgn_dt AND a.end_dt)
)
ORDER BY 1,2;
ANSWER:
KEYCOL | BGN_DT | END_DT | BGN_DT | END_DT | DIFF |
---|---|---|---|---|---|
AAA |
1995-10-01 |
1995-10-04 |
1995-10-06 |
1995-10-06 |
2 |
AAA |
1995-10-07 |
1995-10-07 |
1995-10-15 |
1995-10-19 |
8 |
BBB |
1995-10-01 |
1995-10-01 |
1995-10-03 |
1995-10-03 |
2 |
The first predicate in the above sub-query joins the rows together by matching key value. The second predicate makes sure that one row does not match against itself. The final two predicates look for overlapping date ranges.
The above query relies on the sample table data being valid (as defined by the CHECK constraints in the DDL on Sample Table DDL - Time Series. This means that the END_DT is always greater than or equal to the BGN_DT, and each KYY, BGN_DT combination is unique.
Find Gaps in Time-Series
We want to find all those cases in the TIME_SERIES table when the ending of one row is not exactly one day less than the beginning of the next (if there is a next). The following query will answer this question. It consists of both a join and a sub-query. In the join (which is done first), we match each row with every other row that has the same key and a BGN_DT that is more than one day greater than the current END_DT. Next, the sub-query excludes from the result those join-rows where there is an intermediate third row.
SELECT a.kyy
, a.bgn_dt
, a.end_dt
, b.bgn_dt
, b.end_dt
, DAYS(b.bgn_dt) - DAYS(A.end_dt) as diff
FROM time_series a
, time_series b
WHERE a.kyy = b.kyy
AND a.end_dt < b.bgn_dt - 1 DAY
AND NOT EXISTS
(SELECT *
FROM time_series z
WHERE z.kyy = a.kyy
AND z.kyy = b.kyy
AND z.bgn_dt > a.bgn_dt
AND z.bgn_dt < b.bgn_dt)
ORDER BY 1,2;
TIME_SERIES
KYY | BGN_DT | END_DT |
---|---|---|
AAA |
1995-10-01 |
1995-10-04 |
AAA |
1995-10-06 |
1995-10-06 |
AAA |
1995-10-07 |
1995-10-07 |
AAA |
1995-10-15 |
1995-10-19 |
BBB |
1995-10-01 |
1995-10-01 |
BBB |
1995-10-03 |
1995-10-03 |
ANSWER
KEYCOL | BGN_DT | END_DT | BGN_DT | END_DT | DIFF |
---|---|---|---|---|---|
AAA |
1995-10-01 |
1995-10-04 |
1995-10-06 |
1995-10-06 |
2 |
AAA |
1995-10-07 |
1995-10-07 |
1995-10-15 |
1995-10-19 |
8 |
BBB |
1995-10-01 |
1995-10-01 |
1995-10-03 |
1995-10-03 |
2 |
If there are many rows per key value, the above SQL will be very inefficient. This is because the join (done first) does a form of Cartesian Product (by key value) making an internal result table that can be very large. The sub-query then cuts this temporary table down to size by removing results-rows that have other intermediate rows. |
Instead of looking at those rows that encompass a gap in the data, we may want to look at the actual gap itself. To this end, the following SQL differs from the prior in that the SELECT list has been modified to get the start, end, and duration, of each gap.
SELECT a.kyy AS kyy
, a.end_dt + 1 DAY AS bgn_gap
, b.bgn_dt - 1 DAY AS end_gap
, (DAYS(b.bgn_dt) - DAYS(a.end_dt) - 1) AS sz
FROM time_series a
, time_series b
WHERE a.kyy = b.kyy
AND a.end_dt < b.bgn_dt - 1 DAY
AND NOT EXISTS
(SELECT *
FROM time_series z
WHERE z.kyy = a.kyy
AND z.kyy = b.kyy
AND z.bgn_dt > a.bgn_dt
AND z.bgn_dt < b.bgn_dt)
ORDER BY 1,2;
TIME_SERIES
KYY | BGN_DT | END_DT |
---|---|---|
AAA |
1995-10-01 |
1995-10-04 |
AAA |
1995-10-06 |
1995-10-06 |
AAA |
1995-10-07 |
1995-10-07 |
AAA |
1995-10-15 |
1995-10-19 |
BBB |
1995-10-01 |
1995-10-01 |
BBB |
1995-10-03 |
1995-10-03 |
ANSWER
KYY | BGN_GAP | END_GAP | SZ |
---|---|---|---|
AAA |
1995-10-05 |
1995-10-05 |
1 |
AAA |
1995-10-08 |
1995-10-14 |
7 |
BBB |
1995-10-02 |
1995-10-02 |
1 |
Show Each Day in Gap
Imagine that we wanted to see each individual day in a gap. The following statement does this by taking the result obtained above and passing it into a recursive SQL statement which then generates additional rows - one for each day in the gap after the first.
WITH temp (kyy, gap_dt, gsize) AS
(SELECT a.kyy
, a.end_dt + 1 DAY
, (DAYS(b.bgn_dt) DAYS(a.end_dt) - 1)
FROM time_series a
, time_series b
WHERE a.kyy = b.kyy
AND a.end_dt < b.bgn_dt - 1 DAY
AND NOT EXISTS
(SELECT *
FROM time_series z
WHERE z.kyy = a.kyy
AND z.kyy = b.kyy
AND z.bgn_dt > a.bgn_dt
AND z.bgn_dt < b.bgn_dt)
UNION ALL
SELECT kyy
, gap_dt + 1 DAY
, gsize - 1
FROM temp WHERE gsize > 1)
SELECT *
FROM temp
ORDER BY 1, 2;
ANSWER
KEYCOL | GAP_DT | GSIZE |
---|---|---|
AAA |
1995-10-05 |
1 |
AAA |
1995-10-08 |
7 |
AAA |
1995-10-09 |
6 |
AAA |
1995-10-10 |
5 |
AAA |
1995-10-11 |
4 |
AAA |
1995-10-12 |
3 |
AAA |
1995-10-13 |
2 |
AAA |
1995-10-14 |
1 |
BBB |
1995-10-02 |
1 |
23.2. Other Fun Things
23.2.1. Randomly Sample Data
One can use the TABLESAMPLE schema to randomly sample rows for subsequent analysis.
Notes
-
The table-name must refer to a real table. This can include a declared global temporary table, or a materialized query table. It cannot be a nested table expression.
-
The sampling is an addition to any predicates specified in the where clause. Under the covers, sampling occurs before any other query processing, such as applying predicates or doing a join.
-
The BERNOULLI option checks each row individually.
-
The SYSTEM option lets Db2 find the most efficient way to sample the data. This may mean that all rows on each page that qualifies are included. For small tables, this method often results in a misleading percentage of rows selected.
-
The "percent" number must be equal to or less than 100, and greater than zero. It determines what percentage of the rows processed are returns.
-
The REPEATABLE option and number is used if one wants to get the same result every time the query is run (assuming no data changes). Without this option, each run will be both random and different.
Examples
Sample 5% of the rows in the staff table. Get the same result each time:
SELECT *
FROM staff TABLESAMPLE BERNOULLI(5) REPEATABLE(1234)
ORDER BY id;
Sample 18% of the rows in the employee table and 25% of the rows in the employee-activity table, then join the two tables together. Because each table is sampled independently, the fraction of rows that join will be much less either sampling rate:
SELECT *
FROM employee ee TABLESAMPLE BERNOULLI(18)
, emp_act ea TABLESAMPLE BERNOULLI(25)
WHERE ee.empno = ea.empno
ORDER BY ee.empno;
Sample a declared global temporary table, and also apply other predicates:
DECLARE GLOBAL TEMPORARY TABLE session.nyc_staff
LIKE staff;
SELECT *
FROM session.nyc_staff TABLESAMPLE SYSTEM(34.55)
WHERE id < 100
AND salary > 100
ORDER BY id;
23.3. Convert Character to Numeric
The DOUBLE, DECIMAL, INTEGER, SMALLINT, and BIGINT functions call all be used to convert a character field into its numeric equivalent:
WITH temp1 (c1) AS
(VALUES '123 ', ' 345 ', ' 567')
SELECT c1
, DOUBLE(c1) AS dbl
, DECIMAL(c1,3) AS dec
, SMALLINT(c1) AS sml
, INTEGER(c1) AS int
FROM temp1;
ANSWER (numbers shortened)
C1 | DBL | DEC | SML | INT |
---|---|---|---|---|
123 |
+1.2300E+2 |
123. |
123 |
123 |
345 |
+3.4500E+2 |
345. |
345 |
345 |
567 |
+5.6700E+2 |
567. |
567 |
567 |
Not all numeric functions support all character representations of a number. The following table illustrates what’s allowed and what’s not: .Acceptable conversion values
INPUT STRING | COMPATIBLE FUNCTIONS |
---|---|
" 1234" |
DOUBLE, DECIMAL, INTEGER, SMALLINT, BIGINT |
" 12.4" |
DOUBLE, DECIMAL |
" 12E4" |
DOUBLE |
23.3.1. Checking the Input
There are several ways to check that the input character string is a valid representation of a number - before doing the conversion. One simple solution involves converting all digits to blank, then removing the blanks. If the result is not a zero length string, then the input must have had a character other than a digit:
WITH temp1 (c1) AS
(VALUES ' 123', '456 ', ' 1 2', ' 33%', NULL)
SELECT c1
, TRANSLATE(c1, ' ', '1234567890') AS c2
, LENGTH(LTRIM(TRANSLATE(c1,' ', '1234567890'))) AS c3
FROM temp1;
ANSWER
C1 C2 C3 ---- ---- -- 123 0 456 0 1 2 0 33% % 1 - - -
One can also write a user-defined scalar function to check for non-numeric input, which is what is done below. This function returns "Y" if the following is true:
-
The input is not null.
-
There are no non-numeric characters in the input.
-
The only blanks in the input are to the left of the digits.
-
There is only one "+" or "-" sign, and it is next to the left-side blanks, if any.
-
There is at least one digit in the input.
Now for the code:
This example uses an "!" as the stmt delimiter |
--#SET DELIMITER !
CREATE FUNCTION isnumeric(instr VARCHAR(40))
RETURNS CHAR(1)
BEGIN ATOMIC
DECLARE is_number CHAR(1) DEFAULT 'Y';
DECLARE bgn_blank CHAR(1) DEFAULT 'Y';
DECLARE found_num CHAR(1) DEFAULT 'N';
DECLARE found_pos CHAR(1) DEFAULT 'N';
DECLARE found_neg CHAR(1) DEFAULT 'N';
DECLARE found_dot CHAR(1) DEFAULT 'N';
DECLARE ctr SMALLINT DEFAULT 1;
IF instr IS NULL THEN
RETURN NULL;
END IF;
wloop:
WHILE ctr <= LENGTH(instr) AND is_number = 'Y'
DO
----------------------------
--- ERROR CHECKS ---
---------------------------
IF SUBSTR(instr, ctr, 1) NOT IN (' ', '.', '+', '-', '0', '1', '2'
,'3', '4', '5', '6', '7', '8', '9') THEN
SET is_number = 'N';
ITERATE wloop;
END IF;
IF SUBSTR(instr, ctr, 1) = ' ' AND bgn_blank = 'N' THEN
SET is_number = 'N';
ITERATE wloop;
END IF;
IF SUBSTR(instr, ctr, 1) = '.' AND found_dot = 'Y' THEN
SET is_number = 'N';
ITERATE wloop;
END IF;
IF SUBSTR(instr,ctr,1) = '+' AND (found_neg = 'Y' OR bgn_blank = 'N') THEN
SET is_number = 'N';
ITERATE wloop;
END IF;
IF SUBSTR(instr,ctr,1) = '-' AND (found_neg = 'Y' OR bgn_blank = 'N') THEN
SET is_number = 'N';
ITERATE wloop;
END IF;
----------------------------
--- MAINTAIN FLAGS & CTR ---
---------------------------
IF SUBSTR(instr,ctr,1) IN ('0', '1', '2', '3', '4'
,'5', '6', '7', '8', '9') THEN
SET found_num = 'Y';
END IF;
IF SUBSTR(instr,ctr,1) = '.' THEN
SET found_dot = 'Y';
END IF;
IF SUBSTR(instr,ctr,1) = '+' THEN
SET found_pos = 'Y';
END IF;
IF SUBSTR(instr,ctr,1) = '-' THEN
SET found_neg = 'Y';
END IF;
IF SUBSTR(instr,ctr,1) <> ' ' THEN
SET bgn_blank = 'N';
END IF;
SET ctr = ctr + 1;
END WHILE wloop;
IF found_num = 'N' THEN
SET is_number = 'N';
END IF;
RETURN is_number;
END!
WITH TEMP1 (C1) AS
(VALUES ' 123'
, '+123.45'
, '456 '
, ' 10 2 '
, ' -.23'
, '++12356'
, '.012349'
, ' 33%'
, ' '
, NULL)
SELECT C1 AS C1
, isnumeric(C1) AS C2
, CASE
WHEN isnumeric(C1) = 'Y' THEN DECIMAL(C1, 10, 6)
ELSE NULL
END AS C3
FROM TEMP1!
ANSWER
C1 | C2 | C3 |
---|---|---|
123 |
Y |
123.00000 |
+123.45 |
Y |
123.45000 |
456 |
N |
- |
10 2 |
N |
- |
.23 |
Y |
-0.23000 |
++12356 |
N |
- |
.012349 |
Y |
0.01234 |
33% |
N |
- |
N |
- |
|
- |
- |
- |
See Check Data Value Type for a much simpler function that is similar to the above. |
23.4. Convert Number to Character
The CHAR and DIGITS functions can be used to convert a Db2 numeric field to a character representation of the same, but as the following example demonstrates, both functions return problematic output:
SELECT d_sal
, CHAR(d_sal) AS d_chr
, DIGITS(d_sal) AS d_dgt
, i_sal
, CHAR(i_sal) AS i_chr
, DIGITS(i_sal) AS i_dgt
FROM (SELECT DEC(salary - 11000, 6, 2) AS d_sal
, SMALLINT(salary - 11000) AS i_sal
FROM staff
WHERE salary > 10000
AND salary < 12200) AS xxx
ORDER BY d_sal;
ANSWER
D_SAL | D_CHR | D_DGT | I_SAL | I_CHR | I_DGT |
---|---|---|---|---|---|
494.10 |
-0494.10 |
049410 |
-494 |
-494 |
00494 |
-12.00 |
-0012.00 |
001200 |
-12 |
-12 |
00012 |
508.60 |
0508.60 |
050860 |
508 |
508 |
00508 |
1009.75 |
1009.75 |
100975 |
1009 |
1009 |
01009 |
The DIGITS function discards both the sign indicator and the decimal point, while the CHAR function output is (annoyingly) left-justified, and (for decimal data) has leading zeros. We can do better.
Below are three user-defined functions that convert integer data from numeric to character, displaying the output right-justified, and with a sign indicator if negative. There is one function for each flavor of integer that is supported in Db2:
CREATE FUNCTION char_right(inval SMALLINT)
RETURNS CHAR(06)
RETURN RIGHT(CHAR('',06) CONCAT RTRIM(CHAR(inval)),06);
CREATE FUNCTION char_right(inval INTEGER)
RETURNS CHAR(11)
RETURN RIGHT(CHAR('',11) CONCAT RTRIM(CHAR(inval)),11);
CREATE FUNCTION char_right(inval BIGINT)
RETURNS CHAR(20)
RETURN RIGHT(CHAR('',20) CONCAT RTRIM(CHAR(inval)),20);
Each of the above functions works the same way (working from right to left):
-
First, convert the input number to character using the CHAR function.
-
Next, use the RTRIM function to remove the right-most blanks.
-
Then, concatenate a set number of blanks to the left of the value. The number of blanks appended depends upon the input type, which is why there are three separate functions.
-
Finally, use the RIGHT function to get the right-most "n" characters, where "n" is the maximum number of digits (plus the sign indicator) supported by the input type.
The next example uses the first of the above functions:
SELECT i_sal
, char_right(i_sal) AS i_chr
FROM (SELECT SMALLINT(salary - 11000) AS i_sal
FROM staff
WHERE salary > 10000
AND salary < 12200) AS xxx
ORDER BY i_sal;
ANSWER
I_SAL | I_CHR |
---|---|
494 |
-494 |
-12 |
-12 |
508 |
508 |
1009 |
1009 |
23.5. Decimal Input
Creating a similar function to handle decimal input is a little more tricky. One problem is that the CHAR function adds leading zeros to decimal data, which we don’t want. A more serious problem is that there are many sizes and scales of decimal data, but we can only create one function (with a given name) for a particular input data type. Decimal values can range in both length and scale from 1 to 31 digits. This makes it impossible to define a single function to convert any possible decimal value to character with possibly running out of digits, or losing some precision.
The fact that one can only have one user-defined function, with a given name, per Db2 data type, presents a problem for all variable-length data types - notably character, varchar, and decimal. For character and varchar data, one can address the problem, to some extent, by using maximum length input and output fields. But decimal data has both a scale and a length, so there is no way to make an all-purpose decimal function. |
Despite the above, below is a function that converts decimal data to character. It compromises by assuming an input of type decimal(22,2), which should handle most monetary values:
CREATE FUNCTION char_right(inval DECIMAL(20,2))
RETURNS CHAR(22)
RETURN RIGHT(CHAR('', 19) CONCAT
REPLACE(SUBSTR(CHAR(inval * 1), 1, 1), '0', '') CONCAT
STRIP(CHAR(ABS(BIGINT(inval)))) CONCAT
'.' CONCAT
SUBSTR(DIGITS(inval), 19, 2), 22);
The function works as follows:
-
The input value is converted to CHAR and the first byte obtained. This will be a minus sign if the number is negative, else blank.
-
The non-fractional part of the number is converted to BIGINT then to CHAR.
-
A period (dot) is included.
-
The fractional digits (converted to character using the DIGITS function) are appended to the back of the output.
-
All of the above is concatenation together, along with some leading blanks. Finally, the 22 right-most characters are returned.
Below is the function in action:
WITH temp1 (num, tst) AS
(VALUES (1, DEC(0.01, 20, 2))
UNION ALL
SELECT num + 1
, tst * -3.21
FROM temp1
WHERE num < 8)
SELECT num
, tst
, char_right(tst) AS tchar
FROM temp1;
ANSWER
NUM | TST | TCHAR |
---|---|---|
1 |
0.01 |
0.01 |
2 |
-0.03 |
-0.03 |
3 |
0.09 |
0.09 |
4 |
-0.28 |
-0.28 |
5 |
0.89 |
0.89 |
6 |
-2.85 |
-2.85 |
7 |
9.14 |
9.14 |
8 |
-29.33 |
-29.33 |
Floating point data can be processed using the above function, as long as it is first converted to decimal using the standard DECIMAL function.
23.5.1. Adding Commas
The next function converts decimal input to character, with embedded comas. It first coverts the value to character - as per the above function. It then steps though the output string, three bytes at a time, from right to left, checking to see if the next-left character is a number. If it is, it insert a comma, else it adds a blank byte to the front of the string:
CREATE FUNCTION comma_right(inval DECIMAL(20, 2))
RETURNS CHAR(27)
LANGUAGE SQL
DETERMINISTIC
NO EXTERNAL ACTION
BEGIN ATOMIC
DECLARE i INTEGER DEFAULT 17;
DECLARE abs_inval BIGINT;
DECLARE out_value CHAR(27);
SET abs_inval = ABS(BIGINT(inval));
SET out_value = RIGHT(CHAR('', 19) CONCAT
RTRIM(CHAR(BIGINT(inval))),19) CONCAT
'.' CONCAT
SUBSTR(DIGITS(inval),19,2);
WHILE i > 2 DO
IF SUBSTR(out_value, i-1, 1) BETWEEN '0' AND '9' THEN
SET out_value = SUBSTR(out_value,1,i-1) CONCAT
',' CONCAT
SUBSTR(out_value,i);
ELSE
SET out_value = ' ' CONCAT out_value;
END IF;
SET i = i - 3;
END WHILE;
RETURN out_value;
END
Below is an example of the above function in use:
WITH temp1 (num) AS
(VALUES (DEC(+1,20,2))
, (DEC(-1,20,2))
UNION ALL
SELECT num * 987654.12
FROM temp1
WHERE ABS(num) < 1E10)
, temp2 (num) AS
(SELECT num - 1
FROM temp1)
SELECT num AS input
, comma_right(num) AS output
FROM temp2
ORDER BY num;
ANSWER
INPUT | OUTPUT |
---|---|
-975460660753.97 |
-975,460,660,753.97 |
-987655.12 |
-987,655.12 |
-2.00 |
-2.00 |
0.00 |
0.00 |
987653.12 |
987,653.12 |
975460660751.97 |
975,460,660,751.97 |
23.5.2. Convert Timestamp to Numeric
There is absolutely no sane reason why anyone would want to convert a date, time, or timestamp value directly to a number. The only correct way to manipulate such data is to use the provided date/time functions. But having said that, here is how one does it:
WITH tab1(ts1) AS
(VALUES CAST('1998-11-22-03.44.55.123456' AS TIMESTAMP))
SELECT ts1 --=> 1998-11-22-03.44.55.123456
, HEX(ts1) --=> 19981122034455123456
, DEC(HEX(ts1), 20) --=> 19981122034455123456.
, FLOAT(DEC(HEX(ts1), 20)) --=> 1.99811220344551e+019
, REAL (DEC(HEX(ts1), 20)) --=> 1.998112e+019
FROM tab1;
23.5.3. Selective Column Output
There is no way in static SQL to vary the number of columns returned by a select statement. In order to change the number of columns you have to write a new SQL statement and then rebind. But one can use CASE logic to control whether or not a column returns any data. Imagine that you are forced to use static SQL. Furthermore, imagine that you do not always want to retrieve the data from all columns, and that you also do not want to transmit data over the network that you do not need. For character columns, we can address this problem by retrieving the data only if it is wanted, and otherwise returning to a zero-length string. To illustrate, here is an ordinary SQL statement:
SELECT empno
, firstnme
, lastname
, job
FROM employee
WHERE empno < '000100'
ORDER BY empno;
Here is the same SQL statement with each character column being checked against a hostvariable. If the host-variable is 1, the data is returned, otherwise a zero-length string:
SELECT empno
, CASE :host-var-1
WHEN 1 THEN firstnme
ELSE ''
END AS firstnme
, CASE :host-var-2
WHEN 1 THEN lastname
ELSE ''
END AS lastname
, CASE :host-var-3
WHEN 1 THEN VARCHAR(job)
ELSE ''
END AS job
FROM employee
WHERE empno < '000100'
ORDER BY empno;
23.5.4. Making Charts Using SQL
Imagine that one had a string of numeric values that one wants to display as a line-bar chart. With a little coding, this is easy to do in SQL:
SELECT id
, salary
, INT(salary / 1500) AS len
, REPEAT('*', INT(salary / 1500)) AS salary_chart
FROM staff
WHERE id > 120
AND id < 190
ORDER BY id;
ANSWER
ID | SALARY | LEN | SALARY_CHART |
---|---|---|---|
130 |
10505.90 |
7 |
* |
140 |
21150.00 |
14 |
** |
150 |
19456.50 |
12 |
** |
160 |
22959.20 |
15 |
* |
170 |
12258.50 |
8 |
** |
180 |
12009.75 |
8 |
** |
To create the above graph we first converted the column of interest to an integer field of a manageable length, and then used this value to repeat a single "" character a set number of times. One problem with the above query is that we won’t know how long the chart will be until we run the statement. This may cause problems if we guess wrongly and we are tight for space. The next query addresses this issue by creating a chart of known length. It does it by dividing the row value by the maximum value for the selected rows (all divided by 20). The result is used to repeat the "" character "n" times:
SELECT dept
, id
, salary
, VARCHAR(REPEAT('*', INT(salary / (MAX(salary) OVER() / 20))), 20) AS chart
FROM staff
WHERE dept <= 15
AND id >= 100
ORDER BY 1,2;
ANSWER
DEPT | ID | SALARY | CHART |
---|---|---|---|
10 |
160 |
82959.20 |
** |
10 |
210 |
90010.00 |
|
10 |
240 |
79260.25 |
* |
10 |
260 |
81234.00 |
** |
15 |
110 |
42508.20 |
*** |
15 |
170 |
42258.50 |
*** |
The above code can be enhanced to have two charts in the same column. To illustrate, the next query expresses the salary as a chart, but separately by department. This can be useful to do when the two departments have very different values and one wants to analyze the data in each department independently:
SELECT dept
, id
, salary
, VARCHAR(REPEAT('*',
INT(salary / (MAX(salary)
OVER(PARTITION BY dept) / 20)
)
), 20) AS chart
FROM staff
WHERE dept <= 15
AND id >= 100
ORDER BY 1,2;
ANSWER
DEPT | ID | SALARY | CHART |
---|---|---|---|
10 |
160 |
82959.20 |
** |
10 |
210 |
90010.00 |
|
10 |
240 |
79260.25 |
* |
10 |
260 |
81234.00 |
** |
15 |
110 |
42508.20 |
|
15 |
170 |
42258.50 |
*** |
23.5.5. Multiple Counts in One Pass
Suppose we have a STATS table that has a SEX field with just two values, 'F' (for female) and 'M' (for male). To get a count of the rows by sex we can write the following:
SELECT sex
, COUNT(*) AS num
FROM stats
GROUP BY sex
ORDER BY sex;
ANSWER
SEX | NUM |
---|---|
F |
595 |
M |
405 |
Imagine now that we wanted to get a count of the different sexes on the same line of output. One, not very efficient, way to get this answer is shown below. It involves scanning the data table twice (once for males, and once for females) then joining the result.
WITH f (f) AS
(SELECT COUNT(*) FROM stats WHERE sex = 'F')
, m (m) AS
(SELECT COUNT(*) FROM stats WHERE sex = 'M')
SELECT f, m
FROM f
, m;
It would be more efficient if we answered the question with a single scan of the data table. This we can do using a CASE statement and a SUM function:
SELECT SUM(CASE sex WHEN 'F' THEN 1 ELSE 0 END) AS female
, SUM(CASE sex WHEN 'M' THEN 1 ELSE 0 END) AS male
FROM stats;
We can now go one step further and also count something else as we pass down the data. In the following example we get the count of all the rows at the same time as we get the individual sex counts.
SELECT COUNT(*) AS total
, SUM(CASE sex WHEN 'F' THEN 1 ELSE 0 END) AS female
, SUM(CASE sex WHEN 'M' THEN 1 ELSE 0 END) AS male
FROM stats;
23.5.6. Find Missing Rows in Series / Count all Values
One often has a sequence of values (e.g. invoice numbers) from which one needs both found and not-found rows. This cannot be done using a simple SELECT statement because some of rows being selected may not actually exist. For example, the following query lists the number of staff that have worked for the firm for "n" years, but it misses those years during which no staff joined:
SELECT years
, COUNT(*) AS #staff
FROM staff
WHERE UCASE(name) LIKE '%E%'
AND years <= 5
GROUP BY years;
ANSWER
YEARS | #STAFF |
---|---|
1 |
1 |
4 |
2 |
5 |
3 |
The simplest way to address this problem is to create a complete set of target values, then do an outer join to the data table. This is what the following example does:
WITH list_years (year#) AS
(VALUES (0), (1), (2), (3), (4), (5))
SELECT year# AS years
, COALESCE(#stff, 0) AS #staff
FROM list_years
LEFT OUTER JOIN
(SELECT years
, COUNT(*) AS #stff
FROM staff
WHERE UCASE(name) LIKE '%E%'
AND years <= 5
GROUP BY years) AS xxx
ON year# = years
ORDER BY 1;
ANSWER
YEARS | #STAFF |
---|---|
0 |
0 |
1 |
1 |
2 |
0 |
3 |
0 |
4 |
2 |
5 |
3 |
The use of the VALUES syntax to create the set of target rows, as shown above, gets to be tedious if the number of values to be made is large. To address this issue, the following example uses recursion to make the set of target values:
WITH list_years (year#) AS
(VALUES SMALLINT(0)
UNION ALL
SELECT year# + 1
FROM list_years
WHERE year# < 5)
SELECT year# AS years
, COALESCE(#stff, 0) AS #staff
FROM list_years
LEFT OUTER JOIN
(SELECT years
, COUNT(*) AS #stff
FROM staff
WHERE UCASE(name) LIKE '%E%'
AND years <= 5
GROUP BY years) AS xxx
ON year# = years
ORDER BY 1;
ANSWER
YEARS | #STAFF |
---|---|
0 |
0 |
1 |
1 |
2 |
0 |
3 |
0 |
4 |
2 |
5 |
3 |
If one turns the final outer join into a (negative) sub-query, one can use the same general logic to list those years when no staff joined:
WITH list_years (year#) AS
(VALUES SMALLINT(0)
UNION ALL
SELECT year# + 1
FROM list_years
WHERE year# < 5)
SELECT year#
FROM list_years y
WHERE NOT EXISTS
(SELECT *
FROM staff s
WHERE UCASE(s.name) LIKE '%E%'
AND s.years = y.year#)
ORDER BY 1;
ANSWER
YEAR# |
---|
0 |
2 |
3 |
23.5.7. Multiple Counts from the Same Row
Imagine that we want to select from the EMPLOYEE table the following counts presented in a tabular list with one line per item. In each case, if nothing matches we want to get a zero:
-
Those with a salary greater than $20,000
-
Those whose first name begins 'ABC%'
-
Those who are male.
-
Employees per department.
-
A count of all rows.
Note that a given row in the EMPLOYEE table may match more than one of the above criteria. If this were not the case, a simple nested table expression could be used. Instead we will do the following:
WITH category (cat, subcat, dept) AS
(VALUES ('1ST', 'ROWS IN TABLE ', '')
, ('2ND', 'SALARY > $20K ', '')
, ('3RD', 'NAME LIKE ABC%', '')
, ('4TH', 'NUMBER MALES ', '')
UNION
SELECT '5TH'
, deptname
, deptno
FROM department)
SELECT xxx.cat AS "category"
, xxx.subcat AS "subcategory/dept"
, SUM(xxx.found) AS "#rows"
FROM (SELECT cat.cat
, cat.subcat
, CASE
WHEN emp.empno IS NULL THEN 0
ELSE 1
END AS found
FROM category cat
LEFT OUTER JOIN employee emp
ON cat.subcat = 'ROWS IN TABLE'
OR (cat.subcat = 'NUMBER MALES'
AND
emp.sex = 'M')
OR (cat.subcat = 'SALARY > $20K'
AND
emp.salary > 20000)
OR (cat.subcat = 'NAME LIKE ABC%'
AND
emp.firstnme LIKE 'ABC%')
OR (cat.dept <> ''
AND
cat.dept = emp.workdept)
) AS xxx
GROUP BY xxx.cat
, xxx.subcat
ORDER BY 1,2;
In the above query, a temporary table is defined and then populated with all of the summation types. This table is then joined (using a left outer join) to the EMPLOYEE table. Any matches (i.e. where EMPNO is not null) are given a FOUND value of 1. The output of the join is then feed into a GROUP BY to get the required counts.
CATEGORY | SUBCATEGORY/DEPT | #ROWS |
---|---|---|
1ST |
ROWS IN TABLE |
32 |
2ND |
SALARY > $20K |
25 |
3RD |
NAME LIKE ABC% |
0 |
4TH |
NUMBER MALES |
19 |
5TH |
ADMINISTRATION SYSTEMS |
6 |
5TH |
DEVELOPMENT CENTER |
0 |
5TH |
INFORMATION CENTER |
3 |
5TH |
MANUFACTURING SYSTEMS |
9 |
5TH |
OPERATIONS |
5 |
5TH |
PLANNING |
1 |
5TH |
SOFTWARE SUPPORT |
4 |
5TH |
SPIFFY COMPUTER SERVICE DIV. |
3 |
5TH |
SUPPORT SERVICES |
1 |
23.6. Normalize Denormalized Data
Imagine that one has a string of text that one wants to break up into individual words. As long as the word delimiter is fairly basic (e.g. a blank space), one can use recursive SQL to do this task. One recursively divides the text into two parts (working from left to right). The first part is the word found, and the second part is the remainder of the text:
WITH temp1 (id, data) AS
(VALUES (01, 'SOME TEXT TO PARSE.')
, (02, 'MORE SAMPLE TEXT.')
, (03, 'ONE-WORD.')
, (04,''))
, temp2 (id, word#, word, data_left) AS
(SELECT id
, SMALLINT(1)
, SUBSTR(data, 1, CASE LOCATE(' ', data)
WHEN 0 THEN LENGTH(data)
ELSE LOCATE(' ', data)
END
)
, LTRIM(SUBSTR(data, CASE LOCATE(' ', data)
WHEN 0 THEN LENGTH(data) + 1
ELSE LOCATE(' ', data)
END
)
)
FROM temp1
WHERE data <> ''
UNION ALL
SELECT id
, word# + 1
, SUBSTR(data_left, 1, CASE LOCATE(' ', data_left)
WHEN 0 THEN LENGTH(data_left)
ELSE LOCATE(' ', data_left)
END
)
, LTRIM(SUBSTR(data_left, CASE LOCATE(' ', data_left)
WHEN 0 THEN LENGTH(data_left) + 1
ELSE LOCATE(' ', data_left)
END
)
)
FROM temp2
WHERE data_left <> '')
SELECT *
FROM temp2
ORDER BY 1,2;
The SUBSTR function is used above to extract both the next word in the string, and the remainder of the text. If there is a blank byte in the string, the SUBSTR stops (or begins, when getting the remainder) at it. If not, it goes to (or begins at) the end of the string. CASE logic is used to decide what to do. .Break text into words
ID | WORD# | WORD | DATA_LEFT |
---|---|---|---|
1 |
1 |
SOME |
TEXT TO PARSE. |
1 |
2 |
TEXT |
TO PARSE. |
1 |
3 |
TO |
PARSE. |
1 |
4 |
PARSE. |
|
2 |
1 |
MORE |
SAMPLE TEXT. |
2 |
2 |
SAMPLE |
TEXT. |
2 |
3 |
TEXT. |
|
3 |
1 |
ONE-WORD. |
23.7. Denormalize Normalized Data
The SUM function can be used to accumulate numeric values. To accumulate character values (i.e. to string the individual values from multiple lines into a single long value) is a little harder, but it can also be done. The following example uses the XMLAGG column function to aggregate multiple values into one. The processing goes as follows:
-
The XMLTEXT scalar function converts each character value into XML. A space is put at the end of the each name, so there is a gap before the next.
-
The XMLAGG column function aggregates the individual XML values in name order.
-
The XMLSERIALIZE scalar function converts the aggregated XML value into a CLOB.
-
The SUBSTR scalar function converts the CLOB to a CHAR.
Now for the code:
SELECT dept
, SMALLINT(COUNT(*)) AS #w
, MAX(name) AS max_name
, SUBSTR(
XMLSERIALIZE(
XMLAGG(
XMLTEXT(name || ' ')
ORDER BY name) AS CLOB(1M))
, 1, 50) AS all_names
FROM staff
GROUP BY dept
ORDER BY dept;
Here is the answer:
DEPT | W# | MAX_NAME | ALL_NAMES |
---|---|---|---|
10 |
4 |
Molinare |
Daniels Jones Lu Molinare |
15 |
4 |
Rothman |
Hanes Kermisch Ngan Rothman |
20 |
4 |
Sneider |
James Pernal Sanders Sneider |
38 |
5 |
Quigley |
Abrahams Marenghi Naughton O’Brien Quigley |
42 |
4 |
Yamaguchi |
Koonitz Plotz Scoutten Yamaguchi |
51 |
5 |
Williams |
Fraye Lundquist Smith Wheeler Williams |
66 |
5 |
Wilson |
Burke Gonzales Graham Lea Wilson |
84 |
4 |
Quill |
Davis Edwards Gafney Quill |
The next example uses recursion to do exactly the same thing. It begins by getting the minimum name in each department. It then recursively gets the next to lowest name, then the next, and so on. As the query progresses, it maintains a count of names added, stores the current name in the temporary NAME field, and appends the same to the end of the ALL_NAMES field. Once all of the names have been processed, the final SELECT eliminates from the answer-set all rows, except the last for each department:
WITH temp1 (dept,w#,name,all_names) AS
(SELECT dept
, SMALLINT(1)
, MIN(name)
, VARCHAR(MIN(name), 50)
FROM staff a
GROUP BY dept
UNION ALL
SELECT a.dept
, SMALLINT(b.w#+1)
, a.name
, b.all_names || ' ' || a.name
FROM staff a
, temp1 b
WHERE a.dept = b.dept
AND a.name > b.name
AND a.name =
(SELECT MIN(c.name)
FROM staff c
WHERE c.dept = b.dept
AND c.name > b.name)
)
SELECT dept
, w#
, name AS max_name
, all_names
FROM temp1 d
WHERE w# = (SELECT MAX(w#)
FROM temp1 e
WHERE d.dept = e.dept)
ORDER BY dept;
If there are no suitable indexes, the above query may be horribly inefficient. If this is the case, one can create a user-defined function to string together the names in a department:
This example uses an "!" as the stmt delimiter |
CREATE FUNCTION list_names(indept SMALLINT)
RETURNS VARCHAR(50)
BEGIN ATOMIC
DECLARE outstr VARCHAR(50) DEFAULT '';
FOR list_names AS
SELECT name
FROM staff
WHERE dept = indept
ORDER BY name
DO
SET outstr = outstr || name || ' ';
END FOR;
SET outstr = rtrim(outstr);
RETURN outstr;
END!
SELECT dept AS DEPT
, SMALLINT(cnt) AS W#
, mxx AS MAX_NAME
, list_names(dept) AS ALL_NAMES
FROM (SELECT dept
, COUNT(*) as cnt
, MAX(name) AS mxx
FROM staff
GROUP BY dept) as ddd
ORDER BY dept!
Even the above might have unsatisfactory performance - if there is no index on department. If adding an index to the STAFF table is not an option, it might be faster to insert all of the rows into a declared temporary table, and then add an index to that.
23.8. Transpose Numeric Data
In this section we will turn rows of numeric data into columns. This cannot be done directly in SQL because the language does not support queries where the output columns are unknown at query start. We will get around this limitation by sending the transposed output to a suitably long VARCHAR field. Imagine that we want to group the data in the STAFF sample table by DEPT and JOB to get the SUM salary for each instance, but not in the usual sense with one output row per DEPT and JOB value. Instead, we want to generate one row per DEPT, with a set of "columns" (in a VARCHAR field) that hold the SUM salary values for each JOB in the department. We will also put column titles on the first line of output. To make the following query simpler, three simple scalar functions will be used to convert data from one type to another:
-
Convert decimal data to character - similar to the one on User-defined functions - convert integer to character.
-
Convert smallint data to character - same as the one on User-defined functions - convert integer to character.
-
Right justify and add leading blanks to character data.
Now for the functions:
CREATE FUNCTION num_to_char(inval SMALLINT)
RETURNS CHAR(06)
RETURN RIGHT(CHAR('',06) CONCAT RTRIM(CHAR(inval)), 06);
CREATE FUNCTION num_to_char(inval DECIMAL(9, 2))
RETURNS CHAR(10)
RETURN RIGHT(CHAR('', 7) CONCAT
RTRIM(CHAR(BIGINT(inval))), 7) CONCAT
'.' CONCAT
SUBSTR(DIGITS(inval), 8, 2);
CREATE FUNCTION right_justify(inval CHAR(5))
RETURNS CHAR(10)
RETURN RIGHT(CHAR('', 10) || RTRIM(inval), 10);
The query consists of lots of little steps that are best explained by describing each temporary table built:
-
DATA_INPUT: This table holds the set of matching rows in the STAFF table, grouped by DEPT and JOB as per a typical query (see Transpose Numeric Data for the contents). This is the only time that we touch the original STAFF table. All subsequent queries directly or indirectly reference this table.
-
JOBS_LIST: The list of distinct jobs in all matching rows. Each job is assigned two rownumbers, one ascending, and one descending.
-
DEPT_LIST: The list of distinct departments in all matching rows.
-
DEPT_JOB_LIST: The list of all matching department/job combinations. We need this table because not all departments have all jobs.
-
DATA_ALL_JOBS: The DEPT_JOB_LIST table joined to the original DATA_INPUT table using a left outer join, so we now have one row with a sum-salary value for every JOB and DEPT instance.
-
DATA_TRANSFORM: Recursively go through the DATA_ALL_JOBS table (for each department), adding the a character representation of the current sum-salary value to the back of a VARCHAR column.
-
DATA_LAST_ROW: For each department, get the row with the highest ascending JOB# value. This row has the concatenated string of sum-salary values.
At this point we are done, except that we don’t have any column headings in our output. The rest of the query gets these.
-
JOBS_TRANSFORM: Recursively go through the list of distinct jobs, building a VARCHAR string of JOB names. The job names are right justified - to match the sumsalary values, and have the same output length.
-
JOBS_LAST_ROW: Get the one row with the lowest descending job number. This row has the complete string of concatenated job names.
-
DATA_AND_JOBS: Use a UNION ALL to vertically combine the JOBS_LAST_ROW and DATA_LAST_ROW tables. The result is a new table with both column titles and sum-salary values.
Finally, we select the list of column names and sum-salary values. The output is ordered so that the column names are on the first line fetched.
Now for the query:
WITH data_input AS
(SELECT dept
, job
, SUM(salary) AS sum_sal
FROM staff
WHERE id < 200
AND name <> 'Sue'
AND salary > 10000
GROUP BY dept
, job)
, jobs_list AS
(SELECT job
, ROW_NUMBER() OVER(ORDER BY job ASC) AS job#A
, ROW_NUMBER() OVER(ORDER BY job DESC) AS job#D
FROM data_input
GROUP BY job)
, dept_list AS
(SELECT dept
FROM data_input
GROUP BY dept)
, dept_jobs_list AS
(SELECT dpt.dept
, job.job
, job.job#A
, job.job#D
FROM jobs_list job
FULL OUTER JOIN dept_list dpt
ON 1 = 1)
, data_all_jobs AS
(SELECT djb.dept
, djb.job
, djb.job#A
, djb.job#D
, COALESCE(dat.sum_sal, 0) AS sum_sal
FROM dept_jobs_list djb
LEFT OUTER JOIN data_input dat
ON djb.dept = dat.dept
AND djb.job = dat.job)
, data_transform (dept, job#A, job#D, outvalue) AS
(SELECT dept
, job#A
, job#D
, VARCHAR(num_to_char(sum_sal), 250)
FROM data_all_jobs
WHERE job#A = 1
UNION ALL
SELECT dat.dept
, dat.job#A
, dat.job#D
, trn.outvalue || ',' || num_to_char(dat.sum_sal)
FROM data_transform trn
, data_all_jobs dat
WHERE trn.dept = dat.dept
AND trn.job#A = dat.job#A - 1)
, data_last_row AS
(SELECT dept
, num_to_char(dept) AS dept_char
, outvalue
FROM data_transform
WHERE job#D = 1)
, jobs_transform (job#A, job#D, outvalue) AS
(SELECT job#A
, job#D
, VARCHAR(right_justify(job), 250)
FROM jobs_list
WHERE job#A = 1
UNION ALL
SELECT job.job#A
, job.job#D
, trn.outvalue || ',' || right_justify(job.job)
FROM jobs_transform trn
, jobs_list job
WHERE trn.job#A = job.job#A - 1)
, jobs_last_row AS
(SELECT 0 AS dept
, ' DEPT' AS dept_char
, outvalue
FROM jobs_transform
WHERE job#D = 1)
, data_and_jobs AS
(SELECT ept
, ept_char
, outvalue
FROM jobs_last_row
UNION ALL
SELECT dept
, dept_char
, outvalue
FROM data_last_row)
SELECT dept_char || ',' || outvalue AS output
FROM data_and_jobs
ORDER BY dept;
For comparison, below are the contents of the first temporary table, and the final output:
DATA_INPUT
DEPT | JOB | SUM_SAL |
---|---|---|
10 |
Mgr |
22959.20 |
15 |
Clerk |
24766.70 |
15 |
Mgr |
20659.80 |
15 |
Sales |
16502.83 |
20 |
Clerk |
27757.35 |
20 |
Mgr |
18357.50 |
20 |
Sales |
78171.25 |
38 |
Clerk |
24964.50 |
38 |
Mgr |
77506.75 |
38 |
Sales |
34814.30 |
42 |
Clerk |
10505.90 |
42 |
Mgr |
18352.80 |
42 |
Sales |
18001.75 |
51 |
Mgr |
21150.00 |
51 |
Sales |
19456.50 |
OUTPUT
DEPT | Clerk | Mgr | Sales |
---|---|---|---|
10 |
0.00 |
22959.20 |
0.00 |
15 |
24766.70 |
20659.80 |
16502.83 |
20 |
27757.35 |
18357.50 |
78171.25 |
38 |
24964.50 |
77506.75 |
34814.30 |
42 |
10505.90 |
18352.80 |
18001.75 |
51 |
0.00 |
21150.00 |
19456.50 |
23.9. Reversing Field Contents
Db2 lacks a simple function for reversing the contents of a data field. Fortunately, we can create a function to do it ourselves.
23.9.1. Input vs. Output
Before we do any data reversing, we have to define what the reversed output should look like relative to a given input value. For example, if we have a four-digit numeric field, the reverse of the number 123 could be 321, or it could be 3210. The latter value implies that the input has a leading zero. It also assumes that we really are working with a four digit field. Likewise, the reverse of the number 123.45 might be 54.321, or 543.21. Another interesting problem involves reversing negative numbers. If the value "-123" is a string, then the reverse is probably "321-". If it is a number, then the desired reverse is more likely to be "-321". Trailing blanks in character strings are a similar problem. Obviously, the reverse of "ABC" is "CBA", but what is the reverse of "ABC "? There is no general technical answer to any of these questions. The correct answer depends upon the business needs of the application. Below is a user defined function that can reverse the contents of a character field:
This example uses an "!" as the stmt delimiter |
--#SET DELIMITER !
CREATE FUNCTION reverse(instr VARCHAR(50))
RETURNS VARCHAR(50)
BEGIN ATOMIC
DECLARE outstr VARCHAR(50) DEFAULT '';
DECLARE curbyte SMALLINT DEFAULT 0;
SET curbyte = LENGTH(RTRIM(instr));
WHILE curbyte >= 1 DO
SET outstr = outstr || SUBSTR(instr,curbyte, 1);
SET curbyte = curbyte - 1;
END WHILE;
RETURN outstr;
END!
SELECT id AS ID
, name AS NAME1
, reverse(name) AS NAME2
FROM staff
WHERE id < 40
ORDER BY id!
ANSWER
ID | NAME1 | NAME2 |
---|---|---|
10 |
Sanders |
srednaS |
20 |
Pernal |
lanreP |
30 |
Marenghi |
ihgneraM |
The same function can be used to reverse numeric values, as long as they are positive:
SELECT id AS ID
, salary AS SALARY1
, DEC(reverse(CHAR(salary)), 7, 4) AS SALARY2
FROM staff
WHERE id < 40
ORDER BY id;
ANSWER
ID | SALARY1 | SALARY2 |
---|---|---|
10 |
18357.50 |
5.7538 |
20 |
78171.25 |
52.1718 |
30 |
77506.75 |
57.6057 |
Simple CASE logic can be used to deal with negative values (i.e. to move the sign to the front of the string, before converting back to numeric), if they exist.
23.10. Fibonacci Series
A Fibonacci Series is a series of numbers where each value is the sum of the previous two. Regardless of the two initial (seed) values, if run for long enough, the division of any two adjacent numbers will give the value 0.618 or inversely 1.618. The following user defined function generates a Fibonacci series using three input values:
-
First seed value.
-
Second seed value.
-
Number values to generate in series.
Observe that that the function code contains a check to stop series generation if there is not enough space in the output field for more numbers:
This example uses an "!" as the stmt delimiter |
--#SET DELIMITER !
CREATE FUNCTION Fibonacci (inval1 INTEGER
, inval2 INTEGER
, loopno INTEGER)
RETURNS VARCHAR(500)
BEGIN ATOMIC
DECLARE loopctr INTEGER DEFAULT 0;
DECLARE tempval1 BIGINT;
DECLARE tempval2 BIGINT;
DECLARE tempval3 BIGINT;
DECLARE outvalue VARCHAR(500);
SET tempval1 = inval1;
SET tempval2 = inval2;
SET outvalue = RTRIM(LTRIM(CHAR(tempval1))) || ', ' ||
RTRIM(LTRIM(CHAR(tempval2)));
calc:
WHILE loopctr < loopno DO
SET tempval3 = tempval1 + tempval2;
SET tempval1 = tempval2;
SET tempval2 = tempval3;
SET outvalue = outvalue || ', ' || RTRIM(LTRIM(CHAR(tempval3)));
SET loopctr = loopctr + 1;
IF LENGTH(outvalue) > 480 THEN
SET outvalue = outvalue || ' etc...';
LEAVE calc;
END IF;
END WHILE;
RETURN outvalue;
END!
The following query references the function:
WITH temp1 (v1, v2, lp) AS
(VALUES (00, 01, 11)
, (12, 61, 10)
, (02, 05, 09)
, (01, -1, 08))
SELECT t1.*
, Fibonacci(v1, v2, lp) AS sequence
FROM temp1 t1;
ANSWER
V1 | V2 | LP | SEQUENCE |
---|---|---|---|
0 |
1 |
11 |
0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144 |
12 |
61 |
10 |
12, 61, 73, 134, 207, 341, 548, 889, 1437, 2326, 3763, 6089 |
2 |
5 |
9 |
2, 5, 7, 12, 19, 31, 50, 81, 131, 212, 343 |
1 |
-1 |
8 |
1, -1, 0, -1, -1, -2, -3, -5, -8, -13 |
The above example generates the complete series of values. If needed, the code could easily be simplified to simply return only the last value in the series. Likewise, a recursive join can be used to create a set of rows that are a Fibonacci series.
23.11. Business Day Calculation
The following function will calculate the number of business days (i.e. Monday to Friday) between to two dates:
This example uses an "!" as the stmt delimiter. |
CREATE FUNCTION business_days (lo_date DATE, hi_date DATE)
RETURNS INTEGER
BEGIN ATOMIC
DECLARE bus_days INTEGER DEFAULT 0;
DECLARE cur_date DATE;
SET cur_date = lo_date;
WHILE cur_date < hi_date DO
IF DAYOFWEEK(cur_date) IN (2,3,4,5,6) THEN
SET bus_days = bus_days + 1;
END IF;
SET cur_date = cur_date + 1 DAY;
END WHILE;
RETURN bus_days;
END!
Below is an example of the function in use:
WITH temp1 (ld, hd) AS
(VALUES (DATE('2006-01-10'), DATE('2007-01-01'))
, (DATE('2007-01-01'), DATE('2007-01-01'))
, (DATE('2007-02-10'), DATE('2007-01-01')))
SELECT t1.*
, DAYS(hd) - DAYS(ld) AS diff
, business_days(ld, hd) AS bdays
FROM temp1 t1;
ANSWER
LD | HD | DIFF | BDAYS |
---|---|---|---|
2006-01-10 |
2007-01-01 |
356 |
254 |
2007-01-01 |
2007-01-01 |
0 |
0 |
2007-02-10 |
2007-01-01 |
-40 |
0 |
23.12. Query Runs for "n" Seconds
Imagine that one wanted some query to take exactly four seconds to run. The following query does just this - by looping (using recursion) until such time as the current system timestamp is four seconds greater than the system timestamp obtained at the beginning of the query:
WITH temp1 (num,ts1,ts2) AS
(VALUES (INT(1)
, TIMESTAMP(GENERATE_UNIQUE())
, TIMESTAMP(GENERATE_UNIQUE()))
UNION ALL
SELECT num + 1
, ts1
, TIMESTAMP(GENERATE_UNIQUE())
FROM temp1
WHERE TIMESTAMPDIFF(2, CHAR(ts2 - ts1)) < 4)
SELECT MAX(num) AS #loops
, MIN(ts2) AS bgn_timestamp
, MAX(ts2) AS end_timestamp
FROM temp1;
ANSWER
#LOOPS | BGN_TIMESTAMP | END_TIMESTAMP |
---|---|---|
58327 |
2001-08-09-22.58.12.754579 |
2001-08-09-22.58.16.754634 |
Observe that the CURRENT TIMESTAMP special register is not used above. It is not appropriate for this situation, because it always returns the same value for each invocation within a single query.
23.13. Function to Pause for "n" Seconds
We can take the above query and convert it into a user-defined function that will loop for "n" seconds, where "n" is the value passed to the function. However, there are several caveats:
-
Looping in SQL is a "really stupid" way to hang around for a couple of seconds. A far better solution would be to call a stored procedure written in an external language that has a true pause command.
-
The number of times that the function is invoked may differ, depending on the access path used to run the query.
-
The recursive looping is going to result in the calling query getting a warning message.
Now for the code:
CREATE FUNCTION pause(inval INT)
RETURNS INTEGER
NOT DETERMINISTIC
EXTERNAL ACTION
RETURN
WITH ttt (num, strt, stop) AS
(VALUES (1
, TIMESTAMP(GENERATE_UNIQUE())
, TIMESTAMP(GENERATE_UNIQUE()))
UNION ALL
SELECT num + 1
, strt
, TIMESTAMP(GENERATE_UNIQUE())
FROM ttt
WHERE TIMESTAMPDIFF(2, CHAR(stop - strt)) < inval)
SELECT MAX(num)
FROM ttt;
Below is a query that calls the above function:
SELECT id
, SUBSTR(CHAR(TIMESTAMP(GENERATE_UNIQUE())),18) AS ss_mmmmmm
, pause(id / 10) AS #loops
, SUBSTR(CHAR(TIMESTAMP(GENERATE_UNIQUE())),18) AS ss_mmmmmm
FROM staff
WHERE id < 31;
ANSWER
ID | SS_MMMMMM | #LOOPS | SS_MMMMMM |
---|---|---|---|
10 |
50.068593 |
76386 |
50.068587 |
20 |
52.068744 |
144089 |
52.068737 |
30 |
55.068930 |
206101 |
55.068923 |
23.14. Sort Character Field Contents
The following user-defined scalar function will sort the contents of a character field in either ascending or descending order. There are two input parameters:
-
The input string: As written, the input can be up to 20 bytes long. To sort longer fields, change the input, output, and OUT-VAL (variable) lengths as desired.
-
The sort order (i.e. 'A' or 'D').
The function uses a very simple, and not very efficient, bubble-sort. In other words, the input string is scanned from left to right, comparing two adjacent characters at a time. If they are not in sequence, they are swapped - and flag indicating this is set on. The scans are repeated until all of the characters in the string are in order:
--#SET DELIMITER !
CREATE FUNCTION sort_char(in_val VARCHAR(20), sort_dir VARCHAR(1))
RETURNS VARCHAR(20)
BEGIN ATOMIC
DECLARE cur_pos SMALLINT;
DECLARE do_sort CHAR(1);
DECLARE out_val VARCHAR(20);
IF UCASE(sort_dir) NOT IN ('A','D') THEN
SIGNAL SQLSTATE '75001'
SET MESSAGE_TEXT = 'Sort order not ''A'' or ''D''';
END IF;
SET out_val = in_val;
SET do_sort = 'Y';
WHILE do_sort = 'Y' DO
SET do_sort = 'N';
SET cur_pos = 1;
WHILE cur_pos < length(in_val) DO
IF (UCASE(sort_dir) = 'A'
AND SUBSTR(out_val, cur_pos+1, 1) < SUBSTR(out_val, cur_pos, 1)
) OR
(UCASE(sort_dir) = 'D'
AND SUBSTR(out_val, cur_pos+1, 1) > SUBSTR(out_val, cur_pos, 1)) THEN
SET do_sort = 'Y';
SET out_val = CASE
WHEN cur_pos = 1
THEN ''
ELSE SUBSTR(out_val, 1, cur_pos-1)
END CONCAT SUBSTR(out_val, cur_pos+1, 1)
CONCAT SUBSTR(out_val, cur_pos , 1)
CONCAT CASE WHEN cur_pos = length(in_val) - 1
THEN ''
ELSE SUBSTR(out_val,cur_pos+2)
END;
END IF;
SET cur_pos = cur_pos + 1;
END WHILE;
END WHILE;
RETURN out_val;
END!
Here is the function in action:
WITH word1 (w#, word_val) AS
(VALUES(1, '12345678')
, (2, 'ABCDEFG')
, (3, 'AaBbCc')
, (4,'abccb')
, (5,'''%#.')
, (6,'bB')
, (7,'a')
, (8,''))
SELECT w#
, word_val
, sort_char(word_val, 'a') sa
, sort_char(word_val, 'D') sd
FROM word1
ORDER BY w#;
ANSWER
W# | WORD_VAL | SA | SD |
---|---|---|---|
1 |
12345678 |
12345678 |
87654321 |
2 |
ABCDEFG |
ABCDEFG |
GFEDCBA |
3 |
AaBbCc |
aAbBcC |
CcBbAa |
4 |
abccb |
abbcc |
ccbba |
5 |
'%#. |
.'#% |
%#'. |
6 |
bB |
bB |
Bb |
7 |
a |
a |
a |
8 |
23.15. Calculating the Median
The median is defined at that value in a series of values where half of the values are higher to it and the other half are lower. The median is a useful number to get when the data has a few very extreme values that skew the average. If there are an odd number of values in the list, then the median value is the one in the middle (e.g. if 7 values, the median value is #4). If there is an even number of matching values, there are two formulas that one can use:
-
The most commonly used definition is that the median equals the sum of the two middle values, divided by two.
-
A less often used definition is that the median is the smaller of the two middle values.
Db2 does not come with a function for calculating the median, but it can be obtained using the ROW_NUMBER function. This function is used to assign a row number to every matching row, and then one searches for the row with the middle row number.
23.15.1. Using Formula #1
Below is some sample code that gets the median SALARY, by JOB, for some set of rows in the STAFF table. Two JOB values are referenced - one with seven matching rows, and one with four. The query logic goes as follows:
-
Get the matching set of rows from the STAFF table, and give each row a row-number, within each JOB value.
-
Using the set of rows retrieved above, get the maximum row-number, per JOB value, then add 1.0, then divide by 2, then add or subtract 0.6. This will give one two values that encompass a single row-number, if an odd number of rows match, and two row-numbers, if an even number of rows match.
-
Finally, join the one row per JOB obtained in step 2 above to the set of rows retrieved in step 1 - by common JOB value, and where the row-number is within the high/low range. The average salary of whatever is retrieved is the median.
Now for the code:
WITH numbered_rows AS
(SELECT s.*
, ROW_NUMBER() OVER(PARTITION BY job
ORDER BY salary, id) AS row#
FROM staff s
WHERE comm > 0
AND name LIKE '%e%')
, median_row_num AS
(SELECT job
, (MAX(row# + 1.0) / 2) - 0.5 AS med_lo
, (MAX(row# + 1.0) / 2) + 0.5 AS med_hi
FROM numbered_rows
GROUP BY job)
SELECT nn.job
, DEC(AVG(nn.salary), 7, 2) AS med_sal
FROM numbered_rows nn
, median_row_num mr
WHERE nn.job = mr.job
AND nn.row# BETWEEN mr.med_lo AND mr.med_hi
GROUP BY nn.job
ORDER BY nn.job;
ANSWER
JOB | MED_SAL |
---|---|
Clerk |
13030.50 |
Sales |
17432.10 |
To get consistent results when using the ROW_NUMBER function, one must ensure that the ORDER BY column list encompasses the unique key of the table. Otherwise the row-number values will be assigned randomly - if there are multiple rows with the same value. In this particular case, the ID has been included in the ORDER BY list, to address duplicate SALARY values. |
The next example is the essentially the same as the prior, but there is additional code that gets the average SALARY, and a count of the number of matching rows per JOB value. Observe that all this extra code went in the second step:
WITH numbered_rows AS
(SELECT s.*
, ROW_NUMBER() OVER(PARTITION BY job
ORDER BY salary, id) AS row#
FROM staff s
WHERE comm > 0
AND name LIKE '%e%')
, median_row_num AS
(SELECT job
, (MAX(row# + 1.0) / 2) - 0.5 AS med_lo
, (MAX(row# + 1.0) / 2) + 0.5 AS med_hi
, DEC(AVG(salary),7,2) AS avg_sal
, COUNT(*) AS #rows
FROM numbered_rows
GROUP BY job)
SELECT nn.job
, DEC(AVG(nn.salary),7,2) AS med_sal
, MAX(mr.avg_sal) AS avg_sal
, MAX(mr.#rows) AS #r
FROM numbered_rows nn
, median_row_num mr
WHERE nn.job = mr.job
AND nn.row# BETWEEN mr.med_lo AND mr.med_hi
GROUP BY nn.job
ORDER BY nn.job;
ANSWER
JOB | MED_SAL | AVG_SAL | #R |
---|---|---|---|
Clerk |
13030.50 |
12857.56 |
7 |
Sales |
17432.10 |
17460.93 |
4 |
23.15.2. Using Formula #2
Once again, the following sample code gets the median SALARY, by JOB, for some set of rows in the STAFF table. Two JOB values are referenced - one with seven matching rows, the other with four. In this case, when there is an even number of matching rows, the smaller of the two middle values is chosen. The logic goes as follows:
-
Get the matching set of rows from the STAFF table, and give each row a row-number, within each JOB value.
-
Using the set of rows retrieved above, get the maximum row-number per JOB, then add 1, then divide by 2. This will get the row-number for the row with the median value.
-
Finally, join the one row per JOB obtained in step 2 above to the set of rows retrieved in step 1 - by common JOB and row-number value.
WITH numbered_rows AS
(SELECT s.*
, ROW_NUMBER() OVER(PARTITION BY job
ORDER BY salary, id) AS row#
FROM staff s
WHERE comm > 0
AND name LIKE '%e%')
, median_row_num AS
(SELECT job
, MAX(row# + 1) / 2 AS med_row#
FROM numbered_rows GROUP BY job)
SELECT nn.job
, nn.salary AS med_sal
FROM numbered_rows nn
, median_row_num mr
WHERE nn.job = mr.job
AND nn.row# = mr.med_row#
ORDER BY nn.job;
ANSWER
JOB | MED_SAL |
---|---|
Clerk |
13030.50 |
Sales |
16858.20 |
The next query is the same as the prior, but it uses a sub-query, instead of creating and then joining to a second temporary table:
WITH numbered_rows AS
(SELECT s.*
, ROW_NUMBER() OVER(PARTITION BY job
ORDER BY salary, id) AS row#
FROM staff s
WHERE comm > 0
AND name LIKE '%e%')
SELECT job
, salary AS med_sal
FROM numbered_rows
WHERE (job, row#) IN
(SELECT job
, MAX(row# + 1) / 2
FROM numbered_rows
GROUP BY job)
ORDER BY job;
ANSWER
JOB | MED_SAL |
---|---|
Clerk |
13030.50 |
Sales |
16858.20 |
The next query lists every matching row in the STAFF table (per JOB), and on each line of output, shows the median salary:
WITH numbered_rows AS
(SELECT s.*
, ROW_NUMBER() OVER(PARTITION BY job
ORDER BY salary, id) AS row#
FROM staff s
WHERE comm > 0
AND name LIKE '%e%')
SELECT r1.*
, (SELECT r2.salary
FROM numbered_rows r2
WHERE r2.job = r1.job
AND r2.row# = (SELECT MAX(r3.row# + 1) / 2
FROM numbered_rows r3
WHERE r2.job = r3.job
)
) AS med_sal
FROM numbered_rows r1
ORDER BY job
, salary;
23.16. Converting HEX Data to Number
The following function accepts as input a hexadecimal representation of an integer value, and returns a BIGINT number. It works for any integer type:
CREATE FUNCTION hex_to_int(input_val VARCHAR(16))
RETURNS BIGINT
BEGIN ATOMIC
DECLARE parse_val VARCHAR(16) DEFAULT '';
DECLARE sign_val BIGINT DEFAULT 1;
DECLARE out_val BIGINT DEFAULT 0;
DECLARE cur_exp BIGINT DEFAULT 1;
DECLARE input_len SMALLINT DEFAULT 0;
DECLARE cur_byte SMALLINT DEFAULT 1;
IF LENGTH(input_val) NOT IN (4,8,16) THEN
SIGNAL SQLSTATE VALUE '70001'
SET MESSAGE_TEXT = 'Length wrong';
END IF;
SET input_len = LENGTH(input_val);
WHILE cur_byte <= input_len DO
SET parse_val = parse_val ||
SUBSTR(input_val,cur_byte + 1,1) ||
SUBSTR(input_val,cur_byte + 0,1);
SET cur_byte = cur_byte + 2;
END WHILE;
IF SUBSTR(parse_val,input_len,1) BETWEEN '8' AND 'F' THEN
SET sign_val = -1;
SET out_val = -1;
SET parse_val = TRANSLATE(parse_val, '0123456789ABCDEF', 'FEDCBA9876543210');
END IF;
SET cur_byte = 1;
WHILE cur_byte <= input_len DO
SET out_val = out_val + (cur_exp *
sign_val *
CASE SUBSTR(parse_val, cur_byte, 1)
WHEN '0' THEN 00
WHEN '1' THEN 01
WHEN '2' THEN 02
WHEN '3' THEN 03
WHEN '4' THEN 04
WHEN '5' THEN 05
WHEN '6' THEN 06
WHEN '7' THEN 07
WHEN '8' THEN 08
WHEN '9' THEN 09
WHEN 'A' THEN 10
WHEN 'B' THEN 11
WHEN 'C' THEN 12
WHEN 'D' THEN 13
WHEN 'E' THEN 14
WHEN 'F' THEN 15
END);
IF cur_byte < input_len THEN
SET cur_exp = cur_exp * 16;
END IF;
SET cur_byte = cur_byte + 1;
END WHILE;
RETURN out_val;
END
23.16.1. Function Logic
The function does the following:
-
Check that the input value is the correct length for an integer value. If not, flag an error.
-
Transpose every second byte in the input value. This is done because the HEX representation of an integer does not show the data as it really is.
-
Check the high-order bit of what is now the last byte. If it is a "1", the value is a negative number, so the processing will be slightly different.
-
Starting with the first byte in the (transposed) input, covert each byte to an integer value using CASE logic. Multiply each digit obtained by the (next) power of sixteen.
-
Return the final result.
23.16.2. Usage Examples
WITH temp1 (num) AS
(VALUES (SMALLINT(+0))
, (SMALLINT(+1))
, (SMALLINT(-1))
, (SMALLINT(+32767))
, (SMALLINT(-32768)))
SELECT num
, HEX(num) AS hex
, hex_to_int(HEX(num)) AS h2i
FROM temp1;
ANSWER
NUM | HEX | H2I |
---|---|---|
0 |
0000 |
0 |
1 |
0100 |
1 |
-1 |
FFFF |
-1 |
32767 |
FF7F |
32767 |
-32768 |
0080 |
-32768 |
...
WITH temp1 (num) AS
(VALUES (INTEGER(0))
UNION ALL
SELECT (num + 1) * 7
FROM temp1
WHERE num < 1E6)
, temp2 (sgn) AS
(VALUES (+1)
, (-13))
, temp3 (num) AS
(SELECT DISTINCT num * sgn
FROM temp1
, temp2)
SELECT num
, HEX(num) AS hex
, hex_to_int(HEX(num)) AS h2i
FROM temp3
ORDER BY num;
....
ANSWER
NUM | HEX | H2I |
---|---|---|
87432800 |
A0E1C9FA |
-87432800 |
-12490387 |
6D6941FF |
-12490387 |
-1784328 |
F8C5E4FF |
-1784328 |
-254891 |
551CFCFF |
-254891 |
-36400 |
D071FFFF |
-36400 |
-5187 |
BDEBFFFF |
-5187 |
-728 |
28FDFFFF |
-728 |
-91 |
A5FFFFFF |
-91 |
0 |
00000000 |
0 |
7 |
07000000 |
7 |
56 |
38000000 |
56 |
399 |
8F010000 |
399 |
2800 |
F00A0000 |
2800 |
19607 |
974C0000 |
19607 |
137256 |
28180200 |
137256 |
960799 |
1FA90E00 |
960799 |
6725600 |
E09F6600 |
6725600 |
Usage Notes
-
The above function won’t work on the mainframe because the internal representation of an integer value is different (see below). The modifications required to make it work are minor.
-
The above function won’t work on the HEX representation of packed-decimal or floatingpoint data.
-
One could have three different flavors of the above function - one for each type of integer value. The input value length would determine the output type.
23.17. Endianness
Most computers use one of two internal formats to store binary data. In big-endian, which is used on z/OS, the internal representation equals the HEX value. So the four-byte integer value 1,234,567,890 is stored as "49 96 02 D2". In little-endian, which is used on all Intel chips, the bytes are reversed, so the above value is stored internally as "D2 02 96 49". This is why the above function transposed every two-byte block before converting the HEX value to numeric.
24. Quirks in SQL
One might have noticed by now that not all SQL statements are easy to comprehend. Unfortunately, the situation is perhaps a little worse than you think. In this section we will discuss some SQL statements that are correct, but which act just a little funny.
24.1. Trouble with Timestamps
When does one timestamp not equal another with the same value? The answer is, when one value uses a 24 hour notation to represent midnight and the other does not. To illustrate, the following two timestamp values represent the same point in time, but not according to Db2:
WITH temp1 (c1, t1, t2) AS
(VALUES('A', TIMESTAMP('1996-05-01-24.00.00.000000'), TIMESTAMP('1996-05-02-00.00.00.000000')))
SELECT c1
FROM temp1
WHERE t1 = t2;
ANSWER: No rows
To make Db2 think that both timestamps are actually equal (which they are), all we have to do is fiddle around with them a bit:
WITH temp1 (c1,t1,t2)
AS (VALUES('A', TIMESTAMP('1996-05-01-24.00.00.000000'), TIMESTAMP('1996-05-02-00.00.00.000000')))
SELECT c1
FROM temp1
WHERE t1 + 0 MICROSECOND = t2 + 0 MICROSECOND;
ANSWER
C1 |
---|
A |
Be aware that, as with everything else in this section, what is shown above is not a bug. It is the way that it is because it makes perfect sense, even if it is not intuitive.
24.2. Using 24 Hour Notation
One might have to use the 24-hour notation, if one needs to record (in Db2) external actions that happen just before midnight - with the correct date value. To illustrate, imagine that we have the following table, which records supermarket sales:
CREATE TABLE supermarket_sales
( sales_ts TIMESTAMP NOT NULL
, sales_val DECIMAL(8, 2) NOT NULL
, PRIMARY KEY(sales_ts));
In this application, anything that happens before midnight, no matter how close, is deemed to have happened on the specified day. So if a transaction comes in with a timestamp value that is a tiny fraction of a microsecond before midnight, we should record it thus:
INSERT INTO supermarket_sales
VALUES ('2003-08-01-24.00.00.000000', 123.45);
Now, if we want to select all of the rows that are for a given day, we can write this:
SELECT *
FROM supermarket_sales
WHERE DATE(sales_ts) = '2003-08-01'
ORDER BY sales_ts;
Or this:
SELECT *
FROM supermarket_sales
WHERE sales_ts BETWEEN '2003-08-01-00.00.00' AND '2003-08-01-24.00.00'
ORDER BY sales_ts;
Db2 will never internally generate a timestamp value that uses the 24 hour notation. But it is provided so that you can use it yourself, if you need to.
24.3. No Rows Match
How many rows are returned by a query when no rows match the provided predicates? The answer is that sometimes you get none, and sometimes you get one:
SELECT creator
FROM sysibm.systables
WHERE creator = 'ZZZ';
ANSWER: no row
SELECT MAX(creator)
FROM sysibm.systables
WHERE creator = 'ZZZ';
ANSWER: Null
SELECT MAX(creator)
FROM sysibm.systables
WHERE creator = 'ZZZ'
HAVING MAX(creator) IS NOT NULL;
ANSWER: no row
SELECT MAX(creator)
FROM sysibm.systables
WHERE creator = 'ZZZ'
HAVING MAX(creator) = 'ZZZ';
ANSWER: no row
SELECT MAX(creator)
FROM sysibm.systables
WHERE creator = 'ZZZ'
GROUP BY creator;
ANSWER: no row
SELECT creator
FROM sysibm.systables
WHERE creator = 'ZZZ'
GROUP BY creator;
ANSWER: no row
SELECT COUNT(*)
FROM sysibm.systables
WHERE creator = 'ZZZ'
GROUP BY creator;
ANSWER: no row
SELECT COUNT(*)
FROM sysibm.systables
WHERE creator = 'ZZZ';
ANSWER: 0
There is a pattern to the above, and it goes thus:
-
When there is no column function (e.g. MAX, COUNT) in the SELECT then, if there are no matching rows, no row is returned.
-
If there is a column function in the SELECT, but nothing else, then the query will always return a row - with zero if the function is a COUNT, and null if it is something else.
-
If there is a column function in the SELECT, and also a HAVING phrase in the query, a row will only be returned if the HAVING predicate is true.
-
If there is a column function in the SELECT, and also a GROUP BY phrase in the query, a row will only be returned if there was one that matched.
Imagine that one wants to retrieve a list of names from the STAFF table, but when no names match, one wants to get a row/column with the phrase "NO NAMES", rather than zero rows. The next query does this by first generating a "not found" row using the SYSDUMMY1 table, and then left-outer-joining to the set of matching rows in the STAFF table. The COALESCE function will return the STAFF data, if there is any, else the not-found data:
SELECT COALESCE(name,noname) AS nme
, COALESCE(salary,nosal) AS sal
FROM (SELECT 'NO NAME' AS noname
, 0 AS nosal
FROM sysibm.sysdummy1) AS nnn
LEFT OUTER JOIN
(SELECT *
FROM staff
WHERE id < 5) AS xxx
ON 1 = 1
ORDER BY name;
ANSWER
NME | SAL |
---|---|
NO NAME |
0.00 |
The next query is logically the same as the prior, but it uses the WITH phrase to generate the "not found" row in the SQL statement:
WITH nnn (noname, nosal) AS
(VALUES ('NO NAME', 0))
SELECT COALESCE(name, noname) AS nme
, COALESCE(salary,nosal) AS sal
FROM nnn
LEFT OUTER JOIN
(SELECT *
FROM staff
WHERE id < 5) AS xxx
ON 1 = 1
ORDER BY NAME;
ANSWER
NME | SAL |
---|---|
NO NAME |
0.00 |
24.4. Dumb Date Usage
Imagine that you have some character value that you convert to a Db2 date. The correct way to do it is given below:
SELECT DATE('2001-09-22')
FROM sysibm.sysdummy1;
ANSWER: 2001-09-22
What happens if you accidentally leave out the quotes in the DATE function? The function still works, but the result is not correct:
SELECT DATE(2001-09-22)
FROM sysibm.sysdummy1;
ANSWER: 0006-05-24
Why the 2,000 year difference in the above results? When the DATE function gets a character string as input, it assumes that it is valid character representation of a Db2 date, and converts it accordingly. By contrast, when the input is numeric, the function assumes that it represents the number of days minus one from the start of the current era (i.e. 0001-01-01). In the above query the input was 2001-09-22, which equals (2001-9)-22, which equals 1970 days.
24.5. RAND in Predicate
The following query was written with intentions of getting a single random row out of the matching set in the STAFF table. Unfortunately, it returned two rows:
SELECT id
, name
FROM staff
WHERE id <= 100
AND id = (INT(RAND()* 10) * 10) + 10
ORDER BY id;
ANSWER
ID | NAME |
---|---|
30 |
Marenghi |
60 |
Quigley |
The above SQL returned more than one row because the RAND function was reevaluated for each matching row. Thus the RAND predicate was being dynamically altered as rows were being fetched. To illustrate what is going on above, consider the following query. The results of the RAND function are displayed in the output. Observe that there are multiple rows where the function output (suitably massaged) matched the ID value. In theory, anywhere between zero and all rows could match:
WITH temp AS
(SELECT id
, name
, (INT(RAND(0)* 10) * 10) + 10 AS ran
FROM staff
WHERE id <= 100)
SELECT t.*
, CASE id
WHEN ran THEN 'Y'
ELSE ' '
END AS eql
FROM temp t
ORDER BY id;
ANSWER
ID | NAME | RAN | EQL |
---|---|---|---|
10 |
Sanders |
10 |
Y |
20 |
Pernal |
30 |
|
30 |
Marenghi |
70 |
|
40 |
O’Brien |
10 |
|
50 |
Hanes |
30 |
|
60 |
Quigley |
40 |
|
70 |
Rothman |
30 |
|
80 |
James |
100 |
|
90 |
Koonitz |
40 |
|
100 |
Plotz |
100 |
Y |
To randomly select some fraction of the rows in a table efficiently and consistently, use the TABLESAMPLE feature. See Randomly Sample Data for more details. |
24.6. Getting "n" Random Rows
There are several ways to always get exactly "n" random rows from a set of matching rows. In the following example, three rows are required:
WITH staff_numbered AS
(SELECT s.*
, ROW_NUMBER() OVER() AS row#
FROM staff s
WHERE id <= 100)
, count_rows AS
(SELECT MAX(row#) AS #rows
FROM staff_numbered)
, random_values (RAN#) AS
(VALUES (RAND())
, (RAND())
, (RAND()))
, rows_t0_get AS
(SELECT INT(ran# * #rows) + 1 AS get_row
FROM random_values
, count_rows)
SELECT id
, name
FROM staff_numbered
, rows_t0_get
WHERE row# = get_row
ORDER BY id;
ANSWER
ID | NAME |
---|---|
10 |
Sanders |
20 |
Pernal |
90 |
Koonitz |
The above query works as follows:
-
First, the matching rows in the STAFF table are assigned a row number.
-
Second, a count of the total number of matching rows is obtained.
-
Third, a temporary table with three random values is generated.
-
Fourth, the three random values are joined to the row-count value, resulting in three new row-number values (of type integer) within the correct range.
-
Finally, the three row-number values are joined to the original temporary table.
There are some problems with the above query:
-
If more than a small number of random rows are required, the random values cannot be defined using the VALUES phrase. Some recursive code can do the job.
-
In the extremely unlikely event that the RAND function returns the value "one", no row will match. CASE logic can be used to address this issue.
-
Ignoring the problem just mentioned, the above query will always return three rows, but the rows may not be different rows. Depending on what the three RAND calls generate, the query may even return just one row - repeated three times.
In contrast to the above query, the following will always return three different random rows:
SELECT id
, name
FROM (SELECT s2.*
, ROW_NUMBER() OVER(ORDER BY r1) AS r2
FROM (SELECT s1.*
, RAND() AS r1
FROM staff s1
WHERE id <= 100) AS s2
) as s3
WHERE r2 <= 3
ORDER BY id;
ANSWER
ID | NAME |
---|---|
10 |
Sanders |
40 |
O’Brien |
60 |
Quigley |
In this query, the matching rows are first numbered in random order, and then the three rows with the lowest row number are selected.
24.7. Summary of Issues
The lesson to be learnt here is that one must consider exactly how random one wants to be when one goes searching for a set of random rows:
-
Does one want the number of rows returned to be also somewhat random?
-
Does one want exactly "n" rows, but it is OK to get the same row twice?
-
Does one want exactly "n" distinct (i.e. different) random rows?
24.8. Date/Time Manipulation
I once had a table that contained two fields - the timestamp when an event began, and the elapsed time of the event. To get the end-time of the event, I added the elapsed time to the begin-timestamp - as in the following SQL:
WITH temp1 (bgn_tstamp, elp_sec) AS
(VALUES (TIMESTAMP('2001-01-15-01.02.03.000000'), 1.234)
, (TIMESTAMP('2001-01-15-01.02.03.123456'), 1.234))
SELECT bgn_tstamp
, elp_sec
, bgn_tstamp + elp_sec SECONDS AS end_tstamp
FROM temp1;
ANSWER
BGN_TSTAMP | ELP_SEC | END_TSTAMP |
---|---|---|
2001-01-15-01.02.03.000000 |
1.234 |
2001-01-15-01.02.04.000000 |
2001-01-15-01.02.03.123456 |
1.234 |
2001-01-15-01.02.04.123456 |
As you can see, my end-time is incorrect. In particular, the factional part of the elapsed time has not been used in the addition. I subsequently found out that Db2 never uses the fractional part of a number in date/time calculations. So to get the right answer I multiplied my elapsed time by one million and added microseconds:
WITH temp1 (bgn_tstamp, elp_sec) AS
(VALUES (TIMESTAMP('2001-01-15-01.02.03.000000'), 1.234)
, (TIMESTAMP('2001-01-15-01.02.03.123456'), 1.234))
SELECT bgn_tstamp
, elp_sec
, bgn_tstamp + (elp_sec * 1E6) MICROSECONDS AS end_tstamp
FROM temp1;
ANSWER
BGN_TSTAMP | ELP_SEC | END_TSTAMP |
---|---|---|
2001-01-15-01.02.03.000000 |
1.234 |
2001-01-15-01.02.04.234000 |
2001-01-15-01.02.03.123456 |
1.234 |
2001-01-15-01.02.04.357456 |
Db2 doesn’t use the fractional part of a number in date/time calculations because such a value often makes no sense. For example, 3.3 months or 2.2 years are meaningless values - given that neither a month nor a year has a fixed length.
24.8.1. The Solution
When one has a fractional date/time value (e.g. 5.1 days, 4.2 hours, or 3.1 seconds) that is for a period of fixed length that one wants to use in a date/time calculation, one has to convert the value into some whole number of a more precise time period. For example:
-
5.1 days times 86,400 returns the equivalent number of seconds.
-
6.2 seconds times 1,000,000 returns the equivalent number of microseconds.
24.9. Use of LIKE on VARCHAR
Sometimes one value can be EQUAL to another, but is not LIKE the same. To illustrate, the following SQL refers to two fields of interest, one CHAR, and the other VARCHAR. Observe below that both rows in these two fields are seemingly equal:
WITH temp1 (c0, c1, v1) AS
(VALUES('A', CHAR(' ', 1), VARCHAR(' ', 1))
, ('B', CHAR(' ', 1), VARCHAR('' , 1)))
SELECT c0
FROM temp1
WHERE c1 = v1
AND c1 LIKE ' ';
ANSWER
C0 |
---|
A |
B |
Look what happens when we change the final predicate from matching on C1 to V1. Now only one row matches our search criteria.
WITH temp1 (c0, c1, v1) AS
(VALUES ('A',CHAR(' ', 1),VARCHAR(' ', 1))
, ('B',CHAR(' ', 1),VARCHAR('' , 1)))
SELECT c0
FROM temp1
WHERE c1 = v1
AND v1 LIKE ' ';
ANSWER
C0 |
---|
A |
To explain, observe that one of the VARCHAR rows above has one blank byte, while the other has no data. When an EQUAL check is done on a VARCHAR field, the value is padded with blanks (if needed) before the match. This is why C1 equals C2 for both rows. However, the LIKE check does not pad VARCHAR fields with blanks. So the LIKE test in the second SQL statement only matched on one row. The RTRIM function can be used to remove all trailing blanks and so get around this problem:
WITH temp1 (c0,c1,v1) AS
(VALUES ('A',CHAR(' ',1),VARCHAR(' ',1))
, ('B',CHAR(' ',1),VARCHAR('' ,1)))
SELECT c0
FROM temp1
WHERE c1 = v1
AND RTRIM(v1) LIKE '';
ANSWER
C0 |
---|
A |
B |
24.10. Comparing Weeks
One often wants to compare what happened in part of one year against the same period in another year. For example, one might compare January sales over a decade period. This may be a perfectly valid thing to do when comparing whole months, but it rarely makes sense when comparing weeks or individual days. The problem with comparing weeks from one year to the next is that the same week (as defined by Db2) rarely encompasses the same set of days. The following query illustrates this point by showing the set of days that make up week 33 over a ten-year period. Observe that some years have almost no overlap with the next:
WITH temp1 (yymmdd) AS
(VALUES DATE('2000-01-01')
UNION ALL
SELECT yymmdd + 1 DAY
FROM temp1
WHERE yymmdd < '2010-12-31')
SELECT yy AS year
, CHAR(MIN(yymmdd), ISO) AS min_dt
, CHAR(MAX(yymmdd), ISO) AS max_dt
FROM (SELECT yymmdd
, YEAR(yymmdd) yy
, WEEK(yymmdd) wk
FROM temp1
WHERE WEEK(yymmdd) = 33) AS xxx
GROUP BY yy
, wk;
ANSWER
YEAR | MIN_DT | MAX_DT |
---|---|---|
2000 |
2000-08-06 |
2000-08-12 |
2001 |
2001-08-12 |
2001-08-18 |
2002 |
2002-08-11 |
2002-08-17 |
2003 |
2003-08-10 |
2003-08-16 |
2004 |
2004-08-08 |
2004-08-14 |
2005 |
2005-08-07 |
2005-08-13 |
2006 |
2006-08-13 |
2006-08-19 |
2007 |
2007-08-12 |
2007-08-18 |
2008 |
2008-08-10 |
2008-08-16 |
2009 |
2009-08-09 |
2009-08-15 |
2010 |
2010-08-08 |
2010-08-14 |
24.11. Db2 Truncates, not Rounds
When converting from one numeric type to another where there is a loss of precision, Db2 always truncates not rounds. For this reason, the S1 result below is not equal to the S2 result:
SELECT SUM(INTEGER(salary)) AS s1
, INTEGER(SUM(salary)) AS s2
FROM staff;
ANSWER
S1 | S2 |
---|---|
583633 |
583647 |
If one must do scalar conversions before the column function, use the ROUND function to improve the accuracy of the result:
SELECT SUM(INTEGER(ROUND(salary, -1))) AS s1
, INTEGER(SUM(salary)) AS s2
FROM staff;
ANSWER
S1 | S2 |
---|---|
583640 |
583647 |
24.12. CASE Checks in Wrong Sequence
The case WHEN checks are processed in the order that they are found. The first one that matches is the one used. To illustrate, the following statement will always return the value 'FEM' in the SXX field:
SELECT lastname
, sex
, CASE
WHEN sex >= 'F' THEN 'FEM'
WHEN sex >= 'M' THEN 'MAL'
END AS sxx
FROM employee
WHERE lastname LIKE 'J%'
ORDER BY 1;
ANSWER
LASTNAME | SX | SXX |
---|---|---|
JEFFERSON |
M |
FEM |
JOHNSON |
F |
FEM |
JONES |
M |
FEM |
By contrast, in the next statement, the SXX value will reflect the related SEX value:
SELECT lastname
, sex
, CASE
WHEN sex >= 'M' THEN 'MAL'
WHEN sex >= 'F' THEN 'FEM'
END AS sxx
FROM employee
WHERE lastname LIKE 'J%'
ORDER BY 1;
ANSWER
LASTNAME | SX | SXX |
---|---|---|
JEFFERSON |
M |
MAL |
JOHNSON |
F |
FEM |
JONES |
M |
MAL |
24.13. Division and Average
The following statement gets two results, which is correct?
SELECT AVG(salary) / AVG(comm) AS a1
, AVG(salary / comm) AS a2
FROM staff;
ANSWER
A1 | A2 |
---|---|
-32 |
61.98 |
Arguably, either answer could be correct - depending upon what the user wants. In practice, the first answer is almost always what they intended. The second answer is somewhat flawed because it gives no weighting to the absolute size of the values in each row (i.e. a big SALARY divided by a big COMM is the same as a small divided by a small).
24.14. Date Output Order
Db2 has a bind option (called DATETIME) that specifies the default output format of datetime data. This bind option has no impact on the sequence with which date-time data is presented. It simply defines the output template used. To illustrate, the plan that was used to run the following SQL defaults to the USA date-time-format bind option. Observe that the month is the first field printed, but the rows are sequenced by year:
SELECT hiredate
FROM employee
WHERE hiredate < '1960-01-01'
ORDER BY 1;
ANSWER
HIREDATE |
---|
1947-05-05 |
1949-08-17 |
1958-05-16 |
When the CHAR function is used to convert the date-time value into a character value, the sort order is now a function of the display sequence, not the internal date-time order:
SELECT CHAR(hiredate, USA)
FROM employee
WHERE hiredate < '1960-01-01'
ORDER BY 1;
ANSWER
HIREDATE |
---|
05/05/1947 |
05/16/1958 |
08/17/1949 |
In general, always bind plans so that date-time values are displayed in the preferred format. Using the CHAR function to change the format can be unwise.
24.15. Ambiguous Cursors
The following pseudo-code will fetch all of the rows in the STAFF table (which has ID’s ranging from 10 to 350) and, then while still fetching, insert new rows into the same STAFF table that are the same as those already there, but with ID’s that are 500 larger.
EXEC-SQL
DECLARE fred CURSOR FOR
SELECT *
FROM staff
WHERE id < 1000
ORDER BY id;
END-EXEC;
EXEC-SQL
OPEN fred
END-EXEC;
DO UNTIL SQLCODE = 100;
EXEC-SQL
FETCH fred INTO :HOST-VARS
END-EXEC;
IF SQLCODE <> 100 THEN DO;
SET HOST-VAR.ID = HOST-VAR.ID + 500;
EXEC-SQL
INSERT INTO staff VALUES (:HOST-VARS)
END-EXEC;
END-DO;
END-DO;
EXEC-SQL
CLOSE fred
END-EXEC;
We want to know how many rows will be fetched, and so inserted? The answer is that it depends upon the indexes available. If there is an index on ID, and the cursor uses that index for the ORDER BY, there will 70 rows fetched and inserted. If the ORDER BY is done using a row sort (i.e. at OPEN CURSOR time) only 35 rows will be fetched and inserted.
Be aware that Db2, unlike some other database products, does NOT (always) retrieve all of the matching rows at OPEN CURSOR time. Furthermore, understand that this is a good thing for it means that Db2 (usually) does not process any row that you do not need. Db2 is very good at always returning the same answer, regardless of the access path used. It is equally good at giving consistent results when the same logical statement is written in a different manner (e.g. A=B vs. B=A). What it has never done consistently (and never will) is guarantee that concurrent read and write statements (being run by the same user) will always give the same results.
24.16. Multiple User Interactions
There was once a mythical company that wrote a query to list all orders in the ORDER table for a particular DATE, with the output sequenced by REGION and STATUS. To make the query fly, there was a secondary index on the DATE, REGION, and STATUS columns, in addition to the primary unique index on the ORDER-NUMBER column:
SELECT region_code AS region
, order_status AS status
, order_number AS order#
, order_value AS value
FROM order_table
WHERE order_date = '2006-03-12'
ORDER BY region_code
, order_status
WITH CS;
When the users ran the above query, they found that some orders were seemingly listed twice:
REGION | STATUS | ORDER# | VALUE |
---|---|---|---|
EAST |
PAID |
111 |
4.66 (_) |
EAST |
PAID |
222 |
6.33 |
EAST |
PAID |
333 |
123.45 |
EAST |
SHIPPED |
111 |
4.66 (_) |
EAST |
SHIPPED |
444 |
123.45 |
(*) Same ORDER#
While the above query was running (i.e. traversing the secondary index) another user had come along and updated the STATUS for ORDER# 111 from PAID to SHIPPED, and then committed the change. This update moved the pointer for the row down the secondary index, so that the query subsequently fetched the same row twice.
24.16.1. Explanation
In the above query, Db2 is working exactly as intended. Because the result may seem a little odd, a simple example will be used to explain what is going on:
Imagine that one wants to count the number of cars parked on a busy street by walking down the road from one end to the other, counting each parked car as you walk past. By the time you get to the end of the street, you will have a number, but that number will not represent the number of cars parked on the street at any point in time. And if a car that you counted at the start of the street was moved to the end of the street while you were walking, you will have counted that particular car twice. Likewise, a car that was moved from the end of the street to the start of the street while you were walking in the middle of the street would not have been counted by you, even though it never left the street during your walk.
One way to get a true count of cars on the street is to prevent car movement while you do your walk. This can be unpopular, but it works. The same can be done in Db2 by changing the WITH phrase (i.e. isolation level) at the bottom of the above query:
24.16.2. WITH RR - Repeatable Read
A query defined with repeatable read can be run multiple times and will always return the same result, with the following qualifications:
-
References to special registers, like CURRENT TIMESTAMP, may differ.
-
Rows changed by the user will show in the query results.
No row will ever be seen twice with this solution, because once a row is read it cannot be changed. And the query result is a valid representation of the state of the table, or at least of the matching rows, as of when the query finished. In the car-counting analogy described above, this solution is akin to locking down sections of the street as you walk past, regardless of whether there is a car parked there or not. As long as you do not move a car yourself, each traverse of the street will always get the same count, and no car will ever be counted more than once. In many cases, defining a query with repeatable read will block all changes by other users to the target table for the duration. In theory, rows can be changed if they are outside the range of the query predicates, but this is not always true. In the case of the order system described above, it was not possible to use this solution because orders were coming in all the time.
24.16.3. WITH RS - Read Stability
A query defined with read-stability can be run multiple times, and each row processed previously will always look the same the next time that the query is run - with the qualifications listed above. But rows can be inserted into the table that match the query predicates. These will show in the next run. No row will ever be inadvertently read twice. In our car-counting analogy, this solution is akin to putting a wheel-lock on each parked car as you walk past. The car can’t move, ut new cars can be parked in the street while you are counting. The new cars can also leave subsequently, as you long as you don’t lock them in your next walk down the street. No car will ever be counted more than once in a single pass, but nor will your count ever represent the true state of the street. As with repeatable read, defining a query with read stability will often block all updates by other users to the target table for the duration. It is not a great way to win friends.
24.16.4. WITH CS - Cursor Stability
A query defined with cursor stability will read every committed matching row, occasionally more than once. If the query is run multiple times, it may get a different result each time. In our car-counting analogy, this solution is akin to putting a wheel-lock on each parked car as you count it, but then removing the lock as soon as you move on to the next car. A car that you are not currently counting can be moved anywhere in the street, including to where you have yet to count. In the latter case, you will count it again. This is what happened during our mythical query of the ORDER table. Queries defined with cursor stability still need to take locks, and thus can be delayed if another user has updated a matching row, but not yet done a commit. In extreme cases, the query may get a timeout or deadlock.
24.16.5. WITH UR - Uncommitted Read
A query defined with uncommitted read will read every matching row, including those that have not yet been committed. Rows may occasionally be read more than once. If the query is run multiple times, it may get a different result each time. In our car-counting analogy, this solution is akin to counting each stationary car as one walks past, regardless of whether or not the car is permanently parked. Queries defined with uncommitted read do not take locks, and thus are not delayed by other users who have changed rows, but not yet committed. But some of the rows read may be subsequently rolled back, and so were never valid. Below is a summary of the above options:
SAME RESULT FETCH SAME UNCOMMITTED ROWS CURSOR "WITH" OPTION IF RUN TWICE ROW > ONCE ROWS SEEN LOCKED ===================== ============ ========== =========== ======== RR - Repeatable Read Yes Never Never Many/All RS - Read Stability No (inserts) Never Never Many/All CS - Cusor Stability No (all DML) Maybe Never Current UR - Uncommitted Read No (all DML) Maybe Yes None
24.17. Check for Changes, Using Trigger
The target table can have a column of type timestamp that is set to the current timestamp value (using triggers) every time a row is inserted or updated. The query scanning the table can have a predicate (see below) so it only fetches those rows that were updated before the current timestamp, which is the time when the query was opened:
SELECT region_code AS region
, order_status AS status
, order_number AS order#
, order_value AS value
FROM order_table
WHERE order_date = '2006-03-12'
AND update_ts < CURRENT TIMESTAMP -- <= New predicate
ORDER BY region_code
, order_status
WITH CS;
This solution is almost certainly going to do the job, but it is not quite perfect. There is a very small chance that one can still fetch the same row twice. To illustrate, imagine the following admittedly very improbable sequence of events:
#1 UPDATE statement begins (will run for a long time). #2 QUERY begins (will also run for a long time). #3 QUERY fetches target row (via secondary index). #4 QUERY moves on to the next row, etc... #5 UPDATE changes target row - moves it down index. #6 UPDATE statement finishes, and commits. #7 QUERY fetches target row again (bother).
24.18. Check for Changes, Using Generated TS
A similar solution that will not suffer from the above problem involves adding a timestamp column to the table that is defined GENERATED ALWAYS. This column will be assigned the latest timestamp value (sort of) every time a row is inserted or updated – on a row-by-row basis. Below is an example of a table with this column type:
CREATE TABLE order_table
( order# SMALLINT NOT NULL
, order_date DATE NOT NULL
, order_status CHAR(1) NOT NULL
, order_value DEC(7, 2) NOT NULL
, order_rct TIMESTAMP NOT NULL
GENERATED ALWAYS FOR EACH ROW ON UPDATE
AS ROW CHANGE TIMESTAMP
, PRIMARY KEY (order#));
A query accessing this table that wants to ensure that it does not select the same row twice will include a predicate to check that the order_rct column value is less than or equal to the current timestamp:
SELECT region_code AS region
, order_status AS status
, order_number AS order#
, order_value AS value
FROM order_table
WHERE order_date = '2006-03-12'
AND order_rct <= CURRENT TIMESTAMP -- <= New predicate
ORDER BY region_code
, order_status
WITH CS;
There is just one minor problem with this solution: The generated timestamp value is not always exactly the current timestamp. Sometimes it is every so slightly higher. If this occurs, the above query will not retrieve the affected rows. This problem only occurs during a multi-row insert or update. The generated timestamp value is always unique. To enforce uniqueness, the first row (in a multi-row insert or update) gets the current timestamp special register value. Subsequent rows get the same value, plus "n" microseconds, where "n" incremented by one for each row changed. To illustrate this problem, consider the following statement, which inserts three rows into the above table, but only returns one row- because only the first row inserted has an order_rct value that is equal to or less than the current timestamp special register:
SELECT order#
FROM FINAL TABLE
(INSERT INTO order_table (order#, order_date, order_status, order_value)
VALUES (1, '2007-11-22', 'A', 123.45)
, (2,'2007-11-22','A',123.99)
, (3,'2007-11-22','A',123.99))
WHERE order_rct <= CURRENT TIMESTAMP;
ANSWER
order# |
---|
1 |
The same problem can occur when a query is run immediately after the above insert (i.e. before a commit is done). Occasionally, but by no means always, this query will be use the same current timestamp special register value as the previous insert. If this happens, only the first row inserted will show. NOTE: This problem arises in Db2 running on Windows, which has a somewhat imprecise current timestamp value. It should not occur in environments where Db2 references a system clock with microsecond, or sub-microsecond precision.
24.19. Other Solutions - Good and Bad
Below are some alternatives to the above:
-
Lock Table: If one wanted to see the state of the table as it was at the start of the query, one could use a LOCK TABLE command - in share or exclusive mode. Doing this may not win you many friends with other users.
-
Drop Secondary Indexes: The problem described above does not occur if one accesses the table using a tablespace scan, or via the primary index. However, if the table is large, secondary indexes will probably be needed to get the job done.
-
Two-part Query: One can do the query in two parts: First get a list of DISTINCT primary key values, then join back to the original table using the primary unique index to get the rest of the row:
SELECT region_code AS region
, order_status AS status
, order_number AS order#
, order_value AS value
FROM (SELECT DISTINCT order_number AS distinct_order#
FROM order_table
WHERE order_date = '2006-03-12' ) AS xxx
, order_table
WHERE order_number = distinct_order#
ORDER BY region_code
, order_status
WITH CS;
This solution will do the job, but it is probably going to take about twice as long to complete as the original query.
-
Use Versions: See the chapter titled "Retaining a Record" for a schema that uses lots of complex triggers and views, and that lets one see consistent views of the rows in the table as of any point in time.
24.20. What Time is It
The CURRENT TIMESTAMP special register returns the current time – in local time. There are two other ways to get the something similar the current timestamp. This section discusses the differences:
-
Current Timestamp Special Register: As its name implies, this special register returns the current timestamp. The value will be the same for all references within a single SQL statement, and possibly between SQL statements and/or between users.
-
Generate Unique Scalar Function: With a bit of fudging, this scalar function will return a timestamp value that is unique for every invocation. The value will be close to the current timestamp, but may be a few seconds behind.
-
Generate Always Column Type: This timestamp value will be unique (within a table) for every row changed. In a multi-row insert or update, the first row changed will get the current timestamp. Subsequent rows get the same value, plus "n" microseconds, where "n" incremented by one for each row changed.
The following table will hold the above three values:
CREATE TABLE test_table
( test# SMALLINT NOT NULL
, current_ts TIMESTAMP NOT NULL
, generate_u TIMESTAMP NOT NULL
, generate_a TIMESTAMP NOT NULL
GENERATED ALWAYS FOR EACH ROW ON UPDATE
AS ROW CHANGE TIMESTAMP);
The next statement will insert four rows into the above table:
INSERT INTO test_table (test#, current_ts, generate_u)
WITH temp1 (t1) AS
(VALUES (1),(2),(3),(4))
, temp2 (t1, ts1, ts2) AS
(SELECT t1
, CURRENT TIMESTAMP
, TIMESTAMP(GENERATE_UNIQUE()) + CURRENT TIMEZONE
FROM temp1)
SELECT *
FROM temp2;
Below are the contents of the table after the above insert. Observe the different values: .Table after insert
TEST# | CURRENT_TS | GENERATE_U | GENERATE_A |
---|---|---|---|
1 |
2007-11-13-19.12.43.139000 |
2007-11-13-19.12.42.973805 |
2007-11-13-19.12.43.139000 |
2 |
2007-11-13-19.12.43.139000 |
2007-11-13-19.12.42.974254 |
2007-11-13-19.12.43.154000 |
3 |
2007-11-13-19.12.43.139000 |
2007-11-13-19.12.42.974267 |
2007-11-13-19.12.43.154001 |
4 |
2007-11-13-19.12.43.139000 |
2007-11-13-19.12.42.974279 |
2007-11-13-19.12.43.154002 |
24.21. Floating Point Numbers
The following SQL repetitively multiplies a floating-point number by ten:
WITH temp (f1) AS
(VALUES FLOAT(1.23456789)
UNION ALL
SELECT f1 * 10
FROM temp
WHERE f1 < 1E18)
SELECT f1 AS float1
, DEC(f1,31,8) AS decimal1
, BIGINT(f1) AS bigint1
FROM temp;
After a while, things get interesting:
FLOAT1 | DECIMAL1 | BIGINT1 |
---|---|---|
+1.23456789000000E+000 |
1.23456789 |
1 |
+1.23456789000000E+001 |
12.34567890 |
12 |
+1.23456789000000E+002 |
123.45678900 |
123 |
+1.23456789000000E+003 |
1234.56789000 |
1234 |
+1.23456789000000E+004 |
12345.67890000 |
12345 |
+1.23456789000000E+005 |
123456.78900000 |
123456 |
+1.23456789000000E+006 |
1234567.89000000 |
1234567 |
+1.23456789000000E+007 |
12345678.90000000 |
12345678 |
+1.23456789000000E+008 |
123456789.00000000 |
123456788 |
+1.23456789000000E+009 |
1234567890.00000000 |
1234567889 |
+1.23456789000000E+010 |
12345678900.00000000 |
12345678899 |
+1.23456789000000E+011 |
123456789000.00000000 |
123456788999 |
+1.23456789000000E+012 |
1234567890000.00000000 |
1234567889999 |
+1.23456789000000E+013 |
12345678900000.00000000 |
12345678899999 |
+1.23456789000000E+014 |
123456789000000.00000000 |
123456788999999 |
+1.23456789000000E+015 |
1234567890000000.00000000 |
1234567889999999 |
+1.23456789000000E+016 |
12345678900000000.00000000 |
12345678899999998 |
+1.23456789000000E+017 |
123456789000000000.00000000 |
123456788999999984 |
+1.23456789000000E+018 |
1234567890000000000.00000000 |
1234567889999999744 |
Why do the BIGINT values differ from the original float values? The answer is that they don’t, it is the decimal values that differ. Because this is not what you see in front of your eyes, we need to explain. Note that there are no bugs here, everything is working fine. Perhaps the most insidious problem involved with using floating point numbers is that the number you see is not always the number that you have. Db2 stores the value internally in binary format, and when it displays it, it shows a decimal approximation of the underlying binary value. This can cause you to get very strange results like the following:
WITH temp (f1, f2) AS
(VALUES (FLOAT(1.23456789E1 * 10 * 10 * 10 * 10 * 10 * 10 * 10)
, FLOAT(1.23456789E8)))
SELECT f1
, f2
FROM temp
WHERE f1 <> f2;
ANSWER
F1 | F2 |
---|---|
+1.23456789000000E+008 |
+1.23456789000000E+008 |
We can use the HEX function to show that, internally, the two numbers being compared above are not equal:
WITH temp (f1, f2) AS
(VALUES (FLOAT(1.23456789E1 * 10 * 10 * 10 * 10 * 10 * 10 * 10)
, FLOAT(1.23456789E8)))
SELECT HEX(f1) AS hex_f1
, HEX(f2) AS hex_f2
FROM temp
WHERE f1 <> f2;
ANSWER
HEX_F1 | HEX_F2 |
---|---|
FFFFFF53346F9D41 |
00000054346F9D41 |
Now we can explain what is going on in the recursive code shown at the start of this section. The same value is being displayed using three different methods:
-
The floating-point representation (on the left) is really a decimal approximation (done using rounding) of the underlying binary value.
-
When the floating-point data was converted to decimal (in the middle), it was rounded using the same method that is used when it is displayed directly.
-
When the floating-point data was converted to BIGINT (on the right), no rounding was done because both formats hold binary values.
In any computer-based number system, when you do division, you can get imprecise results due to rounding. For example, when you divide 1 by 3 you get "one third", which can not be stored accurately in either a decimal or a binary number system. Because they store numbers internally differently, dividing the same number in floating-point vs. decimal can result in different results. Here is an example:
WITH temp1 (dec1, dbl1) AS
(VALUES (DECIMAL(1),DOUBLE(1)))
, temp2 (dec1, dec2, dbl1, dbl2) AS
(SELECT dec1
, dec1 / 3 AS dec2
, dbl1
, dbl1 / 3 AS dbl2
FROM temp1)
SELECT *
FROM temp2
WHERE dbl2 <> dec2;
ANSWER (1 row returned)
DEC1 = 1.0 DEC2 = 0.33333333333333333333 DBL1 = +1.00000000000000E+000 DBL2 = +3.33333333333333E-001
When you do multiplication of a fractional floating-point number, you can also encounter rounding differences with respect to decimal. To illustrate this, the following SQL starts with two numbers that are the same, and then keeps multiplying them by ten:
WITH temp (f1, d1) AS
(VALUES (FLOAT(1.23456789)
, DEC(1.23456789,20,10))
UNION ALL
SELECT f1 * 10
, d1 * 10
FROM temp
WHERE f1 < 1E9)
SELECT f1
, d1
, CASE
WHEN d1 = f1 THEN 'SAME'
ELSE 'DIFF'
END AS compare
FROM temp;
Here is the answer:
F1 | D1 | COMPARE |
---|---|---|
+1.23456789000000E+000 |
1.2345678900 |
SAME |
+1.23456789000000E+001 |
12.3456789000 |
SAME |
+1.23456789000000E+002 |
123.4567890000 |
DIFF |
+1.23456789000000E+003 |
1234.5678900000 |
DIFF |
+1.23456789000000E+004 |
12345.6789000000 |
DIFF |
+1.23456789000000E+005 |
123456.7890000000 |
DIFF |
+1.23456789000000E+006 |
1234567.8900000000 |
SAME |
+1.23456789000000E+007 |
12345678.9000000000 |
DIFF |
+1.23456789000000E+008 |
123456789.0000000000 |
DIFF |
+1.23456789000000E+009 |
1234567890.0000000000 |
DIFF |
As we mentioned earlier, both floating-point and decimal fields have trouble accurately storing certain fractional values. For example, neither can store "one third". There are also some numbers that can be stored in decimal, but not in floating-point. One common value is "one tenth", which as the following SQL shows, is approximated in floating-point:
WITH temp (f1) AS
(VALUES FLOAT(0.1))
SELECT f1
, HEX(f1) AS hex_f1
FROM temp
WHERE f1 <> 1.0;
ANSWER
F1 | HEX_F1 |
---|---|
+1.00000000000000E-001 |
9A9999999999B93F |
In conclusion, a floating-point number is, in many ways, only an approximation of a true integer or decimal value. For this reason, this field type should not be used for monetary data, nor for other data where exact precision is required.
24.22. DECFLOAT Usage
We can avoid the problems described above if we use a DECFLOAT value. To illustrate, the following query is exactly the same as that shown on Floating Point Numbers, except that base value is now of type DECFLOAT:
WITH temp (f1) AS
(VALUES DECFLOAT(1.23456789)
UNION ALL
SELECT f1 * 10
FROM temp
WHERE f1 < 1E18)
SELECT f1 AS float1
, DEC(f1,31,8) AS decimal1
, BIGINT(f1) AS bigint1
FROM temp;
Now we get the result that we expect:
FLOAT1 | DECIMAL1 | BIGINT1 |
---|---|---|
+1.23456789000000E+000 |
1.23456789 |
1 |
+1.23456789000000E+001 |
12.34567890 |
12 |
+1.23456789000000E+002 |
123.45678900 |
123 |
+1.23456789000000E+003 |
1234.56789000 |
1234 |
+1.23456789000000E+004 |
12345.67890000 |
12345 |
+1.23456789000000E+005 |
123456.78900000 |
123456 |
+1.23456789000000E+006 |
1234567.89000000 |
1234567 |
+1.23456789000000E+007 |
12345678.90000000 |
12345678 |
+1.23456789000000E+008 |
123456789.00000000 |
123456789 |
+1.23456789000000E+009 |
1234567890.00000000 |
1234567890 |
+1.23456789000000E+010 |
12345678900.00000000 |
12345678900 |
+1.23456789000000E+011 |
123456789000.00000000 |
123456789000 |
+1.23456789000000E+012 |
1234567890000.00000000 |
1234567890000 |
+1.23456789000000E+013 |
12345678900000.00000000 |
12345678900000 |
+1.23456789000000E+014 |
123456789000000.00000000 |
123456789000000 |
+1.23456789000000E+015 |
1234567890000000.00000000 |
1234567890000000 |
+1.23456789000000E+016 |
12345678900000000.00000000 |
12345678900000000 |
+1.23456789000000E+017 |
123456789000000000.00000000 |
123456789000000000 |
+1.23456789000000E+018 |
1234567890000000000.00000000 |
1234567890000000000 |
25. Time Travel
This chapter gives an overview on the time travel feature in Db2. It is based on an article by Jan-Eike Michels and Matthias Nicola published on October 18, 2012 on developerWorks [https://www.ibm.com/developerworks/data/library/techarticle/dm-1210temporaltablesdb2/index.html].
We don’t go that much into detail, just show you how it works. The basic idea is: sometimes you need to keep track on how the data changes over time in the database. While it is possible to track the changes using old database features like triggers (as described in the chapter Retaining a Record), you will get a very robust solution utilizing the native features of the database. The database solution target two different scenarios:
-
track all the changes to data (System-period) and
-
track the validity of data for some time interval (Application-period).
You can have both at the same time defined for one table (Db2 call this a "Bitemporal" table), i.e., you let the database keep track of all changes and at the same time manage the business validity of the data for you.
To show time travel in action let’s take a similar table as we used in chapter Retaining a Record:
CREATE TABLE mycustomer
( cust# INTEGER NOT NULL
, cust_name CHAR(10)
, cust_mgr CHAR(10)
, PRIMARY KEY(cust#));
If you query this table you will always get one manager (the last one) for the customer. If you need to know who was the manager for this customer im some date in the past, you just cannot do it. If this is a need in our business, we can profit from the time travel feature. To implement it for this table we will need the following steps:
-
add new columns for the table (start, end and transaction) with the following characteristics:
-
system_begin TIMESTAMP(12) NOT NULL GENERATED ALWAYS AS ROW BEGIN
-
system_end TIMESTAMP(12) NOT NULL GENERATED ALWAYS AS ROW END
-
trans_start TIMESTAMP(12) GENERATED ALWAYS AS TRANSACTION START ID
-
-
tell the table that this columns will manage the time slices when the data was active in the database:
-
PERIOD SYSTEM_TIME (system_begin, system_end)
-
-
define a new table to save the history data:
-
CREATE TABLE mycustomer_history LIKE customer
-
-
instruct the database management system to activate the time travel feature for this object:
-
ALTER TABLE mycustomer ADD VERSIONING USE HISTORY TABLE mycustomer_history
-
ALTER TABLE mycustomer
ADD COLUMN system_begin TIMESTAMP(12) NOT NULL GENERATED ALWAYS AS ROW BEGIN IMPLICITLY HIDDEN
ADD COLUMN system_end TIMESTAMP(12) NOT NULL GENERATED ALWAYS AS ROW END IMPLICITLY HIDDEN
ADD COLUMN trans_start TIMESTAMP(12) GENERATED ALWAYS AS TRANSACTION START ID
ADD PERIOD SYSTEM_TIME (system_begin, system_end);
CREATE TABLE mycustomer_history LIKE mycustomer;
ALTER TABLE mycustomer ADD VERSIONING USE HISTORY TABLE mycustomer_history;
Now suppose we have the following time line in our company:
Date | Event |
---|---|
01.01.2019 |
First customer is registered in the database. Name: Customer. Manager: Mgr 1 |
01.07.2019 |
Second customer is registered in the database. Name: Customer2. Manager: Mgr 1 |
01.12.2019 |
Third customer is registered in the database. Name: Customer3. Manager: Mgr 2 |
12.01.2020 |
Fourth customer is registerd in the database. Name: Customer4. Manager: Mgr 3 |
15.01.2020 |
Fourth customer is deleted from the database. |
16.01.2020 |
Name of first customer is changed from Customer to Customer1 |
17.01.2020 |
Manager of second customer is changed from Mgr 1 to Mgr 2 |
18.01.2020 |
Manager of third customer is changed from Mgr 2 to Mgr 3 |
Now let us select the current data from our table: .Select current data from table
SELECT *
FROM mycustomer;
CUST# | CUST_NAME | CUST_MGR |
---|---|---|
1 |
Customer1 |
Mgr 1 |
2 |
Customer2 |
Mgr 2 |
3 |
Customer3 |
Mgr 3 |
We don’t see Customer4 because she was already deleted from the database. If we want to see the data as it was on 12.1.2020, we can do it!
SELECT *
FROM mycustomer
FOR SYSTEM_TIME AS OF '2020-01-12'
CUST# | CUST_NAME | CUST_MGR |
---|---|---|
1 |
Customer |
Mgr 1 |
2 |
Customer2 |
Mgr 1 |
3 |
Customer3 |
Mgr 2 |
4 |
Customer4 |
Mgr 3 |
All the changes were done after this date, so we see changes in every single record! Because we defined the columns SYSTEM_BEGIN and SYSTEM_END as IMPLICITLY HIDDEN, we don’t get them in the result set. If we want to see them we need to explicitly code their names in our select statement:
SELECT cust#, cust_name, cust_mgr, system_begin, system_end
FROM mycustomer
FOR SYSTEM_TIME AS OF '2020-01-12'
CUST# | CUST_NAME | CUST_MGR | SYSTEM_BEGIN | SYSTEM_END |
---|---|---|---|---|
1 |
Customer |
Mgr 1 |
2019-01-01 00:00:00.0 |
2020-01-16 00:00:00.0 |
2 |
Customer2 |
Mgr 1 |
2019-07-01 00:00:00.0 |
2020-01-17 00:00:00.0 |
3 |
Customer3 |
Mgr 2 |
2019-12-01 00:00:00.0 |
2020-01-18 00:00:00.0 |
4 |
Customer4 |
Mgr 3 |
2020-01-12 00:00:00.0 |
2020-01-15 00:00:00.0 |
Because the content of SYSTEM_END is always different from the default value of a "ROW END" column in the base table ('9999-12-30 00:00:00.0'), we know that all data came from the history table!
In the first version of the time travel feature IBM used '9999-12-31 00:00:00.0' as default value for the row end. That worked fine when you didn’t work in different time zones! This is the reason why it changed to 9999-12-30: to avoid overflows. |
The history table is a normal table and you can query it. If you want to show all the events in the table mycustomer, you could write the following query:
select 'myCustomer' as Tablename, cust#, cust_name, cust_mgr, system_begin, system_end
from mycustomer
union all
select 'myCustomer_History', cust#, cust_name, cust_mgr, system_begin, system_end
from mycustomer_history
order by system_begin;
TABLENAME | CUST# | CUST_NAME | CUST_MGR | SYSTEM_BEGIN | SYSTEM_END |
---|---|---|---|---|---|
myCustomer_History |
1 |
Customer |
Mgr 1 |
2019-01-01 00:00:00.0 |
2020-01-16 00:00:00.0 |
myCustomer_History |
2 |
Customer2 |
Mgr 1 |
2019-07-01 00:00:00.0 |
2020-01-17 00:00:00.0 |
myCustomer_History |
3 |
Customer3 |
Mgr 2 |
2019-12-01 00:00:00.0 |
2020-01-18 00:00:00.0 |
myCustomer_History |
4 |
Customer4 |
Mgr 3 |
2020-01-12 00:00:00.0 |
2020-01-15 00:00:00.0 |
myCustomer |
1 |
Customer1 |
Mgr 1 |
2020-01-16 00:00:00.0 |
9999-12-30 00:00:00.0 |
myCustomer |
2 |
Customer2 |
Mgr 2 |
2020-01-17 00:00:00.0 |
9999-12-30 00:00:00.0 |
myCustomer |
3 |
Customer3 |
Mgr 3 |
2020-01-18 00:00:00.0 |
9999-12-30 00:00:00.0 |
Now we can answer the question we posted at the beginning: who was the manager Customer2 on Christmas in year 2019?
select cust_name, cust_mgr
FROM mycustomer
FOR SYSTEM_TIME AS OF '2019-12-25'
CUST_NAME | CUST_MGR |
---|---|
Customer |
Mgr 1 |
Customer2 |
Mgr 1 |
Customer3 |
Mgr 2 |
Now if some customer is unsatisfied because she didn’t get a xmas gift, you can blame the right person.
26. Appendix
26.1. Db2 Sample Tables
Sample table DDL follows.
26.1.1. ACT
CREATE TABLE ACT
( ACTNO SMALLINT NOT NULL
, ACTKWD CHARACTER(6) NOT NULL
, ACTDESC VARCHAR(20) NOT NULL)
IN USERSPACE1;
ALTER TABLE ACT
ADD CONSTRAINT PK_ACT PRIMARY KEY (ACTNO);
CREATE UNIQUE INDEX XACT2
ON ACT(ACTNO ASC, ACTKWD ASC)
ALLOW REVERSE SCANS;
26.1.2. CATALOG
CREATE TABLE CATALOG
( NAME VARCHAR(128) NOT NULL
, CATLOG XML)
IN IBMDB2SAMPLEXML;
ALTER TABLE CATALOG
ADD CONSTRAINT PK_CATALOG PRIMARY KEY (NAME);
26.1.3. CL_SCHED
CREATE TABLE CL_SCHED
( CLASS_CODE CHARACTER(7)
, DAY SMALLINT
, STARTING TIME
, ENDING TIME)
IN USERSPACE1;
26.1.4. CUSTOMER
CREATE TABLE CUSTOMER
( CID BIGINT NOT NULL
, INFO XML
, HISTORY XML)
IN IBMDB2SAMPLEXML;
ALTER TABLE CUSTOMER
ADD CONSTRAINT PK_CUSTOMER PRIMARY KEY(CID);
26.1.5. DATA_FILE_NAMES
CREATE TABLE DATA_FILE_NAMES
( DATA_FILE_NAME VARCHAR(40) NOT NULL
, DB2_TABLE_NAME VARCHAR(40) NOT NULL
, EXPORT_FILE_NAM CHARACTER(8) NOT NULL)
IN IBMDB2SAMPLEREL;
26.1.6. DEPARTMENT
CREATE TABLE DEPARTMENT
( DEPTNO CHARACTER(3) NOT NULL
, DEPTNAME VARCHAR(36) NOT NULL
, MGRNO CHARACTER(6)
, ADMRDEPT CHARACTER(3) NOT NULL
, LOCATION CHARACTER(16))
IN USERSPACE1;
ALTER TABLE DEPARTMENT
ADD CONSTRAINT PK_DEPARTMENT PRIMARY KEY(DEPTNO);
CREATE INDEX XDEPT2
ON DEPARTMENT(MGRNO ASC)
ALLOW REVERSE SCANS;
CREATE INDEX XDEPT3
ON DEPARTMENT(ADMRDEPT ASC)
ALLOW REVERSE SCANS;
CREATE ALIAS DEPT FOR DEPARTMENT;
26.1.7. EMPLOYEE
CREATE TABLE EMPLOYEE
( EMPNO CHARACTER(6) NOT NULL
, FIRSTNME VARCHAR(12) NOT NULL
, MIDINIT CHARACTER(1)
, LASTNAME VARCHAR(15) NOT NULL
, WORKDEPT CHARACTER(3)
, PHONENO CHARACTER(4)
, HIREDATE DATE
, JOB CHARACTER(8)
, EDLEVEL SMALLINT NOT NULL
, SEX CHARACTER(1)
, BIRTHDATE DATE
, SALARY DECIMAL(9, 2)
, BONUS DECIMAL(9, 2)
, COMM DECIMAL(9, 2))
IN USERSPACE1;
ALTER TABLE EMPLOYEE
ADD CONSTRAINT PK_EMPLOYEE PRIMARY KEY (EMPNO);
CREATE INDEX XEMP2
ON EMPLOYEE(WORKDEPT ASC)
ALLOW REVERSE SCANS;
CREATE ALIAS EMP FOR EMPLOYEE;
26.1.8. EMPMDC
CREATE TABLE EMPMDC
( EMPNO INTEGER
, DEPT INTEGER
, DIV INTEGER)
IN IBMDB2SAMPLEREL;
26.1.9. EMPPROJACT
CREATE TABLE EMPPROJACT
( EMPNO CHARACTER(6) NOT NULL
, PROJNO CHARACTER(6) NOT NULL
, ACTNO SMALLINT NOT NULL
, EMPTIME DECIMAL(5, 2)
, EMSTDATE DATE
, EMENDATE DATE)
IN USERSPACE1;
CREATE ALIAS EMP_ACT FOR EMPPROJACT;
CREATE ALIAS EMPACT FOR EMPPROJACT;
26.1.10. EMP_PHOTO
CREATE TABLE EMP_PHOTO
( EMPNO CHARACTER(6) NOT NULL
, PHOTO_FORMAT VARCHAR(10) NOT NULL
, PICTURE BLOB(102400)
, EMP_ROWID CHARACTER(40) NOT NULL)
IN USERSPACE1;
ALTER TABLE EMP_PHOTO
ADD CONSTRAINT PK_EMP_PHOTO PRIMARY KEY(EMPNO, PHOTO_FORMAT);
26.1.11. EMP_RESUME
CREATE TABLE EMP_RESUME
( EMPNO CHARACTER(6) NOT NULL
, RESUME_FORMAT VARCHAR(10) NOT NULL
, RESUME CLOB(5120)
, EMP_ROWID CHARACTER(40) NOT NULL)
IN USERSPACE1;
ALTER TABLE EMP_RESUME
ADD CONSTRAINT PK_EMP_RESUME PRIMARY KEY(EMPNO, RESUME_FORMAT);
26.1.12. IN_TRAY
CREATE TABLE IN_TRAY
( RECEIVED TIMESTAMP
, SOURCE CHARACTER(8)
, SUBJECT CHARACTER(64)
, NOTE_TEXT VARCHAR(3000))
IN USERSPACE1;
CREATE TABLE INVENTORY
( PID VARCHAR(10) NOT NULL
, QUANTITY INTEGER
, LOCATION VARCHAR(128))
IN IBMDB2SAMPLEXML;
ALTER TABLE INVENTORY
ADD CONSTRAINT PK_INVENTORY PRIMARY KEY(PID);
26.1.13. ORG
CREATE TABLE ORG
( DEPTNUMB SMALLINT NOT NULL
, DEPTNAME VARCHAR(14)
, MANAGER SMALLINT
, DIVISION VARCHAR(10)
, LOCATION VARCHAR(13))
IN USERSPACE1;
26.1.14. PRODUCT
CREATE TABLE PRODUCT
( PID VARCHAR(10) NOT NULL
, NAME VARCHAR(128)
, PRICE DECIMAL(30,2)
, PROMOPRICE DECIMAL(30, 2)
, PROMOSTART DATE
, PROMOEND DATE
, DESCRIPTION XML)
IN IBMDB2SAMPLEXML;
ALTER TABLE PRODUCT
ADD CONSTRAINT PK_PRODUCT PRIMARY KEY(PID);
26.1.15. PRODUCTSUPPLIER
CREATE TABLE PRODUCTSUPPLIER
( PID VARCHAR(10) NOT NULL
, SID VARCHAR(10) NOT NULL)
IN IBMDB2SAMPLEXML;
26.1.16. PROJACT
CREATE TABLE PROJACT
( PROJNO CHARACTER(6) NOT NULL
, ACTNO SMALLINT NOT NULL
, ACSTAFF DECIMAL(5, 2)
, ACSTDATE DATE NOT NULL
, ACENDATE DATE)
IN USERSPACE1;
ALTER TABLE PROJACT
ADD CONSTRAINT PK_PROJACT PRIMARY KEY(PROJNO, ACTNO, ACSTDATE);
26.1.17. PROJECT
CREATE TABLE PROJECT
( PROJNO CHARACTER(6) NOT NULL
, PROJNAME VARCHAR(24) NOT NULL
, DEPTNO CHARACTER(3) NOT NULL
, RESPEMP CHARACTER(6) NOT NULL
, PRSTAFF DECIMAL(5, 2)
, PRSTDATE DATE
, PRENDATE DATE
, MAJPROJ CHARACTER(6))
IN USERSPACE1;
ALTER TABLE PROJECT
ADD CONSTRAINT PK_PROJECT PRIMARY KEY(PROJNO);
CREATE INDEX XPROJ2
ON PROJECT(RESPEMP ASC)
ALLOW REVERSE SCANS;
CREATE ALIAS PROJ FOR PROJECT;
26.1.18. PURCHASEORDER
CREATE TABLE PURCHASEORDER
( POID BIGINT NOT NULL
, STATUS VARCHAR(10) NOT NULL
, CUSTID BIGINT
, ORDERDATE DATE
, PORDER XML
, COMMENTS VARCHAR(1000))
IN IBMDB2SAMPLEXML;
ALTER TABLE PURCHASEORDER
ADD CONSTRAINT PK_PURCHASEORDER PRIMARY KEY(POID);
26.1.19. SALES
CREATE TABLE SALES
( SALES_DATE DATE
, SALES_PERSON VARCHAR(15)
, REGION VARCHAR(15)
, SALES INTEGER)
IN USERSPACE1;
26.1.20. STAFF
CREATE TABLE STAFF
( ID SMALLINT NOT NULL
, NAME VARCHAR(9)
, DEPT SMALLINT
, JOB CHARACTER(5)
, YEARS SMALLINT
, SALARY DECIMAL(7, 2)
, COMM DECIMAL(7,2))
IN USERSPACE1;
26.1.21. SUPPLIERS
CREATE TABLE SUPPLIERS
( SID VARCHAR(10) NOT NULL
, ADDR XML)
IN IBMDb2SAMPLEXML;
ALTER TABLE SUPPLIERS
ADD CONSTRAINT PK_PRODUCTSUPPLIER PRIMARY KEY(SID);
27. Thank you, Graeme Birchall!
This chapter contains the last introduction, history and comments from Graeme Birchall in his "Db2 SQL Cookbook". It is a kind of tribute to him and his work.
Db2 LUW V9.7 SQL Cookbook Graeme Birchall 16-Aug-2011
27.1. Preface Important!
If you didn’t get this document directly from my personal website, you may have got an older edition. The book is changed very frequently, so if you want the latest, go to the source. Also, the latest edition is usually the best book to have, as the examples are often much better. This is true even if you are using an older version of Db2. This Cookbook is written for Db2 for LUW (i.e. Linux, Unix, Windows). It is not suitable for Db2 for z/OS unless you are running Db2 8 in new-function-mode, or (even better) Db2 9.
27.2. Acknowledgments
I did not come up with all of the ideas presented in this book. Many of the best examples were provided by readers, friends, and/or coworkers too numerous to list. Thanks also to the many people at IBM for their (strictly unofficial) assistance.
27.3. Disclaimer & Copyright
DISCLAIMER: This document is a best effort on my part. However, I screw up all the time, so it would be extremely unwise to trust the contents in its entirety. I certainly don’t. And if you do something silly based on what I say, life is tough. COPYRIGHT: You can make as many copies of this book as you wish. And I encourage you to give it to others. But you cannot charge for it (other than to recover reproduction costs), nor claim the material as your own, nor replace my name with another. You are also encouraged to use the related class notes for teaching. In this case, you can charge for your time and materials - and your expertise. But you cannot charge any licensing fee, nor claim an exclusive right of use. In other words, you can pretty well do anything you want. And if you find the above too restrictive, just let me know. TRADEMARKS: Lots of words in this document, like "Db2", are registered trademarks of the IBM Corporation. Lots of other words, like "Windows", are registered trademarks of the Microsoft Corporation. Acrobat is a registered trademark of the Adobe Corporation.
27.4. Tools Used
This book was written on a Dell PC that came with oodles of RAM. All testing was done in Db2 V9.7 Express-C for Windows. Word for Windows was used to write the document. Adobe Acrobat was used to make the PDF file.
27.5. Book Binding
This book looks best when printed on a doubled sided laser printer and then suitably bound. To this end, I did some experiments a few years ago to figure out how to bind books cheaply using commonly available materials. I came up with what I consider to be a very satisfactory solution that is fully documented on Book Binding.
27.6. Author / Book
Author: Email: Graeme_Birchall@verizon.net Web: http://mysite.verizon.net/Graeme_Birchall/ Title: Db2 9.7 SQL Cookbook © Date: 16-Aug-2011
27.7. Preface
27.7.1. Author
Notes Book History
This book originally began a series of notes for my own use. After a while, friends began to ask for copies, and enemies started to steal it, so I decided to tidy everything up and give it away. Over the years, new chapters have been added as Db2 has evolved, and as I have found new ways to solve problems. Hopefully, this process will continue for the foreseeable future.
27.7.2. Why Free
This book is free because I want people to use it. The more people that use it, and the more that it helps them, the more inclined I am to keep it up to date. For these reasons, if you find this book to be useful, please share it with others. This book is free, rather than formally published, because I want to deliver the best product that I can. If I had a publisher, I would have the services of an editor and a graphic designer, but I would not be able to get to market so quickly, and when a product changes as quickly as Db2 does, timeliness is important. Also, giving it away means that I am under no pressure to make the book marketable. I simply include whatever I think might be useful.
27.7.3. Other Free Documents
The following documents are also available for free from my web site: SAMPLE SQL: The complete text of the SQL statements in this Cookbook is available in an HTML file. Only the first and last few lines of the file have HTML tags, the rest is raw text, so it can easily be cut and paste into other files. CLASS OVERHEADS: Selected SQL examples from this book have been rewritten as class overheads. This enables one to use this material to teach Db2 SQL to others. Use this cookbook as the student notes. OLDER EDITIONS: This book is rewritten, and usually much improved, with each new version of Db2. Some of the older editions are available from my website. The others can be emailed upon request. However, the latest edition is the best, so you should probably use it, regardless of the version of Db2 that you have.
27.7.4. Answering Questions
As a rule, I do not answer technical questions because I need to have a life. But I’m interested in hearing about interesting SQL problems, and also about any bugs in this book. However you may not get a prompt response, or any response. And if you are obviously an idiot, don’t be surprised if I point out (for free, remember) that you are an idiot.
27.7.5. Software Whines
This book is written using Microsoft Word for Windows. I’ve been using this software for many years, and it has generally been a bunch of bug-ridden junk. I do confess that it has been mildly more reliable in recent years. However, I could have written more than twice as much that was twice as good in half the time - if it weren’t for all of the bugs in Word.
Graeme
27.7.6. Graeme Birchall Book Editions Upload Dates
Date Published (Version) | Content |
---|---|
1996-05-08 |
First edition of the Db2 V2.1.1 SQL Cookbook was posted to my web site. This version was in Postscript Print File format. |
1998-02-26 |
The Db2 V2.1.1 SQL Cookbook was converted to an Adobe Acrobat file and posted to my web site. Some minor cosmetic changes were made. |
1998-08-19 |
First edition of Db2 UDB V5 SQL Cookbook posted. Every SQL statement was checked for V5, and there were new chapters on OUTER JOIN and GROUP BY. |
1998-08-26 |
About 20 minor cosmetic defects were corrected in the V5 Cookbook. |
1998-09-03 |
Another 30 or so minor defects were corrected in the V5 ookbook. |
1998-10-24 |
The Cookbook was updated for Db2 UDB V5.2. |
1998-10-25 |
About twenty minor typos and sundry cosmetic defects were fixed. |
1998-12-03 |
This book was based on the second edition of the V5.2 upgrade. |
1999-01-25 |
A chapter on Summary Tables (new in the Dec/98 fixpack) was added and all the SQL was checked for changes. |
1999-01-28 |
Some more SQL was added to the new chapter on Summary Tables. |
1999-02-15 |
The section of stopping recursive SQL statements was completely rewritten, and a new section was added on denormalizing hierarchical data structures. |
1999-02-16 |
Minor editorial changes were made. |
1999-03-16 |
Some bright spark at IBM pointed out that my new and improved section on stopping recursive SQL was all wrong. Damn. I undid everything. |
1999-05-12 |
Minor editorial changes were made, and one new example (on getting multiple counts from one value) was added. |
1999-09-16 |
Db2 V6.1 edition. All SQL was rechecked, and there were some minor additions - especially to summary tables, plus a chapter on "Db2 Dislikes". |
1999-09-23 |
Some minor layout changes were made. |
1999-10-06 |
Some bugs fixed, plus new section on index usage in summary tables. |
2000-04-12 |
Some typos fixed, and a couple of new SQL tricks were added. |
2000-09-19 |
Db2 V7.1 edition. All SQL was rechecked. The new areas covered are: OLAP functions (whole chapter), ISO functions, and identity columns. |
2000-09-25 |
Some minor layout changes were made. |
2000-10-26 |
More minor layout changes. |
2001-01-03 |
Minor layout changes (to match class notes). |
2001-02-06 |
Minor changes, mostly involving the RAND function. |
2001-04-11 |
Document new features in latest fixpack. Also add a new chapter on Identity Columns and completely rewrite sub-query chapter. |
2001-10-24 |
Db2 V7.2 fixpack 4 edition. Tested all SQL and added more examples, plus a new section on the aggregation function. |
2002-03-11 |
Minor changes, mostly to section on precedence rules. |
2002-08-20 |
Db2 V8.1 (beta) edition. A few new functions are added. New section on temporary tables. Identity Column and Join chapters rewritten. Whine chapter removed. |
2003-01-02 |
Db2 V8.1 (post-Beta) edition. SQL rechecked. More examples added. |
2003-07-11 |
New sections added on DML, temporary tables, compound SQL, and user defined functions. Halting recursion section changed to use ser-defined function. |
2003-09-04 |
New sections on complex joins and history tables. |
2003-10-02 |
Minor changes. Some more user-defined functions. |
2003-11-20 |
Added "quick find" chapter. |
2003-12-31 |
Tidied up the SQL in the Recursion chapter, and added a section on the merge statement. Completely rewrote the chapter on materialized query tables. |
2004-02-04 |
Added select-from-DML section, and tidied up some code. Also managed to waste three whole days due to bugs in Microsoft Word. |
2004-07-23 |
Rewrote chapter of identity column and sequences. Made DML separate chapter. Added chapters on protecting data and XML functions. Other minor changes. |
2004-11-03 |
Upgraded to V8.2. Retested all SQL. Documented new SQL features. Some major hacking done on the GROUP BY chapter. |
2005-04-15 |
Added short section on cursors, and a chapter on using SQL to make SQL. |
2005-06-01 |
Added a chapter on triggers. |
2005-11-11 |
Updated MQT table chapter and added bibliography. Other minor changes. |
2005-12-01 |
Applied fixpack 10. Changed my website name. |
2005-12-16 |
Added notes on isolation levels, data-type functions, transforming data. |
2006-01-26 |
Fixed dumb bugs generated by WORD. What stupid software. Also wrote an awesome new section on joining meta-data to real data. |
2006-02-17 |
Touched up the section on joining meta-data to real data. Other minor fixes. |
2006-02-27 |
Added precedence rules for SQL statement processing, and a description of a simplified nested table expression. |
2006-03-23 |
Added better solution to avoid fetching the same row twice. |
2006-04-26 |
Added trigger that can convert HEX value to number. |
2006-09-08 |
Upgraded to V9.1. Retested SQL. Removed the XML chapter as it is now obsolete. I’m still cogitating about XQuery. Looks hard. Added some awesome java code. |
2006-09-13 |
Fixed some minor problems in the initial V9.1 book. |
2006-10-17 |
Fixed a few cosmetic problems that were bugging me. |
2006-11-06 |
Found out that IBM had removed the "UDB" from the Db2 product name, so I did the same. It is now just plain "Db2 V9". |
2006-11-29 |
I goofed. Turns out Db2 is now called "Db2 9". I relabeled accordingly. |
2006-12-15 |
Improved code to update or delete first "n" rows. |
2007-02-22 |
Get unique timestamp values during multi-row insert. Other minor changes. |
2007-11-20 |
Finished the Db2 V9.5 edition. Lots of changes! |
2008-09-20 |
Fixed some minor problems. |
2008-11-28 |
Fixed some minor problems. |
2009-01-18 |
Fixed some minor problems, plus lots of bugs in Microsoft WORD! |
2009-03-12 |
Converted to a new version of Adobe Acrobat, plus minor fixes. |
2010-10-12 |
Finished initial V9.7 edition. Only minor changes. More to come. |
2010-11-05 |
First batch of cute/deranged V9.7 SQL examples added. |
2010-11-14 |
Fixed some minor typos. |
2011-01-11 |
Added LIKE_COLUMN function. Removed bibliography. |
2011-01-14 |
Added HASH function. Other minor edits. |
2011-08-16 |
Fixed some minor problems. |
27.8. Book Binding
Below is a quick-and-dirty technique for making a book out of this book. The object of the exercise is to have a manual that will last a long time, and that will also lie flat when opened up. All suggested actions are done at your own risk.
27.8.1. Tools Required
-
Printer, to print the book.
-
KNIFE, to trim the tape used to bind the book.
-
BINDER CLIPS, (1" size), to hold the pages together while gluing. To bind larger books, or to do multiple books in one go, use two or more cheap screw clamps.
-
CARDBOARD: Two pieces of thick card, to also help hold things together while gluing.
27.8.2. Consumables
Ignoring the capital costs mentioned above, the cost of making a bound book should work out to about $4.00 per item, almost all of which is spent on the paper and toner. To bind an already printed copy should cost less than fifty cents.
-
PAPER and TONER, to print the book.
-
CARD STOCK, for the front and back covers.
-
GLUE, to bind the book. Cheap rubber cement will do the job. The glue must come with an applicator brush in the bottle. Sears hardware stores sell a more potent flavor called Duro Contact Cement that is quite a bit better. This is toxic stuff, so be careful.
-
CLOTH TAPE, (2" wide) to bind the spine. Pearl tape, available from Pearl stores, is fine. Wider tape will be required if you are not printing double-sided.
-
TIME: With practice, this process takes less than five minutes work per book.
27.8.3. Before you Start
-
Make that sure you have a well-ventilated space before gluing.
-
Practice binding on some old scraps of paper.
-
Kick all kiddies out off the room.
27.8.4. Instructions
-
Print the book - double-sided if you can. If you want, print the first and last pages on card stock to make suitable protective covers.
-
Jog the pages, so that they are all lined up along the inside spine. Make sure that every page is perfectly aligned, otherwise some pages won’t bind. Put a piece of thick cardboard on either side of the set of pages to be bound. These will hold the pages tight during the gluing process.
-
Place binder clips on the top and bottom edges of the book (near the spine), to hold everything in place while you glue. One can also put a couple on the outside edge to stop the pages from splaying out in the next step. If the pages tend to spread out in the middle of the spine, put one in the centre of the spine, then work around it when gluing. Make sure there are no gaps between leafs, where the glue might soak in.
-
Place the book spine upwards. The objective here is to have a flat surface to apply the glue on. Lean the book against something if it does not stand up freely.
-
Put on gobs of glue. Let it soak into the paper for a bit, then put on some more.
-
Let the glue dry for at least half an hour. A couple of hours should be plenty.
-
Remove the binder clips that are holding the book together. Be careful because the glue does not have much structural strength.
-
Separate the cardboard that was put on either side of the book pages. To do this, carefully open the cardboard pages up (as if reading their inside covers), then run the knife down the glue between each board and the rest of the book.
-
Lay the book flat with the front side facing up. Be careful here because the rubber cement is not very strong.
-
Cut the tape to a length that is a little longer that the height of the book.
-
Put the tape on the book, lining it up so that about one quarter of an inch (of the tape width) is on the front side of the book. Press the tape down firmly (on the front side only) so that it is properly attached to the cover. Make sure that a little bit of tape sticks out of both the bottom and top ends of the spine.
-
Turn the book over (gently) and, from the rear side, wrap the cloth tape around the spine of the book. Pull the tape around so that it puts the spine under compression.
-
Trim excess tape at either end of the spine using a knife or pair of scissors.
-
Tap down the tape so that it is firmly attached to the book.
-
Let the book dry for a day. Then do the old "hold by a single leaf" test. Pick any page, and gently pull the page up into the air. The book should follow without separating from the page.
27.8.5. More Information
The binding technique that I have described above is fast and easy, but rather crude. It would not be suitable if one was printing books for sale. There are plenty of other binding methods that take a little more skill and better gear that can be used to make "store-quality" books. Search the web for more information.
Index
Note: For technical reasons the index is only shown in the PDF format of this book.