![]() |
![]() |
|
|
![]() |
![]() |
ALLBASE/SQL Performance and Monitoring Guidelines: HP 9000 Computer Systems > Chapter 5 Guidelines on System Administration![]() DBA Guidelines |
|
Several aspects of database administration affect performance. To improve performance in these areas, you need DBA authority. When you run an application, ALLBASE/SQL first checks each section for validity before executing it. Then, ALLBASE/SQL revalidates invalid sections (if possible) before executing them. You may notice a slight delay as the sections are being revalidated. Revalidating sections can reduce concurrency, because it involves updating system catalog tables. For each update, ALLBASE/SQL must obtain an exclusive page lock on the system catalog tables. The locks are retained until the transaction is committed. To improve performance, you can revalidate your applications before run time by either:
You should also be careful when you use the UPDATE STATISTICS statement, because it invalidates sections. For best performance, after the initial statistics for a table have been established, use UPDATE STATISTICS only during periods of low DBEnvironment activity. Then, after you use UPDATE STATISTICS, use VALIDATE to revalidate modules and procedures. You may want to use VALIDATE after executing an UPDATE STATISTICS statement or any DDL statement, since these statements invalidate stored sections. If you issue such statements during a period of low activity for the DBEnvironment (at night, for example), the DBEnvironment experiences minimal performance degradation. The following tips relate to preprocessing of application programs:
Balancing system load is an effort to spread out I/O in your system among the available disk devices for the most efficient operation. Large systems can benefit significantly from an analysis of system load, followed by reallocating the DBEFiles for specific DBEFileSets or within a DBEFileSet. In general, the DBEFiles on which the most I/O is performed should be stored on the fastest devices available. For best performance, separate data DBEFiles, index DBEFiles, and log files across the available disks on your system. Specifically, make sure to:
If two database objects are used concurrently, you should store them on different disk drives to minimize disk head contention. This includes any two tables that show I/O at the same time, any two B-tree indexes that show I/O at the same time, or a table that shows I/O at the same time as the B-tree index defined on it. To monitor I/O, run SQLMON and access the SampleIO subsystem. When you access objects simultaneously, the disk heads may move back and forth between the objects. This movement slows access to data. To avoid this problem, use the SQLUtil MOVEFILE command to place the competing objects on different drives. Remember that the mapping of database objects to DBEFiles is not one-to-one, and that MOVEFILE moves DBEFiles, not DBEFileSets. In order to achieve the goal of placing different objects on different drives, you must ensure that they are in different DBEFiles. Therefore, it is useful to create separate INDEX and TABLE DBEFiles, and to define a different DBEFileSet for each table or index that may need to be separate from other objects. As you adjust the various configurable parameters in ALLBASE/SQL, be sure you do not exceed the amount of shared memory available on your system. You can use the following formula to derive an approximate figure (in kilobytes) for the total shared memory used:
where
The result of this equation must be within the amount of shared memory you have available. This equation also gives an idea of the relative effect of changing different parameters on the total shared memory size. For the default parameters, you can use
And for the maximum parameters, you can use
These equations will help you arrive at a first approximation of the amount of shared memory that has been used. Data buffers are used to cache data pages touched by ALLBASE/SQL. Up to the limit of buffer space, a data page is retained in memory until the space it occupies is needed for another page that is on disk. You can set the number of data buffer pages your DBEnvironment uses. The minimum number of data buffer pages is 15, and the default number is 100. The maximum is determined by the amount of shared memory available. How many data buffer pages should you allocate? There is no explicit equation to find the ideal buffer pool size. The following general suggestions may help, though this is not an exhaustive list:
When should you increase the number of buffer pages? If the set of applications does not have locality of data and index pages, it may not matter how many data buffers there are. Once you have enough data buffer pages, performance is not very sensitive to adding more. How do you keep a small, frequently used group of pages in memory consistently? The trick is to cause other pages to be swapped out first. ALLBASE/SQL employs a least recently used (LRU) algorithm to discard older pages and make room for new pages. To keep a special group of pages in memory, you must have other pages in memory as well that can age past the pages in the special group. If you use a large enough number of pages outside the special group, each page will age to the point of being least recently used before any of the special group is swapped out. Use the following suggestions to determine how many data buffer pages is enough:
For this example, assume that a sample application
Assume that the tables for the sample occupy the following space in DBEFiles:
Also assume that:
At a given instant, the buffer pool looks like this:
Based on the numbers of pages given above, and assuming the default number of data buffer pages (100), the application will require an average of 7 I/Os per transaction, as shown in the following table:
There is a 25% chance that a random access to the small table will want a page that is already in the buffer pool. There is a 10% chance that the large index second-level page is already in the buffer pool. These reduce the I/O count to about 6.4 instead of 7, as shown in the following table:
System performance improves if you add more data buffers. The first threshold of performance gain is when there are enough data buffers to prevent the pages of the small table from being swapped out. This will happen only if the small table's pages never become least recently used. Suppose that we want 62 pages to stay in memory from the previous calculation, and that we will insert about 10 pages. In each transaction, accessing the large table will require touching a root page, two more index pages, and a data page. If the two nonroot index pages and the data page are read each time, we touch three pages that are new. We need enough pages in memory so that pages from the large table are less recently used than pages from the small table. A minimal estimate would be 120 pages; a more comfortable estimate would allow 160 pages. Using the 120-page minimum, the buffer pool would look like this:
Here is the resultant pattern of I/Os:
The next performance gain comes from fitting all the pages from the second level of the large table's index into the data buffer cache. There are 100 pages in the second-level index, so the buffer pool in this scenario looks like the following:
Here is the resultant pattern of I/Os:
This analysis predicts one plateau of performance at about 180-200 pages in the buffer pool, a second plateau at about 360-400 pages. The next plateau would require something like 25,000 pages, which is impractical for most installations. In practice, there is little penalty for extra pages, so use a generous allocation (that is, at least 200 or 400 pages), if you can afford to.
Using SQLMON, you can observe the data buffer miss rate and adjust the number of data buffer pages accordingly. Start by allocating the largest number of data buffer pages you are willing to reserve for ALLBASE/SQL. This number should be less than the following:
As you observe the data buffer miss rate on SQLMON's IO screen, gradually reduce the number of data buffer pages until the miss rate increases sharply. Allocate just enough data buffer pages to avoid the sharp increase of the data buffer miss rate. Since the load on your system probably varies, you should monitor the data buffer miss rate at different times throughout the day. The runtime control block is an area of shared memory containing global, runtime information for the DBEnvironment. ALLBASE/SQL allocates control blocks from the runtime control block. Lock management is the single greatest user of control blocks. Each table, page, or row lock acquired needs one control block. As the granularity of the lock decreases, the number of locks required is likely to increase. For example, locking an entire table with row level locking requires more locks than locking the table with the LOCK TABLE statement. The following table lists the maximum number of locks that can be associated with table, page, and row-level locking:
An application that manages locks well is less likely to deplete the amount of shared memory available. If the runtime control block is too small, ALLBASE/SQL is unable to allocate the necessary control blocks, an error is returned, and the transaction is rolled back. If the runtime control block is too large, other processes on the system may not have adequate access to memory. To allocate pages for the runtime control block, you can use the START DBE statement or the ALTDBE command. To monitor the usage of the runtime control block, start SQLMON and go to the Overview screen, as described in the chapter "Getting Started with SQLMON." The number of log buffers is independent of the number of data buffers. The log buffer pages are only 512 bytes (in contrast with the data buffer pages, which are 4096 bytes). If the log buffer pool becomes full before any transaction completes, then at least two I/Os are needed to complete the transaction. You should provide enough log buffers to allow all active transactions to have log space. Additional buffers do not help performance, but adding pages probably doesn't hurt either. To monitor the I/O incurred by logging, run SQLMON and access the IO screen. For example, suppose that a user is making changes to a sample database and that
The figures listed above total 800 bytes. You can then round up to 1024 bytes, or 2 512-byte pages. This implies that 20 log buffer pages should be sufficient for 10 transactions. If there are occasional bursts of more than the average, 40 or 50 pages might be a good idea. ALLBASE/SQL gives you great flexibility in choosing the number and size of log files. The following suggestions may be useful in making choices for your system. The size of nonarchive log files determines the frequency of automatic system checkpoints, and it therefore determines rollback recovery time. The total size of the log (all files taken together) determines the largest transaction that can be accomodated in the DBEnvironment. Some guidelines for determining the size of the nonarchive log are listed below:
The sort method in ALLBASE/SQL is memory-intensive, using the tournament sort algorithm for sorting and merging. It sorts in two phases:
Use the CREATE TEMPSPACE statement to create a directory that will contain temporary space for scratch files. ALLBASE/SQL needs scratch files to store the intermediate results of sorting and creating indexes. You should create at least one temporary space in the DBEnvironment you are working in. If you do not create a temporary space, ALLBASE/SQL opens scratch files in the /tmp directory. The CREATE TEMPSPACE statement itself does not use any disk space, and ALLBASE/SQL does not create any scratch files when you issue the CREATE TEMPSPACE statement. Instead, they are dynamically created and purged during sort operations. When you define a temporary space, the amount of available disk space may change dynamically as ALLBASE/SQL creates and purges scratch files. By default, the number of pages used for each temporary space file is 256. The total amount of space used is all that is available in the directory /tmp.
All intermediate sort results are written to scratch files, not to any DBEFileSet in the DBEnvironment. Sorting may require 100% to 200% of the input data size in temporary disk files for intermediate results. If a join and an ORDER BY are both involved, the amount could be as much as 300%. However, scratch files are purged as soon as the sort operation is finished. Final sort output for queries with GROUP BY, ORDER BY or DISTINCT is written to the SYSTEM DBEFileSet, requiring space worth 100% of the sorted output. For CREATE INDEX, the final output is directly written to the DBEFileSet containing the index, requiring space worth the size of the index. The disk space required for sort operations is summarized below:
The order in which you carry out database operations may determine the amount of temporary space required. For example, if you create a large table, load it, then create an index, you will probably use scratch file space for sorting. However, if you create the table, then create the index, then load the data, you will not use scratch files. In the second case, the load is 3 to 10 times slower because ALLBASE/SQL creates index entries during loading. However, if you load unsorted data, the size of the index is about 50% larger. Sorting is memory-intensive, and the amount of memory required is estimated from the number of input tuples. The number of input tuples is in turn estimated from table statistics. Therefore, it is important to have up-to-date statistics when you execute ORDER BY, GROUP BY, or CREATE INDEX statements, especially on large tables.
ALLBASE/SQL uses two join methods: nested loop join and sort/merge join. Nested loop joins are usually much faster than sort/merge joins. By default, the optimizer chooses a join method depending on the query and the statistics in the system catalog for the tables involved. To override the join method the optimizer chooses, you can use the SETOPT statement. The nested loop method scans the second table once for each qualifying row in the first table. If the scan on the second table is a serial scan (rather than an index scan), the nested loop join can require considerable I/O and CPU time when the table is larger than the buffer cache. However, a nested loop join may be the best approach for joins involving small tables. A sort/merge join has two phases: the sort phase and the merge/join phase.
A sort/merge join can help with tables that have indexes that are larger than the buffer cache size. Query results from queries using ORDER BY and GROUP BY clauses are stored in temporary pages in the SYSTEM DBEFileSet. Be sure to include enough TABLE or MIXED DBEFile space in SYSTEM to accomodate your largest sorted query. As your system catalog grows, you should monitor the DBEFile space in SYSTEM occasionally (after doing an UPDATE STATISTICS on at least one system view) to make sure enough temporary space is still available. As needed, add DBEFiles to SYSTEM to supply temporary pages. Section caching holds a number of sections in memory between transactions so that they do not need to be read again for the next execution. ALLBASE/SQL can keep up to 12 sections in memory. Section caching is more efficient when DDL Enabled is set to NO. Directory caching stores in memory the DBCore directory, which contains the locations of tables and indexes. Directory caching is turned on when you set the DDL Enabled flag to NO by using ALTDBE in SQLUtil. Section caching is subject to the following rules:
This feature is helpful for an application that executes the same sections repeatedly, especially if it has a small number of sections. By default, ALLBASE/SQL allocates shared memory for up to 12 sections. You can increase this number by setting the environment variable HPSQLsectcache to any value from 4 to 128. It is possible to establish a maximum of 32 simultaneous database connections. When your application must access more than one DBEnvironment, there is no need to release one before connecting to another. Performance is greatly improved by using this method rather than connecting to and releasing each DBEnvironment individually. For more detailed information, refer to the chapter "Transaction Management with Multiple DBEnvironment Connections" in the ALLBASE/SQL Advanced Application Programming Guide and the section "Using Multiple Connections and Transactions with Timeouts" in the ALLBASE/SQL Reference Manual chapter "Using ALLBASE/SQL." When an application requests a database resource that is unavailable, the application is placed on a wait queue. If the application waits longer than the timeout value specified for its DBEnvironment connection, an error occurs, and the transaction is rolled back. Your strategy for specifying timeout values and handling timeout errors depends on the specific needs of your application and on your business procedures. By default, the timeout value is infinite. For example, a transaction may require resources that are locked by a second transaction that requires a great deal of execution time. In this case, you could specify a reasonable amount of time for the first transaction to wait before a timeout occurs. Such a strategy might be essential in a distributed environment. For more detailed information, refer to the chapter "Transaction Management with Multiple DBEnvironment Connections" in the ALLBASE/SQL Advanced Application Programming Guide and the section "Setting Timeout Values" in the ALLBASE/SQL Reference Manual chapter "Using ALLBASE/SQL." |
![]() |
||
![]() |
![]() |
![]() |
|||||||||
|