An Adabas session involves the execution of the Adabas nucleus which controls access/update to a single database. This chapter describes the job control statements needed when executing an Adabas session under each supported operating system. For examples of the Adabas utility jobs, see the Adabas Utilities documentation.
This document covers the following topics:
The following table contains all files which are used when executing an Adabas session under BS2000.
File | Link Name | Logical Unit | Storage Medium | Additional Information |
---|---|---|---|---|
ADARUN parameters | DDCARD | SYSDTA | disk | note 1 |
ADARUN / Adabas messages | DDPRINT | SYSOUT | disk | note 2 |
Associator | DDASSORn or DDASSOnn |
disk | note 3 | |
Data Storage | DDDATARn or DDDATAnn |
disk | note 3 | |
Work | DDWORKR1 DDWORKR4 |
disk | note 4 | |
Recovery Aid log | DDRLOGR1 | disk | note 5 | |
Protection log multiple log 1 multiple log 2 |
DDSIBA DDPLOGR1 DDPLOGR2 |
tape disk disk disk |
note 6 note 7 note 7 |
|
Command log multiple log 1 multiple log 2 |
DDLOG DDCLOGR1 DDCLOGR2 |
tape disk disk disk |
note 8 note 9 note 9 |
|
ECS encoding objects | DDECSOJ | tape/disk | note 10 | |
Time zone objects | TZINFO | disk | note 11 |
This job includes multiple protection logging, multiple command logging, Recovery Aid logging, and universal encoding support (UES):
/.ADANUC SET-LOGON=PARAMETERS /MODIFY-TEST-OPTIONS DUMP=YES /REMARK * /REMARK * ADABAS NUCLEUS /REMARK * /ASS-SYSLST L.NUC /ASS-SYSDTA *SYSCMD /SET-FILE-LINK DDLIB,ADAvrs.MOD /SET-FILE-LINK BLSLIB01,$.SYSLNK.LMS /SET-FILE-LINK DDASSOR1,ADAyyyyy.ASSO,SHARE-UPD=YES /SET-FILE-LINK DDDATAR1,ADAyyyyy.DATA,SHARE-UPD=YES /SET-FILE-LINK DDWORKR1,ADAyyyyy.WORK /SET-FILE-LINK DDPLOGR1,ADAyyyyy.PLOGR1,SHARE-UPD=YES /SET-FILE-LINK DDPLOGR2,ADAyyyyy.PLOGR2,SHARE-UPD=YES /SET-FILE-LINK DDCLOGR1,ADAyyyyy.CLOGR1,SHARE-UPD=YES /SET-FILE-LINK DDCLOGR2,ADAyyyyy.CLOGR2,SHARE-UPD=YES /SET-FILE-LINK DDRLOGR1,ADAyyyyy.RLOGR1,SHARE-UPD=YES /SET-FILE-LINK TZINFO,ADAvrs.TZ00 /START-PROGRAM *M(ADA.MOD,ADARUN),PR-MO=ANY,RUN-MODE=A(ALT-LIB=Y) ADARUN PROG=ADANUC,DB=yyyyy ADARUN LBP=600000 ADARUN LWP=320000 ADARUN LS=80000 ADARUN LP=400 ADARUN NAB=24 ADARUN NC=50 ADARUN NH=2000 ADARUN NU=100 ADARUN TNAE=180,TNAA=180,TNAX=600,TT=90 ADARUN NPLOG=2,PLOGSIZE=1800,PLOGDEV=dddd ADARUN NCLOG=2,CLOGSIZE=1800,CLOGDEV=dddd /REMARK /EXIT-JOB MODE=NORMAL, SYS-OUT=NONE
/.ADANUC LOGON /OPTION MSG=FH,DUMP=YES /REMARK * /REMARK * ADABAS NUCLEUS /REMARK * /SYSFILE SYSLST=L.NUC /FILE ADAvrs.MOD ,LINK=DDLIB /FILE EXAMPLE.ADAyyyyy.ASSOR1,LINK=DDASSOR1,SHARUPD=YES /FILE EXAMPLE.ADAyyyyy.DATAR1,LINK=DDDATAR1,SHARUPD=YES /FILE EXAMPLE.ADAyyyyy.WORKR1,LINK=DDWORKR1 /FILE EXAMPLE.ADAyyyyy.PLOGR1,LINK=DDPLOGR1,SHARUPD=YES /FILE EXAMPLE.ADAyyyyy.PLOGR2,LINK=DDPLOGR2,SHARUPD=YES /FILE EXAMPLE.ADAyyyyy.CLOGR1,LINK=DDCLOGR1,SHARUPD=YES /FILE EXAMPLE.ADAyyyyy.CLOGR2,LINK=DDCLOGR2,SHARUPD=YES /FILE EXAMPLE.ADAyyyyy.RLOGR1,LINK=DDRLOGR1,SHARUPD=YES /FILE ADAvrs.TZ00 /EXEC (ADARUN,ADAvrs.MOD) ADARUN PROG=ADANUC,DB=yyyyy ADARUN LBP=600000 ADARUN LWP=320000 ADARUN LS=80000 ADARUN LP=400 ADARUN NAB=24 ADARUN NC=50 ADARUN NH=2000 ADARUN NU=100 ADARUN TNAE=180,TNAA=180,TNAX=600,TT=90 ADARUN NPLOG=2,PLOGSIZE=1800,PLOGDEV=dddd ADARUN NCLOG=2,CLOGSIZE=1800,CLOGDEV=dddd /REMARK /LOGOFF NOSPOOL
If you are using universal encoding support (UES), the following additional JCL is required for BS2000 environments:
/SET-FILE-LINK BLSLIB00,ADAvrs.MOD /SET-FILE-LINK BLSLIB01,APSvrs.LIB /CLP (APSvrs.LIB,LMSLINKLIB),(LNK-NAME=BLSLIB02) . . /SET-FILE-LINK DDECSOJ.ADAvrs.ALLECSO . /START-PROGRAM *M(ADAvrs.MOD,ADARUN),PR-M0=ANY,R-M=*A(ALT=Y,LOAD-INF=*REF,- / UNRES=*DELAY,MESSAGE=*ERROR) .
/FILE ADAvrs.ALLECSO,LINK=DDECSOJ
If you are connecting your UES-enabled database directly through a TCP/IP link, you must also:
Include the ADATCP library in the BLSLIB chain:
/SET-FILE-LINK BLSLIB03,DSN=WCPvrs.LIB /SET-FILE-LINK BLSLIB04,DSN=WTCvrs.LIB
Note:
These libraries are distributed with Entire Net-Work.
Set the ADARUN SUBMPSZ parameter as follows:
ADARUN SUBMPSZ=1M
In some exceptional circumstances, you may find it necessary to increase this size set by the SUBMPSZ parameter.
Include the following setting in the nucleus startup JCL:
/SET-FILE-LINK DDLNKPAR,link_parameter_file
where link_parameter_file identifies a permanent sequential file containing an ADALNK IDTNAME=idtname statement (where idtname is the same name as the one set by the ADARUN IDTNAME parameter.
The following notes apply to Adabas startup jobs in various platforms (as described by the notes).
This file is used to provide the Adabas session parameters.
This file is used to print messages produced by the control module ADARUN or the Adabas nucleus.
The Adabas Associator and Data Storage files. These files are mandatory.
n and nn represent the number of the Associator and Data Storage files, respectively.
If more than one file exists for Associator or Data Storage, a separate statement is required for each. For example, if the Associator consists of two files, FILE-LINK statements for DDASSOR1 and DDASSOR2 are required.
If less than 10 files exist for each, you must use the DDASSOR* and DDDATAR* link names in the JCL. For example, if the Associator consists of two files and Data Storage consists of three files, the following names would be used in the JCL: DDASSOR1, DDASSOR2, DDDATAR1, DDDATAR2, and DDDATAR3. If 10 or more files exist, the first nine must use the DDASSOR* and DDDATAR* link names in the JCL and the remainder must use the DDASSO* and DDDATA* link names in the JCL (dropping the "R" in the link names). For example, the tenth Associator file would be identified in the JCL using the name DDASSO10, while the third Associator file in the same JCL would be identified using the name DDASSOR3.
A maximum of 99 physical extents is now set for Associator and Data Storage files. However, your actual real maximum could be less because the extent descriptions of all Associator, Data Storage, and Data Storage Space Table (DSST) extents must fit into the general control blocks (GCBs). For example, on the default device type, there could be more than 75 Associator, Data Storage, and DSST extents each (or there could be more of one extent type if there are less for another).
So, the range of Associator link names that can be used in JCL is DDASSOR1 to DDASSOR9 and DDASSO10 to DDASSO99. In addition, the range of Data Storage DD names that can be used in JCL is DDDATAR1 to DDDATAR9 and DDDATA10 to DDDATA99.
The Adabas Work files. The WORKR1 file is mandatory. If you have Adabas Transaction Manager version 7.5 or later installed, an additional work file, WORKR4 is also mandatory.
Software AG recommends running the nucleus with the FILE-LINK option SUP=DISK(SHARE-UPD=YES) not specified for the WORKR1 file as a way of preventing two nuclei from writing to the same WORK file and corrupting the database. This could otherwise happen if the ADARUN parameters FORCE and IGNDIB are improperly used.
Work part 4 of WORKR1 is no longer supported if you have Adabas Transaction Manager Version 7.5 or later installed. Instead, you should use the WORKR4 file. WORKR4 is used for the same purpose as Work part 4, but it can be used in parallel by all members in a cluster. It is used to store the PET (preliminary end-of transaction) overflow transactions (those that cause a work overflow) of a database or of all members in a multiplex/SMP cluster.
The WORKR4 file is a container file that should be allocated and formatted in the normal way (use ADAFRM WORKFRM), using a block size greater than or equal to WORKR1. WORKR4 can have the same or a different device type than WORKR1. It should be at least as large as the cluster’s LP parameter of the database or cluster. The smaller WORKR1 Work part 1 is, the larger WORKR4 should be. This is because the nucleus must prevent a work overflow due to incomplete DTP transactions, but the nucleus must keep all PET transactions; they cannot be backed out.
If the Adabas Recovery Aid is being used, this logging file is required.
The data protection log file. This file is required if the database will be updated during the session and logging of protection information is desired. This file is not applicable if multiple protection logging is used.
The data protection log may be assigned to tape or disk. A new file must be used for each Adabas session (DISP=MOD may not be used). See Adabas Restart and Recovery for additional information.
Multiple (two to eight) data protection log files. These files are required only if multiple data protection logging is to be in effect for the session.
Multiple data protection logging is activated by the ADARUN NPLOG and PLOGSIZE parameters. The device type of the multiple protection logs is specified with the ADARUN PLOGDEV parameter.
Whenever one of multiple protection log files is full, Adabas switches automatically to another file and notifies the operator through a console message that the log which is full should be copied using the PLCOPY function of the ADARES utility. This copy procedure may also be implemented using the user exit 12 facility as described in the User Exits documentation.
If no command logging is to be performed, this file may be omitted.
The command log file. This file is required if command logging is to be performed during the session. Command logging is activated by the ADARUN LOGGING parameter.
Multiple (two to eight) command log files. These files are required only if multiple command logging is to be in effect for the session.
Multiple command logging is activated by the ADARUN NCLOG and CLOGSIZE parameters. The device type of the multiple command log files is specified with the ADARUN CLOGDEV parameter.
Whenever one of multiple command log files is full, Adabas switches automatically to another file and notifies the operator through a console message that the log which is full should be copied using the CLCOPY function of the ADARES utility. This copy procedure may also be implemented using the user exit 12 facility as described in the Adabas User Exits documentation.
The Entire Conversion Services (ECS) objects file is required for universal encoding support (UES) in all operating environments.
When running JCL for UES and TCP/IP support, a job will be spawned for each TCPURL defined.
If not declared, DDPRINT writes to SYSOUT . The subtasking automatically assigns a file with the following structure:
L.O.db_task_number.ADATCP_load_name.timestamp
The following substitutions are used in this syntax:
Substitution | Description |
---|---|
db_task_number | The task number of the UES TCP/IP database. |
ADATCP_load_name | The value "ADATCHPS" |
timestamp | A character representation of the time when the subtask was started. |
The TZINFO file is required when defining fields with the TZ option and using the user session open parameter TZ. This file contains information on local time offsets and daylight savings time offsets and its transition times.
The following data sets are required when executing an Adabas session under z/OS.
Data Set | DD Name | Storage Medium | Additional Information |
---|---|---|---|
ADARUN parameters | DDCARD | card image | note 1 |
ADARUN / Adabas messages | DDPRINT | printer | note 2 |
Associator | DDASSORn or DDASSOnn |
disk | note 3 |
Data Storage | DDDATARn or DDDATAnn |
disk | note 3 |
Work | DDWORKR1 DDWORKR4 |
disk | note 4 |
Recovery Aid log | DDRLOGR1 | disk | note 5 |
Protection log multiple log 1 multiple log 2 |
DDSIBA DDPLOGR1 DDPLOGR2 |
tape/disk disk disk |
note 6 note 7 note 7 |
Command log multiple log 1 multiple log 2 |
DDLOG DDCLOGR1 DDCLOGR2 |
tape/disk disk disk |
note 8 note 9 note 9 |
ECS encoding objects | DDECSOJ | disk | note 10 |
Abnormal termination | MPMDUMP | printer | note 11 |
Abnormal termination (if large buffer pools are used) | SVCDUMP | printer | note 12 |
SMGT dump and snap dump | ADASNAP | printer | note 13 |
SMGT print file | DDTRACE1 | printer | note 14 |
Time zone files | TZINFO | disk | note 15 |
This job includes multiple protection logging, multiple command logging, and Recovery Aid logging:
//NUC099 EXEC PGM=ADARUN //STEPLIB DD DISP=SHR,DSN=ADABAS.ADAvrs.LOAD //DDASSOR1 DD DISP=SHR,DSN=EXAMPLE.ADAyyyyy.ASSOR1 //DDDATAR1 DD DISP=SHR,DSN=EXAMPLE.ADAyyyyy.DATAR1 //DDWORKR1 DD DISP=OLD,DSN=EXAMPLE.ADAyyyyy.WORKR1 //DDPLOGR1 DD DISP=SHR,DSN=EXAMPLE.ADAyyyyy.PLOGR1 //DDPLOGR2 DD DISP=SHR,DSN=EXAMPLE.ADAyyyyy.PLOGR2 //DDCLOGR1 DD DISP=SHR,DSN=EXAMPLE.ADAyyyyy.CLOGR1 //DDCLOGR2 DD DISP=SHR,DSN=EXAMPLE.ADAyyyyy.CLOGR2 //DDRLOGR1 DD DISP=SHR,DSN=EXAMPLE.ADAyyyyy.RLOGR1 //TZINFO DD DISP=SHR,DSN=ADABAS.Vvrs.TZ00 //DDPRINT DD SYSOUT=X //DDTRACE1 DD SYSOUT=X //SYSUDUMP DD SYSOUT=X //MPMDUMP DD SYSOUT=X //ADASNAP DD SYSOUT=X //DDCARD DD * ADARUN PROG=ADANUC,DB=yyyyy ADARUN LBP=600000 ADARUN LWP=320000 ADARUN LS=80000 ADARUN LP=400 ADARUN NAB=24 ADARUN NC=1000 ADARUN NH=2000 ADARUN NU=100 ADARUN TNAE=180,TNAA=180,TNAX=600,TT=90 ADARUN NPLOG=2,PLOGSIZE=1800,PLOGDEV=dddd ADARUN NCLOG=2,CLOGSIZE=1800,CLOGDEV=dddd //
where:
dddd | is a valid device type. |
vrs | is the version of the product. |
yyyyy | is the physical database ID. |
A sample startup job for a UES-enabled database is provided in member ADANUCU in the ADAvrs.JOBS data set. If you are using universal encoding support (UES), you must
include the following additional libraries for internal products in the STEPLIB:
//STEPLIB DD DISP=SHR,DSN=ADABAS.APSvrs.LDnn // DD DISP=SHR,DSN=ADABAS.APSvrs.LD00
where nn is the load library level. If the library with a higher level number is not a full replacement for the lower level load library(s), the library with the higher level must precede those with lower numbers in the STEPLIB concatenation.
add the following additional JCL related to internal product libraries:
//DDECSOJ DD DISP=SHR,DSN=ADABAS.ADAvrs.EC00 //SYSPARM DD * SYSTEM_ID=ADAAPS ABEND_RECOVERY=NO THREAD_ABEND_RECOVERY=NO
If you are connecting your UES-enabled database directly through a TCP/IP link, you must also:
Include the ADATCP library in the STEPLIB:
//STEPLIB DD .... // DD DISP=SHR,DSN=WCPvrs.LOAD // DD DISP=SHR,DSN=WTCvrs.LOAD
Note:
These libraries are distributed with Entire Net-Work.
The dump produced by MPMDUMP may be too slow for users with very large buffer pools. You may instead elect to use the z/OS SVC dump facility to speed up nucleus dump processing. An SVC dump is triggered by the presence of an //SVCDUMP DD statement in the nucleus startup JCL.
If //SVCDUMP DD DUMMY
is specified, and the job is
running with APF authorization, a z/OS SVC dump is produced on the system dump
data set, normally SYS1.DUMPxx. If //SVCDUMP
DD DUMMY
is specified and the job is not running with APF
authorization, message ADAM77 is issued and dump
processing continues as if the SVCDUMP DD statement had not been
specified.
If //SVCDUMP DD DSN=dsn
is
specified (with an appropriate data set name), a z/OS SVC dump is produced on
the specified data set. Note that the SVCDUMP data set needs to be allocated
with DCB attributes RECFM=FB,LRECL=4160,BLKSIZE=4160
. Note also
that, for APF-authorized jobs, secondary extents are ignored.
For APF-authorized jobs, the SVC dump title consists of
"ADABAS System Dump" plus additional job name, DBID,
and timestamp information in the form: Job
jjjjjjjj DBID nnnnn
yyyy/mm/dd
hh:mm:ss.hhttmm
.
For non-authorized jobs, the title consists of "ADABAS Sys Tx
Dump" plus the additional job name, DBID, and timestamp
information. Note that the timestamp reflects the time at which the SVC dump
request was passed to z/OS -- not the time at which any preceding abend
occurred.
The shortest dump processing time occurs when you specify
//SVCDUMP DD DUMMY
. This is because the nucleus only needs to wait
for the dump to be captured, not written out to a dump data set. A
specification of //SVCDUMP DD DSN=dsn
will give you a shorter processing time for an APF-authorized job than for a
non-APF-authorized job. In both cases, the time taken for dump processing may
actually be longer than with MPMDUMP.
When an SVCDUMP DD statement is included in your JCL, any MPMDUMP DD statement is ignored, unless Adabas detects that it is unable to proceed with SVC dump processing.
If an error is encountered while writing the SVC dump, message ADAM78 appears. If dump writing completes successfully, message ADAM79 appears. For more information on these messages, refer to your Adabas messages and codes documentation.
No error message is produced when a dump can only be partially written. You should therefore ensure that sufficient space is available on the dump data set to accommodate the dump.
When an SVCDUMP DD statement is included in the JCL, but the SVC dump is unable to complete successfully, dump processing reverts to the standard dump options as specified in the JCL via the MPMDUMP, SYSUDUMP, SYSABEND or SYSMDUMP DD statements.
Note:
SVC dump processing might be suppressed due to installation SLIP or
DAE options. If dump processing is still required in this case, the relevant
MPMDUMP, SYSUDUMP, SYSABEND or SYSMDUMP DD statement should be specified in the
JCL in addition to the SVCDUMP DD statement.
When the SVCDUMP DD statement is omitted from the JCL, existing dump options, specified via the MPMDUMP, SYSUDUMP, SYSABEND or SYSMDUMP DD statements, continue to operate as normal.
The following notes apply to Adabas startup jobs in various platforms (as described by the notes).
This data set is used to provide the Adabas session parameters.
This data set is used to print messages produced by the control module ADARUN or the Adabas nucleus.
The Adabas Associator and Data Storage data sets. These data sets are mandatory.
n and nn represent the number of the Associator and Data Storage data set, respectively.
If more than one data set exists for Associator or Data Storage, a separate statement is required for each. For example, if the Associator consists of two data sets, DD statements for DDASSOR1 and DDASSOR2 are required.
If less than 10 data sets exist for each, you must use the DDASSOR* and DDDATAR* DD names in the JCL. For example, if the Associator consists of two data sets and Data Storage consists of three data sets, the following names would be used in the JCL: DDASSOR1, DDASSOR2, DDDATAR1, DDDATAR2, and DDDATAR3. If 10 or more data sets exist, the first nine must use the DDASSOR* and DDDATAR* DD names in the JCL and the remainder must use the DDASSO* and DDDATA* DD names in the JCL (dropping the "R" in the DD names). For example, the tenth Associator data set would be identified in the JCL using the name DDASSO10, while the third Associator data set in the same JCL would be identified using the name DDASSOR3.
A maximum of 99 physical extents is now set for Associator and Data Storage data sets. However, your actual real maximum could be less because the extent descriptions of all Associator, Data Storage, and Data Storage Space Table (DSST) extents must fit into the general control blocks (GCBs). For example, on a standard 3390 device type, there could be more than 75 Associator, Data Storage, and DSST extents each (or there could be more of one extent type if there are less for another).
So, the range of Associator DD names that can be used in JCL is DDASSOR1 to DDASSOR9 and DDASSO10 to DDASSO99. The range of Data Storage DD names that can be used in JCL is DDDATAR1 to DDDATAR9 and DDDATA10 to DDDATA99.
The Adabas Work data sets. The WORKR1 data set is mandatory. If you have Adabas Transaction Manager installed, an additional work data set, WORKR4 is also mandatory.
Software AG recommends running the nucleus with DISP=OLD for the WORKR1 data set as a way of preventing two nuclei from writing to the same WORK data set and corrupting the database. This could otherwise happen if the ADARUN parameters FORCE and IGNDIB are improperly used.
Work part 4 of WORKR1 is no longer supported if you have Adabas Transaction Manager installed. Instead, you should use the WORKR4 data set. WORKR4 is used for the same purpose as Work part 4, but it can be used in parallel by all members in a cluster. It is used to store the PET (preliminary end-of transaction) overflow transactions (those that cause a work overflow) of a database or of all members in a multiplex/SMP cluster.
The WORKR4 data set is a container data set that should be allocated and formatted in the normal way (use ADAFRM WORKFRM), using a block size greater than or equal to WORKR1. WORKR4 can have the same or a different device type than WORKR1. It should be at least as large as the cluster’s LP parameter of the database or cluster. The smaller WORKR1 Work part 1 is, the larger WORKR4 should be. This is because the nucleus must prevent a work overflow due to incomplete DTP transactions, but the nucleus must keep all PET transactions; they cannot be backed out.
If the Adabas Recovery Aid is being used, this logging data set is required.
The data protection log data set. This data set is required if the database will be updated during the session and logging of protection information is desired. This data set is not applicable if multiple protection logging is used.
The data protection log may be assigned to tape or disk. A new data set must be used for each Adabas session (DISP=MOD may not be used). See Adabas Restart and Recovery for additional information.
Multiple (two to eight) data protection log data sets. These data sets are required only if multiple data protection logging is to be in effect for the session.
Multiple data protection logging is activated by the ADARUN NPLOG and PLOGSIZE parameters. The device type of the multiple protection logs is specified with the ADARUN PLOGDEV parameter.
Whenever one of multiple protection log data sets is full, Adabas switches automatically to another data set and notifies the operator through a console message that the log which is full should be copied using the PLCOPY function of the ADARES utility. This copy procedure may also be implemented using the user exit 12 facility as described in the User Exits documentation.
If no command logging is to be performed, this data set may be omitted.
The command log data set. This data set is required if command logging is to be performed during the session. Command logging is activated by the ADARUN LOGGING parameter.
Multiple (two to eight) command log data sets. These data sets are required only if multiple command logging is to be in effect for the session.
Multiple command logging is activated by the ADARUN NCLOG and CLOGSIZE parameters. The device type of the multiple command log data sets is specified with the ADARUN CLOGDEV parameter.
Whenever one of multiple command log data sets is full, Adabas switches automatically to another data set and notifies the operator through a console message that the log which is full should be copied using the CLCOPY function of the ADARES utility. This copy procedure may also be implemented using the user exit 12 facility as described in the Adabas User Exits documentation.
The Entire Conversion Services (ECS) objects data set is required for universal encoding support (UES) in all operating environments.
This data set is used to take an Adabas dump including SVC, ID-TABLE and allocated CSA in the event that an abnormal termination occurs.
The line count in the JCL must be set appropriately; otherwise, the dump cannot be printed in its entirety.
The z/OS SVC dump facility can be used to speed up nucleus dump processing if an MPMDUMP is too slow for users with very large buffer pools. We recommend that you specify both an MPMDUMP and an SVCDUMP DD statement in your JCL to ensure that one of the dumps is produced when needed. If an SVCDUMP DD statement is included, an SVC dump is created if possible and the MPMDUMP DD statement is ignored. Should problems arise during processing of the SVC dump, an MPMDUMP will be taken, but only if the MPMDUMP DD statement is also specified in the JCL. For more information, read Using the z/OS SVC Dump Facility.
This data set is used under z/OS to take an Adabas dump (SMGT,DUMP) or snap dump (SMGT,SNAP) when using the error handling and message buffering facility.
This data set is used when the PIN output is to be directed to DDTRACE1 rather than DDPRINT as specified by the user in ADASMXIT when using the error handling facility. If DDTRACE1 has not been specified in the JCL and PRINTDD is set to "NO" in ADASMXIT, output will be lost. The PIN output will not be written to DDPRINT unless "YES" is specified.
The TZINFO data set is required when defining fields with the TZ option and using the user session open parameter TZ. This data set is a library or partioned data set containing information on local time offsets and daylight savings time offsets and its transition times.
The following table contains all data sets used when executing an Adabas session under z/VSE. SYSnnn means that any programmer logical unit may be used.
Data Set | File Name | Logical Unit | Storage Medium | Additional Information |
---|---|---|---|---|
ADARUN parameters | none CARD CARD |
SYSRDR SYS000 SYSnnn |
reader tape disk |
note 1 |
ADARUN/Adabas messages | none | SYSLST | printer | note 2 |
Associator | ASSORn or ASSOnn |
SYSnnn | disk | note 3 |
Data Storage | DATARn or DATAnn |
SYSnnn | disk | note 3 |
Work | WORKR1 | SYSnnn | disk | note 4 |
Recovery Aid log | RLOGR1 | SYSnnn | disk | note 5 |
Protection log multiple log 1 multiple log 2 |
SIBA PLOGR1 PLOGR2 |
SYS014 SYSnnn SYSnnn SYSnnn |
tape disk disk disk |
note 6 note 7 note 7 |
Command log multiple log 1 multiple log 2 |
LOG CLOGR1 CLOGR2 |
SYS012 SYSnnn SYSnnn SYSnnn |
tape disk disk disk |
note 8 note 9 note 9 |
ECS encoding objects | DDECSOJ | * | disk | note 10 |
Time zone files | TZINFO | not required | disk | note 11 |
This job includes multiple protection logging, multiple command logging, and Recovery Aid logging:
// ASSGN SYS031,dddd,VOL=ADA001,SHR // ASSGN SYS032,dddd,VOL=ADA002,SHR // ASSGN SYS033,dddd,DISK,VOL=ADA003,SHR // ASSGN SYS034,dddd,VOL=ADA004,SHR // DLBL ASSOR1,'EXAMPLE.ADAyyyyy.ASSOR1',2099/365,DA // EXTENT SYS031,ADA001,,,15,1500 // DLBL DATAR1,'EXAMPLE.ADAyyyyy.DATAR1',2099/365,DA // EXTENT SYS032,ADA002,,,15,3000 // DLBL WORKR1,'EXAMPLE.ADAyyyyy.WORKR1',2099/365,DA // EXTENT SYS033,ADA003,,,15,600 // DLBL PLOGR1,'EXAMPLE.ADAyyyyy.PLOGR1',2099/365,DA // EXTENT SYS034,ADA004,,,15,600 // DLBL PLOGR2,'EXAMPLE.ADAyyyyy.PLOGR2',2099/365,DA // EXTENT SYS034,ADA004,,,615,600 // DLBL CLOGR1,'EXAMPLE.ADAyyyyy.CLOGR1',2099/365,DA // EXTENT SYS034,ADA004,,,1215,600 // DLBL CLOGR2,'EXAMPLE.ADAyyyyy.CLOGR2',2099/365,DA // EXTENT SYS034,ADA004,,,1815,600 // DLBL RLOGR1,'EXAMPLE.ADAyyyyy.RLOGR1',2099/365,DA // EXTENT SYS034,ADA004,,,1300,600 // DLBL ADALIB,'ADABAS.ADAvrs.LOADLIB',2099/365 // EXTENT ,ADADSK ADARUN PROG=ADANUC,SVC=xxx,DEVICE=dddd,DB=yyyyy ADARUN LBP=600000 ADARUN LWP=320000 ADARUN LS=80000 ADARUN LP=400 ADARUN NAB=24 ADARUN NC=1000 ADARUN NH=2000 ADARUN NU=100 ADARUN TNAE=180,TNAA=180,TNAX=600,TT=90 ADARUN NPLOG=2,PLOGSIZE=1800,PLOGDEV=dddd ADARUN NCLOG=2,CLOGSIZE=1800,CLOGDEV=dddd /* /*
The following additional JCL is required for universal encoding support (UES):
// ASSGN SYS020,disk,VOL=volume,SHR // DLBL DDECSOJ,'ADABAS.ADAvrs.ECSLIB' // EXTENT SYS020
To support both UES and TCP/IP in z/VSE environments, you only need to add the additional JCL described in Installing TCP/IP Support for the Adabas Nucleus.
The following additional JCL is required for universal encoding support (UES):
// DLBL TZINFO,'/ADALIB/ADAvrsTZ/.TIMEZONE'
See note 11 below for more information.
The following notes apply to Adabas startup jobs in various platforms (as described by the notes).
This data set is used to provide the Adabas session parameters.
This data set is used to print messages produced by the control module ADARUN or the Adabas nucleus.
The Adabas Associator and Data Storage data sets. These data sets are mandatory.
n and nn represent the number of the Associator and Data Storage data set, respectively.
If more than one data set exists for Associator or Data Storage, a separate statement is required for each. For example, if the Associator consists of two data sets, DLBL statements for ASSOR1 and ASSOR2 are required.
If less than 10 data sets exist for each, you must use the ASSOR* and DATAR* DLBL names in the JCL. For example, if the Associator consists of two data sets and Data Storage consists of three data sets, the following names would be used in the JCL: ASSOR1, ASSOR2, DATAR1, DATAR2, and DATAR3. If 10 or more data sets exist, the first nine must use the ASSOR* and DATAR* DLBL names in the JCL and the remainder must use the ASSO* and DATA* DLBL names in the JCL (dropping the "R" in the DLBL names). For example, the tenth Associator data set would be identified in the JCL using the name ASSO10, while the third Associator data set in the same JCL would be identified using the name ASSOR3.
A maximum of 99 physical extents is now set for Associator and Data Storage data sets. However, your actual real maximum could be less because the extent descriptions of all Associator, Data Storage, and Data Storage Space Table (DSST) extents must fit into the general control blocks (GCBs). For example, on a standard 3390 device type, there could be more than 75 Associator, Data Storage, and DSST extents each (or there could be more of one extent type if there are less for another).
So, the range of Associator DLBL names that can be used in JCL is ASSOR1 to ASSOR9 and ASSO10 to ASSO99. The range of Data Storage DLBL names that can be used in JCL is DATAR1 to DATAR9 and DATA10 to DATA99.
The Adabas Work data sets. The WORKR1 data set is mandatory. If you have Adabas Transaction Manager installed, an additional work data set, WORKR4 is also mandatory.
Work part 4 of WORKR1 is no longer supported if you have Adabas Transaction Manager installed. Instead, you should use the WORKR4 data set. WORKR4 is used for the same purpose as Work part 4, but it can be used in parallel by all members in a cluster. It is used to store the PET (preliminary end-of transaction) overflow transactions (those that cause a work overflow) of a database or of all members in a multiplex/SMP cluster.
The WORKR4 data set is a container data set that should be allocated and formatted in the normal way (use ADAFRM WORKFRM), using a block size greater than or equal to WORKR1. WORKR4 can have the same or a different device type than WORKR1. It should be at least as large as the cluster’s LP parameter of the database or cluster. The smaller WORKR1 Work part 1 is, the larger WORKR4 should be. This is because the nucleus must prevent a work overflow due to incomplete DTP transactions, but the nucleus must keep all PET transactions; they cannot be backed out.
If the Adabas Recovery Aid is being used, this logging file is required.
The data protection log data set. This data set is required if the database will be updated during the session and logging of protection information is desired. This data set is not applicable if multiple protection logging is used.
The data protection log may be assigned to tape or disk. A new data set must be used for each Adabas session (DISP=MOD may not be used). See Adabas Restart and Recovery for additional information.
Multiple (two to eight) data protection log data sets. These data sets are required only if multiple data protection logging is to be in effect for the session.
Multiple data protection logging is activated by the ADARUN NPLOG and PLOGSIZE parameters. The device type of the multiple protection logs is specified with the ADARUN PLOGDEV parameter.
Whenever one of multiple protection log data sets is full, Adabas switches automatically to another data set and notifies the operator through a console message that the log which is full should be copied using the PLCOPY function of the ADARES utility. This copy procedure may also be implemented using the user exit 12 facility as described in the User Exits documentation.
If no command logging is to be performed, this data set may be omitted.
The command log data set. This data set is required if command logging is to be performed during the session. Command logging is activated by the ADARUN LOGGING parameter.
Multiple (two to eight) command log data sets. These data sets are required only if multiple command logging is to be in effect for the session.
Multiple command logging is activated by the ADARUN NCLOG and CLOGSIZE parameters. The device type of the multiple command log data sets is specified with the ADARUN CLOGDEV parameter.
Whenever one of multiple command log data sets is full, Adabas switches automatically to another data set and notifies the operator through a console message that the log which is full should be copied using the CLCOPY function of the ADARES utility. This copy procedure may also be implemented using the user exit 12 facility as described in the Adabas User Exits documentation.
The Entire Conversion Services (ECS) objects data set is required for universal encoding support (UES).
The TZINFO DLBL is required when defining fields with the
TZ option and using the user session open parameter TZ. This data set is a
library containing information on local time offsets and daylight savings time
offsets and its transition times. The TZINFO DLBL defines the library and
sublibrary to be used for accessing time zone data. The physical name
associated with the DLBL must be in the form
'library/sublib/.TIMEZONE'
.
in the example above, time zone
data is accessed from the ADALIB sublibrary
"ADAvrsTZ". In this case,
a label for ADALIB must be available to the Adabas session job.
Adabas uses operating system services to synchronize the start and end of nucleus and utility executions. Only one program can modify the data integrity block (DIB) at a time.
The operating system services used are as follows:
Operating System | Service |
---|---|
z/OS | systems-wide ENQ/DEQ macros (SCOPE=SYSTEMS) with major name (QNAME) 'ADABAS' |
BS2000 | system-wide ENQAR/DEQAR macros (SCOPE=GLOBAL) |
z/VSE | system-wide LOCK/UNLOCK macros |
This feature reliably and efficiently guarantees proper synchronization of DIB updates within a single operating-system image.
If your database resides on disks that are shared among multiple images of the operating system and you run nucleus or utility jobs against the same database on more than one of the system images, you need to ensure that
the system images are installed in such a way that synchronization is effective on all systems where nucleus and utility jobs execute; or
nucleus and utility jobs do not execute concurrently on different system images.
Consult your system programmer for the needed information.
Warning: If different nucleus or utility jobs updating the same file are allowed to start or terminate on different system images at the same time without proper synchronization, a DIB update may be lost. If this happens, a lock in the DIB may be violated, thereby opening the file to the possibility of destruction due to concurrent unsynchronized updates by utilities. |
Although the normal mode of operation is multiuser mode, it is also possible to execute Adabas together with a user program or Adabas utility in the same region.
For single-user mode, you must include the Adabas nucleus job control that you use along with the job control for the utility or user program.
The Adabas prefetch option cannot be used in single-user mode; however, single-user mode must be used when running a read-only nucleus and an update nucleus simultaneously.
Some information within an Adabas database is user-related and must be retained from session to session. One such kind of information is ET data records; another is the priority value assigned to a user.
A set of user-related information can be stored in a profile table. The values stored in this table are read at OPEN time and assigned to the user. The direct call user must OPEN the Adabas session with the proper call; that is, as an ID user with an ETID in the Additions 1 field of the Adabas control block. For Natural users, the profile table is identified by the Natural ETID.
The associated fields are user-related timeout and threshold values, and the OWNERID for multiclient fields. One record per user is stored. The profile table is maintained using Adabas Online System.
The user-related values shown below are currently stored in the profile table.
Value | Description |
---|---|
PRIORITY | User's priority (0-255) |
TNAA* | Access user non-activity time |
TNAE* | ET user non-activity time |
TNAX* | EXU/EXF user non-activity time |
TT* | Transaction time threshold |
TLSCMD* | Sx command threshold |
NSISN* | Maximum number of ISNs per TBI element |
NSISNHQ* | Maximum number of records held by user |
NQCID* | Maximum number of active command IDs per user |
OWNERID | Owner ID for multiclient file access |
* The decision sequence for determining the values for a user at the time of an open call is shown in Managing the User Profile.
Adabas Online System (AOS) must be used to maintain the profile table. See the Adabas Online System documentation for detailed information about managing the profile table.