This document describes what Entire Output Management can do and how it works. It covers the following topics:
When you start using Entire Output Management, you need not change your present system of print-data handling all at once. You can gradually transfer the processing of your print data step by step to Entire Output Management. This allows you a smooth transition from you present print-data handling to using the full range of possibilities for the processing of print data offered by Entire Output Management.
For print data to be processed by Entire Output Management, the applications producing these data need not be changed.
Entire Output Management is an output management system which is used for the processing and distribution of print data. Entire Output Management can process any kind of print data in heterogeneous environments. The print data can come from a variety of data sources:
print data produced by applications,
data from spooling systems,
data stored in files.
The print data actually processed by Entire Output Management are called reports. In the definition of a report, you select which print data from a data source are to be processed, and you determine how they are to be processed and distributed.
A report can be:
sent to a printer for printing,
sent to a file,
sent to a follow-up process for further processing.
For the sake of convenience, all these "output devices" are generally referred to as printers within Entire Output Management.
In addition, Entire Output Management can be used to archive the print data by storing the reports in sequential files.
A report definition consists of various attributes. By specifying these attributes, you determine:
the data source of the print data to be used for the report,
which print data are to be extracted from the data source,
the Entire Output Management users who are to receive the report for further processing,
how and on which printer the report is to be printed,
if and how the report is to be archived.
The actual report which Entire Output Management creates from the report definition is called an active report.
Reports can be combined to form larger packets, which are called bundles. This bundling is possible even if the reports come from different data sources. The reports grouped in a bundle are processed together as a unit.
In a bundle definition, you specify which reports are to be part of the bundle. In addition, you can define a number of other bundle attributes to control the processing of the bundle.
When an active report assigned to a bundle is processed, Entire Output Management creates an active bundle based on the bundle definition.
You can assign users to a report. These users can display and review the active report on his/her screen, export it (for example, to a PC) for further processing, and print it.
Every Entire Output Management user has a folder named
#Inbasket
. By default, an active report assigned to a user appears
in his/her #Inbasket
folder.
In addition, you can define public folders to which all Entire Output Management users have access, and distribute active reports to these.
Users can also define their individual private folders and file active reports in them.
To make the distribution of reports to various users easier, you can create distribution lists. A distribution list can contain individual users, but you can also have distribution lists within distribution lists. Instead of assigning a report to multiple users, you assign it to a distribution list. It will then be distributed to all members of the distribution list.
The printing of an active report can be triggered either automatically depending on attributes specified in the report definition, or manually by a user.
In the report definition, you specify on which printer the active report is to be printed. A printer assigned to a report is called a logical printer. In the definition of the logical printer you specify a set of attributes linked to an actual physical printer, which determine the printing characteristics and printing format of the report on the physical printer.
If all output to be printed on a physical printer is to be printed in the same way, you only need to define one logical printer referring to the physical printer. If you wish to print different reports differently on the same physical printer, you define multiple logical printers which refer to the same physical printer, but with different printing characteristics.
You can determine how long an active report is to be available online to users in Entire Output Management. You do this by defining a retention period in the report definition. When this period expires, the active report is longer available online.
If you wish to keep an active report, you can archive it on a sequential file. In the report definition, you specify how long an active report is to be kept in the archive.
If you need an active report online again, you can revive it; that is, you can retrieve it from the archive and copy it to the online system.
To reduce the storage space for archive data on the sequential files to a minimum, Entire Output Management performs so-called condense jobs. These jobs delete from the archive all active reports whose archiving period has expired.
With the function Automatic Archiving Defaults you can define a threshold for the number of reports in an archive file that will trigger its being marked for condensation.
This section provides an overview of the various cleanup functions provided by Entire Output Management. It covers the following topics:
There are three types of cleanup:
daily cleanup,
source cleanup,
report cleanup.
The primary cleanup type is daily cleanup. Entire Output Management performs its daily cleanup once per day on its next activation after the time specified in the system defaults (option 8.1.1). If no time is specified there, the daily cleanup will be done on the first activation after midnight.
Source cleanup and report cleanup are optional and can be activated in option 8.1.7. Each can be activated individually and a schedule defined for it. They may be scheduled to run whenever you wish, using a calendar to specify on which days of the week or month they are to run, whether they should run before or after non-working days, the time window during which they are to run and how often. For example, you could define that source/report cleanup is to run every 2 hours between 08:00 and 20:00 on Mondays, Wednesdays and Fridays.
The daily cleanup processes all expired entities:
Active reports that have expired are deleted or marked for archiving, as appropriate.
Active bundles, printouts and log records that have expired are deleted.
Revived reports are “unrevived”, that is, the contents are deleted if the report was revived to the database and the active report is reset to its archived state.
Archived active reports are deleted.
Active report sources (spool files or container file entries) are checked and, if no longer needed, deleted.
Source cleanup checks Entire Output Management sources (for example, JES or Power spool files in Entire Output Management’s temporary classes or entries in the container files) to see if they are still needed - and deletes them if they are not needed. They are no longer needed when there is no active report with location "S" referring to them. Container file entries which are no longer needed are deleted even if Spool Cleanup is set to "N".
Report cleanup checks all active reports with location "S" to establish if the source file is still available. If it is not, the active report is deleted.
The daily cleanup process may entail a lot of work (typically it can run between 30 and 60 minutes in a production environment), and during this time the monitor can do nothing else; this means that no reports, bundles or printouts can be processed until the daily cleanup has finished. This is because the monitor is a single task.
To avoid this problem, the daily cleanup can run in an
asynchronous batch job. To do this, you execute the program
NOMCLEAN
in a standard batch-mode Natural (with the necessary
parameters set, such as LFILE 206). This batch job can then be activated by a
job scheduler to run a couple of hours before the Entire Output Management
monitor is scheduled to perform the daily cleanup. When the monitor is about to
perform the daily cleanup, it can ascertain that NOMCLEAN
has ran
and so it will not have to do anything.
NOMCLEAN
and the monitor keep each other informed of
the current status by using two control records in the Entire Output Management
system file. These control records are accessed using the view
NRM-ARCHIVE-TASK
and the descriptor M-MONITOR-ID =
'CLEANUP'
or M-MONITOR-ID = SYSNOM
(actually the library
Entire Output Management is running from).
The communication paths are as follows:
NOMCLEAN
at its start sets SYSNOM’s
CLEANUP-IDENTIFIER
to CLEAN
,
CLEANUP-ACTIVE-NOW
to TRUE
and
CLEANUP-RESCHEDULE
to FALSE
to show that it is
active.
RMONITOR (monitor main loop) only deletes no longer needed
sources if SYSNOM’s CLEANUP-IDENTIFIER
is not CLEAN
or CLEANUP-ACTIVE-NOW
is not TRUE
(that is, only if
NOMCLEAN
is not currently active). In addition, to avoid
concurrency problems while creating active reports, RMONITOR sets CLEANUP’s
REPORT-PROCESSING-NOW
to TRUE
while creating active
reports, and to FALSE
when it has finished. This is necessary to
prevent NOMCLEAN
from deleting reports or sources that are being
worked on by the monitor.
RMARCSCH (monitor scheduled functions) only performs daily
cleanup if SYSNOM’s CLEANUP-RESCHEDULE
and
CLEANUP-ACTIVE-NOW
are both FALSE
(that is, only if
NOMCLEAN
is not currently active and has not already run on that
day). If CLEANUP-RESCHEDULE
is TRUE
, this means that
NOMCLEAN
has already run so RMARCSCH resets
CLEANUP-RESCHEDULE
to FALSE
and updates the next
scheduled daily cleanup to the same time on the next day. RMARCSCH only
performs source/report cleanup if SYSNOM’s CLEANUP-ACTIVE-NOW
is
FALSE
. This is to avoid collision with NOMCLEAN
which
may also be cleaning up sources.
RMSRCK (source/report cleanup) may be invoked by the monitor
or NOMCLEAN
. When invoked by NOMCLEAN
, it needs to
avoid the concurrency problem described above. It does this by checking
REPORT-PROCESSING-NOW
and, if it is TRUE
, uses the
NPR EVENTING view to make itself dormant (repeatedly, for 5 seconds at a time)
until REPORT-PROCESSING-NOW
is FALSE
, which is when
RMSRCK resumes processing where it left off.
If multi-tasking is active (more than one monitor task is defined), the monitor cleanup functions will be performed by the main task, task 1. This will allow cleanup to be separated from other monitor functions like active report and printout creation.
Entire Output Management provides methods for accessing printers: It is able to directly print to TCP/IP printers. This print method is intended to access printers that have either got a dedicated IP address or can be reached using printer queue names of printer servers.
This print method is intended for direct printing of shorter documents directly from online applications. However, it is executed in a Natural subtask and no online space (such as CICS memory) is affected. No spooling system is required, and no extra address space will be used, not even an ESY server. This makes this print method independent from batch jobs, Broker, CA-SPOOL, ESY, JES, and Natural Advanced Facilities.
Performance enhancements are achieved, and 4 MB of space in an Entire Output Management printer subtask per 1000 pages of text output are needed as soon as the printer (server) acknowledges the reception of the printout.
At print time, the printout will be initiated online or by the monitor and executed in one of the defined print tasks of Entire Output Management. At execution time, the printout will be kept in virtual memory and then sent to TCP/IP directly (via socket programming) using the LPR/LPD protocol (RFC1179).
Entire Output Management also allows you to process binary data. This means that any kind of file or printout can be kept in Entire Output Management, archived, printed, distributed, and passed to a destination system.
This is possible with UNIX and Windows file systems, but not with mainframe spool systems.
Note:
If binary data are read from a mainframe source, subsequent
processing may cause Natural abends like IW060
or
0C7
. On mainframes, only text data can be processed. For
performance reasons, however, no check is performed to ensure that the input
data do not contain binary data.
Basically there are two ways to transfer binary data to Entire Output Management:
You can use the direct-input interface together with the Open Print Option (OPO). A Windows client passes output of the Windows systems to Entire Output Management. This can be the output of a Windows printer driver or any program that can forward data to OPO using the pipe mechanism of Windows. The output is redirected to Entire Output Management using the Remote Procedure Call (RPC) facility of EntireX.
You can define a report that receives binary data from a UNIX or Windows directory. In the report definition, the parameter "binary reading" must be set to "B" to indicate that the file is to be processed in binary format.
The report is passed through the Entire Output Management trigger queue, similar to the NOMPUT interface. Therefore the trigger queue must be activated (Trigger Container File). The opened active report receives the type "binary" (a special CC type).
A binary document cannot be browsed with the mainframe character interface. However, the Entire Output Management GUI Client is able to receive the data, store them on the PC (please note that the whole file has to be transferred) and browse (view) them with the Windows default application of the specific file type (such as ".doc" or ".pdf").
Binary active reports can be printed with UNIXLP printers or on BS2000 with a SYSPRBS2 printer type. Under UNIX, a printer type NATUNIX is available which can pass the data to the "lp" or "lpr" utility (in which case the option "-l" has to be used to ensure that the data are forwarded in a transparent mode) or to a program or script, or to a file in a UNIX directory.
Please note that a binary report that is an output of a Windows printer driver will be bound to the hardware it was created for. The decision where to print the data Entire Output Management may have received a long time ago is made at creation time. Consider to use file formats like PDF if you want to keep binary data for a long time.
Entire Output Management converts binary data to the BASE64 format. This makes it possible to pass the data across platforms and code pages, because BASE64 consists of printable characters only. However, this also means that the data are larger than the original binary data stream. When active reports are output to destinations, the BASE64 data stream is decoded to binary format again.
The output of printer drivers often result in large data streams, because a printer driver creates print pages with all graphical statements of the appropriate printer language.
In theory, separation exits can be used. However, binary data cannot be separated with standard exits, as they are not readable.
Any report Entire Output Management receives via the OPO can be
enriched with user-defined meta data. The meta data can be viewed via
PF2 on the spooling attributes screen of an active report, or in any
user exit that uses the field SPOOL-ATTRIBUTES-EXTENDED
. Meta-data
values that exceed the length of the terminal line size of the Entire Output
Management character interface will be truncated to one line and marked with
the continuation marker "...". Meta data are passed to Entire Output Management
via an XML file that can be defined in an OPO printer definition.
For further details and examples, see the section Setting Up Environments for Binary Documents in the System Administration documentation.
For systems which run on multiple nodes in heterogeneous environments, please observe the following recommendations concerning the use of code pages.
Always ensure that the code page used complies with you terminal emulation and system software.
Recommended code page for IBM systems: IBM01140
on
English systems, IBM01140
or IBM01141
on German
systems.
Recommended code page for BS2000: EDF03DRV
.
The code page used has to be specified with the Natural profile
parameter CP
.
Recommended code page for Windows: WINDOWS-1252
.
Recommended code page for UNIX: ISO-8859-15
.
These are recommended for English and Germans systems, as they both provide English characters, German umlauts and the Euro sign.
Define the code page used:
in you Windows/UNIX operating system;
in Entire System Server for UNIX in the file
npr.ini
as, e.g.: Locale_String=ISO-8859-15
;
in the Natural parameter module, e.g.:
CP=ISO-8859-15
;
in the Natural RPC parameter module, e.g.:
CPRPC=ISO-8859-15,CP=ISO-8859-15
;
in the Broker attribute file, e.g.:
DEFAULTS=CODEPAGE DEFAULT_ASCII=ISO-8859-15 DEFAULTS=SERVICE CONVERSION=SAGTCHA *ICU* CLASS=NPR, SERVER=*, SERVICE=*, DEFERRED=NO, APPMON=NO DEFAULTS=SERVICE CONVERSION=SAGTRPC *ICU* CLASS=RPC, SERVER=*, SERVICE=*, DEFERRED=NO, APPMON=NO
For the definition of nodes and code pages within Entire Output Management, see Default Code Pages and Node Definitions in the System Administration documentation.
Entire Output Management can handle text output from mainframe operating systems and from UNIX or Windows systems. This output has to be stored as ASCII or EBCDIC files, on mainframes with or without control characters.
Basically, Entire Output Management on UNIX handles the same output formats as on mainframes unless binary data are to be processed. Binary data have to be converted into the ASCII data format before being processed by Entire Output Management.
To get output data from UNIX or Windows directories, use the Node Definitions described in the System Administration documentation.
UNIX applications can print into a Entire Output Management file
directory (ASCII files) or access spoolers like CUPS to print output to the
queue of a virtual printer. In this case, the printer queue of CUPS has to be
connected to a CUPS backend like "pipe" (default:
/usr/lib/cups/backend
) which stores the data as ASCII files in the
Entire Output Management directory defined in the node definitions.
However, some interfaces do not exist under UNIX and can therefore not be used in an Entire Output Management UNIX environment:
CA Spool
NOMPUT
Under UNIX, Entire Output Management is designed to only print on the printer types NATUNIX and DISKUNIX. They are described under Special Printer Types in the System Administration documentation.
Instead of printing output from Natural applications in a spooling system, you can route it directly to an Entire Output Management container file, without having to create any intermediate files in a file system.
See Printing from Natural to Entire Output Management Directly in the Installation and Customization on Mainframes documentation.
Direct print from Natural on UNIX is only possible on Linux systems. As a prerequisite, the Open Print Option (OPO) has to be installed on Linux.
To use direct printing from Natural on UNIX, you have to do the following:
With the Natural Configuration utility, assign a Natural report
to a printer of profile type NOM
, as described in the
Natural Configuration Utility documentation under
Overview of Profile Parameters > Natural Execution Configuration
> Device/Report Assignments.
With the Natural Configuration utility, attach a printer profile
of the method type NOM
, as described in the Natural
Configuration Utility documentation under Overview of
Configuration File Parameters > Global Configuration File > Printer
Profiles > NOM Printer Profiles.
Customize the configuration file nomrptConf.xml
, as
described in the Installing the Open Print Option
documentation.
This enables Natural on UNIX to call OPO, and OPO will transfer the Natural print data to an Entire Output Management configuration file on UNIX or on a mainframe. From the container file, the print data can then be further processed by Entire Output Management.
It is possible to convert the output of a report to common multimedia file formats when it is printed. The printer type DISKUNIX is available to convert a file after it has been written to disk. As a prerequisite the report to be converted must be stored in Entire Output Management in the format "text" (ASCII, ASA, machine code), PDF or PostScript (or mixed in bundles). The DISKUNIX printer will convert text reports to the desired format according to the format definitions. Reports that are already in PDF or PostScript format will not be formatted because they are already rendered. PDF and PostScript formatted reports will only be converted to the desired output format without applying the Enscript parameters (see Formatting Attributes for File-Format Conversion under DISKUNIX Printers in the System Administration documentation). If separator pages are defined, however, they will follow the format definitions as they are text portions of the report. If bundles are to be printed DISKUNIX will collect all reports, regardless if they are in text, PDF, or PostScript format and put them into a single file of the desired output format. All text portions (text reports, separator pages) will be converted as defined in the format parameters. DISKUNIX uses the utilities Ghostscript and Enscript to perform the conversion; these haves to be installed on the destination UNIX system to which DISKUNIX writes.
IF UTF-8 is used as the file code page for reading and printing, you can use the UNIX utility Uniprint as the renderer instead of Enscript. Uniprint is part of the UNIX package "yudit" (Unicode editor), which also has to be installed on the node on which the conversion is done.
The conversion is triggered by specifying the attribute field Output Format.
PDF formatted reports can use a mask file for every page. This mask file is a PDF file with transparent background. This file is treated as a stamp on each page: Only the parts of the mask file which are transparent will show the original report. This means that logos or forms can be integrated into a PDF report. If the mask file contains more than one page, the corresponding pages of the original report will be overlaid.
To use the overlay function in Entire Output Management, the package "pdftk" (PDF Toolkit) has to be installed on the converting node. It is available for Windows and Linux machines. If the field Command is filled, this command line can be used to further process the resulting file after the conversion For example, the utilities "ps2afp" or "pdf2afp" (contained in the IBM product Infoprint) can be used to finally create AFP-formatted files for further processing.
The above-mentioned UNIX utilities are third-party products which are not part of Entire Output Management; Software AG neither delivers them nor provides support for them.
Entire Output Management transfers data from and to Entire System Server on UNIX and the Open Print Option (OPO). If the transfer is performed remotely, the code pages of the source and destination machines will differ in most cases. This means that you have to implement code-page translation using one or more middleware products. As a result, you may end up with complex environments in which special characters are still not always translated to the desired characters on the destination machine.
To avoid this problem, Entire Output Management, Entire System Server and OPO send named characters instead of the characters themselves. Entire Output Management uses the trigraph feature of Entire System Server and HTML named characters with OPO. This means that only the ASCII 7-bit and basis EBCDIC tables are used for the conversion of characters across platforms.
Entire Output Management uses the following Adabas files:
the definition-data file (logical file number 206),
the active-data file (logical file number 91),
one or more container files,
the trigger container file.
You can use the same Adabas file for both the definition-data file and the active-data file (however, the logical file specifications for both 206 and 91 are necessary).
Entire Output Management can copy print data from their original source (for example, JES spool) into a container file or into the active-data file, as described below.
For information on the trigger container file, see Trigger Container File in the System Administration documentation.
Entire Output Management stores output in arrays of 16 KB. Adabas compresses null values and trailing blanks in alphanumeric fields. Text data of output often contain blocks of blanks to present the data in the form of lists or tables. These blocks of blanks are compressed by Entire Output Management itself. This results in a more efficient use of Adabas database space and fewer databases accesses, thus improving performance.
Entire Output Management copies report sources into a container file under any of the following circumstances:
If Entire Output Management on UNIX is used.
If the print data come from CA Spool, Natural Advanced
Facilities or the 3GL interface (including the VTAM virtual-printer application
NOMVPRNT
with the parameter STORE=DB
).
In BS2000, if the option Copy files is set to "Y" in the Monitor Defaults and at least one destination is defined (see Container Files).
In JES2, JES3 and POWER, if a spool file is processed with a
DEST
specification that matches one of the destinations defined in
the Monitor Defaults (see
Container
Files).
Separating, browsing and printing in parts would require direct access to record ranges of the output. Access to the original print data at the source, however, may not always be possible.
When container files are used, Entire Output Management stores the complete print data in a container file before processing any report definitions. The original print data are copied as a whole. All subsequent separation processing, browsing and printing in parts requires no access to the original data source, but uses the print data in the container file. Consequently processing is also very fast.
Entire Output Management keeps the whole original output stored in the database as long as there is at least one active report with location "S" (= source) pointing to it. This could mean that the container file is filled with very large output, even though only a fraction is actually needed.
You should store print data in a container file whenever intensive separation processing (separation exits), browsing or printing in parts is necessary.
If you use intensive separation processing, but the resulting active reports use only a small part of the original print data, you should copy the print data into the active-data file.
The storing of print data in the active-data file is controlled by the option Copy report content to NOM database in the report definition.
If this option is set to "Yes", the resulting active report will be copied from the original output stored in the container file into the active-data file (active report location = "D"). When the next cleanup is done and there are no active reports with location "S" pointing to the original source, it will be deleted from the container file.
You should use the active-data file if you have intensive separation processing, but only small parts of the original print data are needed or if the resulting active reports have very different expiration dates.
On local z/OS systems, active reports should only be copied into the active-data file if absolutely necessary (for example, to avoid accidental loss through spool deletion), because the storing of large reports in the database creates a considerable overhead.
Be aware though, that for the lifetime of the original output in the container file, reports created with the Copy report content to NOM database set to "Yes" mean that those parts of the output are actually stored again in the active-data file, whereas with location "S" Entire Output Management keeps only pointers to the container file.
Bear in mind that the active report with the highest expiration date determines the lifetime of the whole original output in the container file.
Container files for mainframe spooling and job-entry systems, together with the spool destination that will be copied into the associated container file, are defined in the Monitor Defaults.
Container files for Entire Output Management on UNIX are defined in the Node Definitions for a UNIX or Windows node.
Container files for CA Spool are defined in the CA Spool Defaults.
Container files for Natural Advanced Facilities are defined in the Natural Advanced Facilities Defaults.
Container files for the 3GL Interface are defined in 3GL Interface Maintenance.