My client did an unload of an ADABAS system containing approximately 1,500,000,000 records. This took approximately 7 hours. using ADARUN with the ADAULD command. It then took an additional 20 hours to decompress this data using ADARUN with ADACMP DECOMPRESS
They are looking for ways to perform these tasks in less time - Could ADABAS sustain multiple concurrent unloads and decompresses?
I’ve also asked them to try using the INFILE command with DECOMPRESS rather than doing separate UNLOADs and DECOMPRESSes
Have you tried setting SORTSEQ=ISN, STARTISN and NUMREC options in parallel unloads? You might have to experiment with different number of parallel unloads (and the corresponding parallel decompresses) to see what works best in your environment (largely a function of the number of available processors and existing load).
Doing parallel unloads will be resource intensive, but should be able to reduce the total wall time needed to unload. Parallel decompresses can then be run against each unload file.
The exact STARTISN / NUMREC values to use will depend on the file being unloaded and how many parallel unloads you want to run. The values will need to be selected carefully to ensure neither dropping records nor including duplicates.
eww. Apparently ADAULD opens as UTYPE=EXU and only one user can have a file open for exclusive use (sorta what EXU says…). Not sure why EXU is needed for unload operation, but parallel unloads don’t appear to be a valid option presently. My apologies.
Hey thanks Wolfgang … I will set up a quick test and see if this works. These parallel unloads do need to be done without any other “user” accessing the files, so hopefully this works
For parallelism - have you tried creating an offline save with ADASAV and then passing the backup dataset to multiple ADAULD SAVETAPE with SELCRIT and SELVAL?
And have you tried ADACMP DECOMPRESS with the INFILE option? That may save you enough time to not need parallelism…
The issue with ADAUNL suggestion above is the backup dataset is on about 15 tapes and can only be mounted to one job at a time.
The most serious issue is with the DECOMPRESS performance. We haven’t been able to find a way to improve performance except to run 4 decompresses concurrently. If we run these 4 jobs serially, it take 16 hours. If we run them concurrently it take about 7 hours.