• Status: Solved
  • Priority: Medium
  • Security: Public
  • Views: 624
  • Last Modified:

Spooling stopping around 25G from scheduler; running process manually spools entire contents

When using a scheduler in production, the shell script called is spooling to a file.  The file stops midline around 25Gb; it is not always stopping on the same record or point within the record, so it is not a data issue.  Calling the shell script manually, the entire contents of the spool successfully (around 200Gb). Using the same scheduler job in QA, the entire contents of the spool successfully (around 150Gb).  

In production, the scheduler calling the same shell script, with different paramters and extracting smaller dataset,  spools successfully.

The output from the sqlplus session is logged in a log file.  All of the records are logged (the 200Gb sized set). At the end of the log the following error appears:
         SP2-0308: cannot close spool file
Here is the version of SQL*Plus (it is the same in QA and Production):
         SQL*Plus: Release

The scheduler and the user when running the job manually is the same.

Any ideas are appreciated. Thanks.

1 Solution
johnsoneSenior Oracle DBACommented:
From the Oracle manual, here is the description of that error:

SP2-0308 Cannot close spool file

Cause: The file is currently being used.

Action: Release the file from the other process.

Is it possible the scheduled job kicks off twice, or kicks off a second time before the first one is finished?
HangingCurveAuthor Commented:
Thanks Sage, I was thinking the same thing...but upon further review...
Well, of course after submitting this, we had an epiphany at the end of the day yesterday.
We determined that the scheduler is running the script from another directory. We manually ran the script from that directory.  It turns out that from that directory it is spooling to file system that is different from where the shell scripts are.  That filesystem didn't have enough memory, so the spooling hit the limit and stopped.
The sys admin has increased the size and the spooling extracts all of the records successfully when running the script manually from the directory the scheduler is running from.
Question has a verified solution.

Are you are experiencing a similar issue? Get a personalized answer when you ask a related question.

Have a better answer? Share it in a comment.

Join & Write a Comment

Featured Post

Get your problem seen by more experts

Be seen. Boost your question’s priority for more expert views and faster solutions

Tackle projects and never again get stuck behind a technical roadblock.
Join Now