How to troubleshoot a DSX scheduled notebook?

407 views Asked by At

I have a DSX notebook that I can run manually usng the DSX user interface and it populates some data in a Cloudant database.

I have scheduled the notebook to run hourly. Overnight I would have expected the job to have run many times, but the Cloudant database has not been updated.

How can I debug the scheduled job? Are there any logs that I can check to verify that the notebook has actually been executed? Is the output from my notebook saved to log files? Where can I find these files?

3

There are 3 answers

1
Sven Hafeneger On BEST ANSWER

One possibility is to look into the kernel logs of your notebook kernel. For that you need to use a Python notebook.

Check the following location on the gpfs in your Python notebook:

!ls /gpfs/fs01/user/USERID/logs/notebook/

To get the USERID execute the following code:

!whoami

You should find for each kernel a log file, e.g. kernel-python3-20170105_102510.

0
Sumit Goyal On

You can find the notebook execution logs at the location ~/notebook/logs/. If your notebook was executed multiple times over the night you will find several log entries (ls -la ~/notebook/logs/). Open up one of the relevant log files (according to the timestamp) to see if the notebooks had connection issues with Cloudant or something else (cat ~/notebook/logs/kernel-pyspark-20170105_095103.log).

0
Chris Snow On

One gotcha is print() statements will not make it to the log file, so you need to use the spark logging functionality. In pyspark, I created a utility function that sends the output to a log file, but also prints it to the notebook when the notebook is run manually:

# utility method for logging
log4jLogger = sc._jvm.org.apache.log4j

# give a meaningful name to your logger (mine is CloudantRecommender)
LOGGER = log4jLogger.LogManager.getLogger("CloudantRecommender")

def info(*args):
    print(args) # sends output to notebook
    LOGGER.info(args) # sends output to kernel log file

def error(*args): 
    print(args) # sends output to notebook
    LOGGER.error(args) # sends output to kernel log file

Using the function like so in my notebook:

info("some log output")

If I check the log files I can see my logout is getting written:

! grep 'CloudantRecommender' $HOME/logs/notebook/*pyspark* 

kernel-pyspark-20170105_164844.log:17/01/05 10:49:08 INFO CloudantRecommender: [Starting load from Cloudant: , 2017-01-05 10:49:08]
kernel-pyspark-20170105_164844.log:17/01/05 10:53:21 INFO CloudantRecommender: [Finished load from Cloudant: , 2017-01-05 10:53:21]

Exceptions don't appear to get sent to the log file either, so you will need to wrap code in a try block and log the error, e.g.

import traceback
try:
    # your spark code that may throw an exception
except Exception as e:
    # send the exception to the spark logger
    error(str(e), traceback.format_exc(), ts())
    raise e

CAUTION: Another gotcha that hit me during debugging is that scheduled jobs run a specific version of a notebook. Check that you update the schedule job when you save new versions of your notebook.