When developing python code, I make use of the package ipdb.
This halts the execution of the python code there, where I have inserted ipdb.set_trace(), and presents me with a python interpreter command line.
However, in the python code that I develop for pyspark, and which I send off using spark-submit, the ipdb package does not work.
So my question is: is there a way, in which I can debug my pyspark code in a manner similar to using the ipdb package?
Note: Obviously, for python code executed on remote nodes, this would not be possible. But when using spark-submit with option --master local[1] I have hopes that it might be possible.
PS. There is a related question, but with a narrower scope, here: How to PySpark Codes in Debug Jupyter Notebook