I'm trying to pass and get arguments in my databricks job it's a spark_python_task type IT IS NOT A NOTEBOOK. I deployed my job with dbx from pycharm. I have deployment.json file where I configure deployment stuff.
If you follow documentation about deployment file, you can see that you can specify parameters as parameters
array:
{
"name": "this-parameter-is-required!",
"spark_python_task": {
"python_file": "path/to/entrypoint.py"
},
"parameters": [
"--conf-file",
"conf/test/sample.json"
]
}
Parameters are passed as command-line parameters, so you can get them from the code just using the sys.argv, or built-in argparse library.