Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

new:usr:SDK-415:Adding a spark command example #328

Open
wants to merge 4 commits into
base: unreleased
Choose a base branch
from
Open
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
115 changes: 115 additions & 0 deletions example/qds_spark_example.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,115 @@
"""
This is a sample code used for submitting a Spark script (SparkCommand) and getting the result.
"""

from qds_sdk.qubole import Qubole
from qds_sdk.commands import SparkCommand
import time


def get_results_filename(command_id):
"""
A helper method to generate a file name to write the downloaded result
:param command_id:
:return:
"""
return "/tmp/result_{}.tsv".format(command_id)


def get_content(script_file_path):
"""
Helper method to read a script file in the given location.
:param script_file_path:
:return:
"""
with open(script_file_path, 'r') as content_file:
content = content_file.read()
return content


def execute_query(cluster_label, cmd_to_run, language, user_program_arguments=None, arguments=None):
"""
Helper method to execute a script
:param cluster_label:
:param cmd_to_run:
:param language:
:param user_program_arguments:
:param arguments:
:return:
"""
if script is None or script == "":
print("script cannot be None or empty")
return None

if not language:
print("language cannot be None or empty")
return

if language in ["command_line"]:
# A Shell command needs to be invoked in this fashion
cmd = SparkCommand.create(label=cluster_label, cmdline=cmd_to_run, arguments=arguments,
user_program_arguments=user_program_arguments)
elif language == "sql":
# A SQL command needs to be invoked in this fashion
cmd = SparkCommand.create(label=cluster_label, sql=cmd_to_run, arguments=arguments)
else:
# A python, R or scala command needs to be invoked in this fashion.
cmd = SparkCommand.create(label=cluster_label, program=cmd_to_run, language=language,
arguments=arguments, user_program_arguments=user_program_arguments)

while not SparkCommand.is_done(cmd.status):
print("Waiting for completion of command : {}".format(cmd.id))
cmd = SparkCommand.find(cmd.id)
time.sleep(5)

if SparkCommand.is_success(cmd.status):
print("\nCommand Executed: Completed successfully")
else:
print("\nCommand Executed: Failed!!!. The status returned is: {}".format(cmd.status))
print(cmd.get_log())
return cmd


def get_results(command):
"""
A helper method to get the results
:param command:
:return:
"""
if command is None:
return None

results_file_name = get_results_filename(command.id)
fp = open(results_file_name, 'w')

command.get_results(fp, delim="\n")
print("results are written to {}".format(results_file_name))


if __name__ == '__main__':
# Set the API token. If you are using any other environment other then api.qubole.com then set api_url to that url
# as <env_url>/api
Qubole.configure(api_token='<auth_token>')

filename = "<your script location>"
user_program_arguments = None # arguments for your script
arguments = None # spark configuration for your program for ex : "--conf spark.executor.memory=1024M"
cluster_label = "<your cluster label>" # the cluster on which the command will run

# Running a python command. In case of scala or R get the script content and then set the langauage field to scala
# or R as required
script = get_content(filename)
command = execute_query(cluster_label, script, "python", user_program_arguments=user_program_arguments,
mayankahuja marked this conversation as resolved.
Show resolved Hide resolved
arguments=arguments)
get_results(command)

# Running a SQL command
script = "show tables"
command = execute_query(cluster_label, script, "sql", arguments=arguments)
get_results(command)

# Running a shell command
script = "/usr/lib/spark/bin/spark-submit --class org.apache.spark.examples.SparkPi --master yarn " \
"--deploy-mode client /usr/lib/spark/spark-examples* 1"
command = execute_query(cluster_label, script, "command_line", arguments=arguments)
get_results(command)