diff --git a/sahara_tests/scenario/custom_checks/check_kafka.py b/sahara_tests/scenario/custom_checks/check_kafka.py index 80f66b64..d8c922fa 100644 --- a/sahara_tests/scenario/custom_checks/check_kafka.py +++ b/sahara_tests/scenario/custom_checks/check_kafka.py @@ -89,7 +89,7 @@ class CustomCheckKafka(object): def _prepare_spark_kafka_job_running(self, shs): ip = self._get_node_ip(shs) utils_url = ( - "http://central.maven.org/maven2/org/apache/spark" + "https://repo1.maven.org/maven2/org/apache/spark" "/spark-streaming-kafka-assembly_2.10/1.4.1" "/spark-streaming-kafka-assembly_2.10-1.4.1.jar") # try to search spark-kafka assembly utils diff --git a/sahara_tests/scenario/defaults/edp-examples/edp-spark/README.rst b/sahara_tests/scenario/defaults/edp-examples/edp-spark/README.rst index 0f222828..0cbf1d47 100644 --- a/sahara_tests/scenario/defaults/edp-examples/edp-spark/README.rst +++ b/sahara_tests/scenario/defaults/edp-examples/edp-spark/README.rst @@ -54,7 +54,7 @@ Launching spark-kafka-example and ``path`` with your own values. 2. Download the Spark Streaming utils to the node with your ``Spark History Server`` from this URL: - ``http://central.maven.org/maven2/org/apache/spark/spark-streaming-kafka-assembly_2.10/1.4.1/spark-streaming-kafka-assembly_2.10-1.4.1.jar``. + ``https://repo1.maven.org/maven2/org/apache/spark/spark-streaming-kafka-assembly_2.10/1.4.1/spark-streaming-kafka-assembly_2.10-1.4.1.jar``. Now you are ready to launch your job from sahara UI. 3. Create a job binary that points to ``spark-kafka-example.py``. Also you need to create a job that uses this job binary as a main binary.