From 408d30d2a1dd562771dec1a4a48509319152f943 Mon Sep 17 00:00:00 2001 From: Markus Cozowicz Date: Fri, 26 Oct 2018 12:57:46 +0200 Subject: [PATCH] Update README.md --- README.md | 26 ++++++++++++++++++++++++++ 1 file changed, 26 insertions(+) diff --git a/README.md b/README.md index a32206c..bf42cb7 100644 --- a/README.md +++ b/README.md @@ -43,3 +43,29 @@ On [Spark](https://spark.apache.org/) one can install all 3 components (C++, Pyt ```bash sbt spPublish ``` + +# Jupyter Notebook Setup + +# Spark Setup + +One must set the crossJoin property to enable calculation of the similarity matrix. + +``` +spark.sql.crossJoin.enabled true +``` + +# Databricks Setup + +One must set the crossJoin property to enable calculation of the similarity matrix. + +``` +spark.sql.crossJoin.enabled true +``` + +1. Navigate to your workspace +2. Create library +3. Under 'Source' select 'Maven Coordinate' +4. Enter eisber:sarplus:0.2.1 +5. Hit 'Create Libbrary' + +This will install C++, Python and Scala code on your cluster.