consolidating build scripts in one directory
This commit is contained in:
Родитель
ef63d27014
Коммит
8add53fa56
|
@ -63,8 +63,8 @@ Refer to the [docs folder](docs).
|
|||
|
||||
## Building, Running and Debugging SparkCLR
|
||||
|
||||
* [Windows Instructions](windows-instructions.md)
|
||||
* [Linux Instructions](linux-instructions.md)
|
||||
* [Windows Instructions](notes/windows-instructions.md)
|
||||
* [Linux Instructions](notes/linux-instructions.md)
|
||||
|
||||
(Note: Tested only with Spark 1.5.2)
|
||||
|
||||
|
@ -79,4 +79,4 @@ SparkCLR is licensed under the MIT license. See [LICENSE](LICENSE) file for full
|
|||
[![Issue Stats](http://issuestats.com/github/Microsoft/SparkCLR/badge/pr)](http://issuestats.com/github/Microsoft/SparkCLR)
|
||||
[![Issue Stats](http://issuestats.com/github/Microsoft/SparkCLR/badge/issue)](http://issuestats.com/github/Microsoft/SparkCLR)
|
||||
|
||||
We welcome contributions. To contribute, follow the instructions in [CONTRIBUTING.md](CONTRIBUTING.md).
|
||||
We welcome contributions. To contribute, follow the instructions in [CONTRIBUTING.md](notes/CONTRIBUTING.md).
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
@echo OFF
|
||||
REM @echo OFF
|
||||
setlocal
|
||||
|
||||
@rem check prerequisites
|
||||
|
@ -35,7 +35,7 @@ if NOT EXIST "%SPARKCLR_HOME%\samples" mkdir "%SPARKCLR_HOME%\samples"
|
|||
if NOT EXIST "%SPARKCLR_HOME%\scripts" mkdir "%SPARKCLR_HOME%\scripts"
|
||||
|
||||
@echo Assemble SparkCLR Scala components
|
||||
pushd "%CMDHOME%\scala"
|
||||
pushd "%CMDHOME%\..\scala"
|
||||
|
||||
@rem clean the target directory first
|
||||
call mvn.cmd clean
|
||||
|
@ -67,7 +67,7 @@ if EXIST "%CMDHOME%\lib" (
|
|||
)
|
||||
|
||||
@echo Assemble SparkCLR C# components
|
||||
pushd "%CMDHOME%\csharp"
|
||||
pushd "%CMDHOME%\..\csharp"
|
||||
|
||||
@rem clean any possible previous build first
|
||||
call Clean.cmd
|
||||
|
@ -92,12 +92,12 @@ copy /y Samples\Microsoft.Spark.CSharp\data\* "%SPARKCLR_HOME%\data\"
|
|||
popd
|
||||
|
||||
@echo Assemble SparkCLR script components
|
||||
pushd "%CMDHOME%\scripts"
|
||||
pushd "%CMDHOME%\..\scripts"
|
||||
copy /y *.cmd "%SPARKCLR_HOME%\scripts\"
|
||||
popd
|
||||
|
||||
@echo zip run directory
|
||||
pushd %~dp0
|
||||
if not exist ".\target" (mkdir .\target)
|
||||
powershell -f .\scripts\zipdir.ps1 -dir "%SPARKCLR_HOME%" -target ".\target\run.zip"
|
||||
powershell -f .\zipdir.ps1 -dir "%SPARKCLR_HOME%" -target ".\target\run.zip"
|
||||
popd
|
|
@ -19,7 +19,7 @@ fi
|
|||
[ ! -d "$SPARKCLR_HOME/scripts" ] && mkdir "$SPARKCLR_HOME/scripts"
|
||||
|
||||
echo "Assemble SparkCLR Scala components"
|
||||
pushd "$FWDIR/scala"
|
||||
pushd "$FWDIR/../scala"
|
||||
|
||||
# clean the target directory first
|
||||
mvn clean -q
|
||||
|
@ -53,7 +53,7 @@ then
|
|||
fi
|
||||
|
||||
echo "Assemble SparkCLR C# components"
|
||||
pushd "$FWDIR/csharp"
|
||||
pushd "$FWDIR/../csharp"
|
||||
|
||||
# clean any possible previous build first
|
||||
./clean.sh
|
||||
|
@ -79,7 +79,7 @@ cp Samples/Microsoft.Spark.CSharp/data/* "$SPARKCLR_HOME/data/"
|
|||
popd
|
||||
|
||||
echo "Assemble SparkCLR script components"
|
||||
pushd "$FWDIR/scripts"
|
||||
pushd "$FWDIR/../scripts"
|
||||
cp *.sh "$SPARKCLR_HOME/scripts/"
|
||||
popd
|
||||
|
|
@ -345,10 +345,10 @@ function Update-SparkVerboseMode
|
|||
|
||||
# replace {env:TEMP} with temp path
|
||||
$targetFile = "$temp\log4j.properties.temp"
|
||||
Replace-VariableInFile '\${env:TEMP}' "$tempValue" "$scriptDir\scripts\spark.conf\log4j.properties" $targetFile
|
||||
Replace-VariableInFile '\${env:TEMP}' "$tempValue" "$scriptDir\..\scripts\spark.conf\log4j.properties" $targetFile
|
||||
|
||||
# copy customized log4j properties to SPARK_HOME\conf
|
||||
copy-item $scriptDir\scripts\spark.conf\*.properties $S_HOME\conf -force
|
||||
copy-item $scriptDir\..\scripts\spark.conf\*.properties $S_HOME\conf -force
|
||||
copy-item $targetFile $S_HOME\conf\log4j.properties -force
|
||||
}
|
||||
else
|
|
@ -46,7 +46,7 @@ export PATH="$SPARK_HOME/bin:$PATH"
|
|||
if [ ! "$verbose" = "--verbose" ];
|
||||
then
|
||||
# redirect the logs from console (default) to /tmp
|
||||
cp "$FWDIR"/scripts/spark.conf/*.properties "$SPARK_HOME/conf/"
|
||||
cp "$FWDIR"/../scripts/spark.conf/*.properties "$SPARK_HOME/conf/"
|
||||
sed -i "s/\${env:TEMP}/\/tmp/g" "$SPARK_HOME/conf/log4j.properties"
|
||||
else
|
||||
# remove customized log4j.properties, revert back to out-of-the-box Spark logging to console
|
Загрузка…
Ссылка в новой задаче