fixing the issue in YARN clusters introduced during Spark 2.0 upgrade
This commit is contained in:
Коммит
fbf0f8d841
|
@ -65,7 +65,9 @@ namespace Microsoft.Spark.CSharp.Configuration
|
|||
configuration = new SparkCLRConfiguration(appConfig);
|
||||
runMode = RunMode.CLUSTER;
|
||||
}
|
||||
else if (sparkMaster.Equals("yarn-client", StringComparison.OrdinalIgnoreCase) || sparkMaster.Equals("yarn-cluster", StringComparison.OrdinalIgnoreCase))
|
||||
else if (sparkMaster.Equals("yarn-cluster", StringComparison.OrdinalIgnoreCase) ||
|
||||
sparkMaster.Equals("yarn-client", StringComparison.OrdinalIgnoreCase) ||
|
||||
sparkMaster.Equals("yarn", StringComparison.OrdinalIgnoreCase)) //supported in Spark 2.0
|
||||
{
|
||||
configuration = new SparkCLRConfiguration(appConfig);
|
||||
runMode = RunMode.YARN;
|
||||
|
|
|
@ -68,7 +68,7 @@ class CSharpRDD(
|
|||
val func = SQLUtils.createCSharpFunction(command,
|
||||
envVars,
|
||||
cSharpIncludes,
|
||||
cSharpWorkerExecutable,
|
||||
cSharpWorker.getAbsolutePath,
|
||||
unUsedVersionIdentifier,
|
||||
broadcastVars,
|
||||
accumulator)
|
||||
|
|
Загрузка…
Ссылка в новой задаче