Skip to content

Commit 4a46f97

Browse files
icywang86ruiRui Wang Napieralski
andauthored
fix: upgrade pyspark (#135)
* fix: upgrade pyspark * use 2.4.0 instead and fix breaking changes * remove py2 * remove py27 * upgrade spark version Co-authored-by: Rui Wang Napieralski <[email protected]>
1 parent 063af62 commit 4a46f97

File tree

6 files changed

+11
-8
lines changed

6 files changed

+11
-8
lines changed

buildspec.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ phases:
4343
- tox -e flake8,twine,sphinx
4444
- AWS_ACCESS_KEY_ID= AWS_SECRET_ACCESS_KEY= AWS_SESSION_TOKEN=
4545
AWS_CONTAINER_CREDENTIALS_RELATIVE_URI=
46-
tox -e py27,py36,stats -- tests/
46+
tox -e py36,stats -- tests/
4747

4848
# spark integration tests
4949
- cd $CODEBUILD_SRC_DIR/integration-tests/sagemaker-spark-sdk

sagemaker-pyspark-sdk/setup.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -101,7 +101,7 @@ def read_version():
101101
scripts=["bin/sagemakerpyspark-jars", "bin/sagemakerpyspark-emr-jars"],
102102

103103
install_requires=[
104-
"pyspark==2.3.4",
104+
"pyspark==2.4.0",
105105
"numpy",
106106
],
107107
)

sagemaker-pyspark-sdk/src/sagemaker_pyspark/algorithms/XGBoostSageMakerEstimator.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -175,8 +175,8 @@ class XGBoostSageMakerEstimator(SageMakerEstimatorBase):
175175
"Subsample ratio of columns for each split, in each level. Must be in (0, 1].",
176176
typeConverter=TypeConverters.toFloat)
177177

178-
_lambda = Param(
179-
Params._dummy(), "lambda",
178+
lambda_weights = Param(
179+
Params._dummy(), "lambda_weights",
180180
"L2 regularization term on weights, increase this value"
181181
" will make model more conservative.",
182182
typeConverter=TypeConverters.toFloat)
@@ -508,10 +508,10 @@ def setColSampleByLevel(self, value):
508508
self._set(colsample_bylevel=value)
509509

510510
def getLambda(self):
511-
return self.getOrDefault(self._lambda)
511+
return self.getOrDefault(self.lambda_weights)
512512

513513
def setLambda(self, value):
514-
self._set(_lambda=value)
514+
self._set(lambda_weights=value)
515515

516516
def getAlpha(self):
517517
return self.getOrDefault(self.alpha)

sagemaker-pyspark-sdk/src/sagemaker_pyspark/wrapper.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -116,6 +116,7 @@ class Option(SageMakerJavaWrapper):
116116

117117
def __init__(self, value):
118118
self.value = value
119+
self._java_obj = None
119120

120121
@classmethod
121122
def empty(cls):
@@ -147,6 +148,7 @@ class ScalaMap(SageMakerJavaWrapper):
147148

148149
def __init__(self, dictionary):
149150
self.dictionary = dictionary
151+
self._java_obj = None
150152

151153
def _to_java(self):
152154
map = self._new_java_obj(ScalaMap._wrapped_class)
@@ -169,6 +171,7 @@ class ScalaList(SageMakerJavaWrapper):
169171

170172
def __init__(self, p_list):
171173
self.p_list = p_list
174+
self._java_obj = None
172175

173176
def _to_java(self):
174177
# Since py4j cannot deal with scala list directly

sagemaker-pyspark-sdk/tox.ini

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
[tox]
2-
envlist = flake8,twine,sphinx,py{27,36},stats
2+
envlist = flake8,twine,sphinx,py36,stats
33
skip_missing_interpreters = False
44

55
[testenv]

sagemaker-spark-sdk/build.sbt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ scalaVersion := "2.11.7"
1818

1919
// to change the version of spark add -DSPARK_VERSION=2.x.x when running sbt
2020
// for example: "sbt -DSPARK_VERSION=2.1.1 clean compile test doc package"
21-
val sparkVersion = System.getProperty("SPARK_VERSION", "2.2.0")
21+
val sparkVersion = System.getProperty("SPARK_VERSION", "2.4.0")
2222

2323
lazy val SageMakerSpark = (project in file("."))
2424

0 commit comments

Comments
 (0)