diff --git a/spark/spark-tensorflow-connector/pom.xml b/spark/spark-tensorflow-connector/pom.xml index 389ab173..31dd6e45 100644 --- a/spark/spark-tensorflow-connector/pom.xml +++ b/spark/spark-tensorflow-connector/pom.xml @@ -4,7 +4,7 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> 4.0.0 org.tensorflow - spark-tensorflow-connector_2.11 + spark-tensorflow-connector_${scala.binary.version} jar 1.10.0 spark-tensorflow-connector @@ -27,13 +27,14 @@ UTF-8 - 3.2.2 - 2.11 + 4.3.0 + 2.12.10 + 2.12 1.0 - 2.2.6 + 3.0.8 3.0 1.8 - 2.3.1 + 2.4.4 2.7.3 4.11 @@ -86,7 +87,7 @@ incremental true - ${scala.binary.version} + ${scala.version} false diff --git a/spark/spark-tensorflow-connector/src/main/scala/org/tensorflow/spark/datasources/tfrecords/DefaultSource.scala b/spark/spark-tensorflow-connector/src/main/scala/org/tensorflow/spark/datasources/tfrecords/DefaultSource.scala index bea6623d..f7742c12 100644 --- a/spark/spark-tensorflow-connector/src/main/scala/org/tensorflow/spark/datasources/tfrecords/DefaultSource.scala +++ b/spark/spark-tensorflow-connector/src/main/scala/org/tensorflow/spark/datasources/tfrecords/DefaultSource.scala @@ -167,7 +167,7 @@ class DefaultSource extends DataSourceRegister } } -object DefaultSource { +object DefaultSource extends scala.Serializable { // The function run on each worker. // Writes the partition to a file and returns the number of records output. private def writePartitionLocal( @@ -189,7 +189,7 @@ object DefaultSource { // Make the directory if it does not exist dir.mkdirs() // The path to the partition file. - val filePath = localPath + s"/part-" + String.format("%05d", new java.lang.Integer(index)) + val filePath = localPath + s"/part-" + String.format("%05d", java.lang.Integer.valueOf(index)) val fos = new DataOutputStream(new FileOutputStream(filePath)) var count = 0 try {