Skip to content

Commit 123d958

Browse files
author
Marcelo Vanzin
committed
Merge branch 'master' into SPARK-2750
2 parents 66bf7e6 + b3ba1be commit 123d958

File tree

280 files changed

+516
-574
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

280 files changed

+516
-574
lines changed

core/src/test/scala/org/apache/spark/ContextCleanerSuite.scala

Lines changed: 3 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -24,18 +24,14 @@ import scala.language.existentials
2424
import scala.util.Random
2525

2626
import org.scalatest.BeforeAndAfter
27-
import org.scalatest.concurrent.PatienceConfiguration
2827
import org.scalatest.concurrent.Eventually._
28+
import org.scalatest.concurrent.PatienceConfiguration
2929
import org.scalatest.time.SpanSugar._
3030

31-
import org.apache.spark.rdd.{ReliableRDDCheckpointData, RDD}
32-
import org.apache.spark.storage._
31+
import org.apache.spark.rdd.{RDD, ReliableRDDCheckpointData}
3332
import org.apache.spark.shuffle.hash.HashShuffleManager
3433
import org.apache.spark.shuffle.sort.SortShuffleManager
35-
import org.apache.spark.storage.BroadcastBlockId
36-
import org.apache.spark.storage.RDDBlockId
37-
import org.apache.spark.storage.ShuffleBlockId
38-
import org.apache.spark.storage.ShuffleIndexBlockId
34+
import org.apache.spark.storage._
3935

4036
/**
4137
* An abstract base class for context cleaner tests, which sets up a context with a config

core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,7 @@ package org.apache.spark
2020
import scala.collection.mutable
2121

2222
import org.scalatest.{BeforeAndAfter, PrivateMethodTester}
23+
2324
import org.apache.spark.executor.TaskMetrics
2425
import org.apache.spark.scheduler._
2526
import org.apache.spark.scheduler.cluster.ExecutorInfo

core/src/test/scala/org/apache/spark/FailureSuite.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,10 +17,10 @@
1717

1818
package org.apache.spark
1919

20-
import org.apache.spark.util.NonSerializable
21-
2220
import java.io.{IOException, NotSerializableException, ObjectInputStream}
2321

22+
import org.apache.spark.util.NonSerializable
23+
2424
// Common state shared by FailureSuite-launched tasks. We use a global object
2525
// for this because any local variables used in the task closures will rightfully
2626
// be copied for each task, so there's no other way for them to share state.

core/src/test/scala/org/apache/spark/FileServerSuite.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -27,10 +27,10 @@ import org.apache.commons.lang3.RandomUtils
2727

2828
import org.apache.spark.util.Utils
2929

30-
import SSLSampleConfigs._
31-
3230
class FileServerSuite extends SparkFunSuite with LocalSparkContext {
3331

32+
import SSLSampleConfigs._
33+
3434
@transient var tmpDir: File = _
3535
@transient var tmpFile: File = _
3636
@transient var tmpJarUrl: String = _

core/src/test/scala/org/apache/spark/FileSuite.scala

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -21,17 +21,16 @@ import java.io.{File, FileWriter}
2121

2222
import scala.io.Source
2323

24-
import org.apache.spark.input.PortableDataStream
25-
import org.apache.spark.storage.StorageLevel
26-
2724
import org.apache.hadoop.io._
2825
import org.apache.hadoop.io.compress.DefaultCodec
29-
import org.apache.hadoop.mapred.{JobConf, FileAlreadyExistsException, FileSplit, TextInputFormat, TextOutputFormat}
26+
import org.apache.hadoop.mapred.{FileAlreadyExistsException, FileSplit, JobConf, TextInputFormat, TextOutputFormat}
3027
import org.apache.hadoop.mapreduce.Job
3128
import org.apache.hadoop.mapreduce.lib.input.{FileSplit => NewFileSplit, TextInputFormat => NewTextInputFormat}
3229
import org.apache.hadoop.mapreduce.lib.output.{TextOutputFormat => NewTextOutputFormat}
3330

34-
import org.apache.spark.rdd.{NewHadoopRDD, HadoopRDD}
31+
import org.apache.spark.input.PortableDataStream
32+
import org.apache.spark.rdd.{HadoopRDD, NewHadoopRDD}
33+
import org.apache.spark.storage.StorageLevel
3534
import org.apache.spark.util.Utils
3635

3736
class FileSuite extends SparkFunSuite with LocalSparkContext {

core/src/test/scala/org/apache/spark/HeartbeatReceiverSuite.scala

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -25,13 +25,13 @@ import scala.concurrent.Await
2525
import scala.concurrent.duration._
2626
import scala.language.postfixOps
2727

28-
import org.scalatest.{BeforeAndAfterEach, PrivateMethodTester}
29-
import org.mockito.Mockito.{mock, spy, verify, when}
3028
import org.mockito.Matchers
3129
import org.mockito.Matchers._
30+
import org.mockito.Mockito.{mock, spy, verify, when}
31+
import org.scalatest.{BeforeAndAfterEach, PrivateMethodTester}
3232

3333
import org.apache.spark.executor.TaskMetrics
34-
import org.apache.spark.rpc.{RpcCallContext, RpcEndpoint, RpcEnv, RpcEndpointRef}
34+
import org.apache.spark.rpc.{RpcCallContext, RpcEndpoint, RpcEndpointRef, RpcEnv}
3535
import org.apache.spark.scheduler._
3636
import org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages._
3737
import org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend

core/src/test/scala/org/apache/spark/LocalSparkContext.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717

1818
package org.apache.spark
1919

20-
import _root_.io.netty.util.internal.logging.{Slf4JLoggerFactory, InternalLoggerFactory}
20+
import _root_.io.netty.util.internal.logging.{InternalLoggerFactory, Slf4JLoggerFactory}
2121
import org.scalatest.BeforeAndAfterAll
2222
import org.scalatest.BeforeAndAfterEach
2323
import org.scalatest.Suite

core/src/test/scala/org/apache/spark/MapOutputTrackerSuite.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,10 +19,10 @@ package org.apache.spark
1919

2020
import scala.collection.mutable.ArrayBuffer
2121

22-
import org.mockito.Mockito._
2322
import org.mockito.Matchers.{any, isA}
23+
import org.mockito.Mockito._
2424

25-
import org.apache.spark.rpc.{RpcAddress, RpcEndpointRef, RpcCallContext, RpcEnv}
25+
import org.apache.spark.rpc.{RpcAddress, RpcCallContext, RpcEndpointRef, RpcEnv}
2626
import org.apache.spark.scheduler.{CompressedMapStatus, MapStatus}
2727
import org.apache.spark.shuffle.FetchFailedException
2828
import org.apache.spark.storage.{BlockManagerId, ShuffleBlockId}

core/src/test/scala/org/apache/spark/SSLOptionsSuite.scala

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,9 +21,10 @@ import java.io.File
2121
import javax.net.ssl.SSLContext
2222

2323
import com.google.common.io.Files
24-
import org.apache.spark.util.Utils
2524
import org.scalatest.BeforeAndAfterAll
2625

26+
import org.apache.spark.util.Utils
27+
2728
class SSLOptionsSuite extends SparkFunSuite with BeforeAndAfterAll {
2829

2930
test("test resolving property file as spark conf ") {

core/src/test/scala/org/apache/spark/ShuffleSuite.scala

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,17 +17,17 @@
1717

1818
package org.apache.spark
1919

20-
import java.util.concurrent.{Callable, Executors, ExecutorService, CyclicBarrier}
20+
import java.util.concurrent.{Callable, CyclicBarrier, Executors, ExecutorService}
2121

2222
import org.scalatest.Matchers
2323

2424
import org.apache.spark.ShuffleSuite.NonJavaSerializableClass
2525
import org.apache.spark.memory.TaskMemoryManager
2626
import org.apache.spark.rdd.{CoGroupedRDD, OrderedRDDFunctions, RDD, ShuffledRDD, SubtractedRDD}
27-
import org.apache.spark.scheduler.{MyRDD, MapStatus, SparkListener, SparkListenerTaskEnd}
27+
import org.apache.spark.scheduler.{MapStatus, MyRDD, SparkListener, SparkListenerTaskEnd}
2828
import org.apache.spark.serializer.KryoSerializer
2929
import org.apache.spark.shuffle.ShuffleWriter
30-
import org.apache.spark.storage.{ShuffleDataBlockId, ShuffleBlockId}
30+
import org.apache.spark.storage.{ShuffleBlockId, ShuffleDataBlockId}
3131
import org.apache.spark.util.MutablePair
3232

3333
abstract class ShuffleSuite extends SparkFunSuite with Matchers with LocalSparkContext {

0 commit comments

Comments
 (0)