@@ -96,11 +96,13 @@ trait DStreamCheckpointTester { self: SparkFunSuite =>
9696 expectedOutput.take(numBatchesBeforeRestart), stopSparkContextAfterTest)
9797
9898 // Restart and complete the computation from checkpoint file
99- logInfo(
99+ // scalastyle:off println
100+ print(
100101 " \n -------------------------------------------\n " +
101102 " Restarting stream computation " +
102103 " \n -------------------------------------------\n "
103104 )
105+ // scalastyle:on println
104106 val restartedSsc = new StreamingContext (checkpointDir)
105107 generateAndAssertOutput[V ](restartedSsc, batchDuration, checkpointDir, nextNumBatches,
106108 expectedOutput.takeRight(nextNumExpectedOutputs), stopSparkContextAfterTest)
@@ -125,9 +127,11 @@ trait DStreamCheckpointTester { self: SparkFunSuite =>
125127 ssc.start()
126128 val numBatches = expectedOutput.size
127129 val clock = ssc.scheduler.clock.asInstanceOf [ManualClock ]
128- logDebug(" Manual clock before advancing = " + clock.getTimeMillis())
130+ // scalastyle:off println
131+ logInfo(" Manual clock before advancing = " + clock.getTimeMillis())
129132 clock.advance((batchDuration * numBatches).milliseconds)
130- logDebug(" Manual clock after advancing = " + clock.getTimeMillis())
133+ logInfo(" Manual clock after advancing = " + clock.getTimeMillis())
134+ // scalastyle:on println
131135
132136 val outputStream = ssc.graph.getOutputStreams().filter { dstream =>
133137 dstream.isInstanceOf [TestOutputStreamWithPartitions [V ]]
0 commit comments