Skip to content

Commit 4b53667

Browse files
committed
review feedback
1 parent 6c57e4d commit 4b53667

File tree

2 files changed

+2
-3
lines changed

2 files changed

+2
-3
lines changed

core/src/main/scala/org/apache/spark/util/io/ChunkedByteBufferFileRegion.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ import org.apache.spark.network.util.AbstractFileRegion
2727

2828
/**
2929
* This exposes a ChunkedByteBuffer as a netty FileRegion, just to allow sending > 2gb in one netty
30-
* message. This is because netty cannot send a ByteBuf > 2g, but it can send a large FileRegion,
30+
* message. This is because netty cannot send a ByteBuf > 2g, but it can send a large FileRegion,
3131
* even though the data is not backed by a file.
3232
*/
3333
private[io] class ChunkedByteBufferFileRegion(
@@ -37,7 +37,7 @@ private[io] class ChunkedByteBufferFileRegion(
3737
private var _transferred: Long = 0
3838
// this duplicates the original chunks, so we're free to modify the position, limit, etc.
3939
private val chunks = chunkedByteBuffer.getChunks()
40-
private val size = chunks.foldLeft(0) { _ + _.remaining()}
40+
private val size = chunks.foldLeft(0) { _ + _.remaining() }
4141

4242
protected def deallocate: Unit = {}
4343

core/src/test/scala/org/apache/spark/io/ChunkedByteBufferFileRegionSuite.scala

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,6 @@ class ChunkedByteBufferFileRegionSuite extends SparkFunSuite with MockitoSugar
126126
/**
127127
* This mocks a channel which only accepts a limited number of bytes at a time. It also verifies
128128
* the written data matches our expectations as the data is received.
129-
* @param maxWriteSize
130129
*/
131130
private class LimitedWritableByteChannel(maxWriteSize: Int) extends WritableByteChannel {
132131
val bytes = new Array[Byte](maxWriteSize)

0 commit comments

Comments
 (0)