From ac43895d11805c4f6c343efa0dff3c7c48693701 Mon Sep 17 00:00:00 2001 From: sfc-gh-astachowski Date: Wed, 6 Nov 2024 14:55:39 +0100 Subject: [PATCH] Formatting --- .../client/core/SFArrowResultSetIT.java | 34 +++++++++---------- 1 file changed, 17 insertions(+), 17 deletions(-) diff --git a/src/test/java/net/snowflake/client/core/SFArrowResultSetIT.java b/src/test/java/net/snowflake/client/core/SFArrowResultSetIT.java index c7c5cfaaf..d7ea2faac 100644 --- a/src/test/java/net/snowflake/client/core/SFArrowResultSetIT.java +++ b/src/test/java/net/snowflake/client/core/SFArrowResultSetIT.java @@ -98,11 +98,11 @@ public void testNoOfflineData() throws Throwable { int dataSize = (int) file.length(); byte[] dataBytes = new byte[dataSize]; - try (InputStream is = new FileInputStream(file)) { - is.read(dataBytes, 0, dataSize); - } + try (InputStream is = new FileInputStream(file)) { + is.read(dataBytes, 0, dataSize); + } - SnowflakeResultSetSerializableV1 resultSetSerializable = new SnowflakeResultSetSerializableV1(); + SnowflakeResultSetSerializableV1 resultSetSerializable = new SnowflakeResultSetSerializableV1(); resultSetSerializable.setRootAllocator(new RootAllocator(Long.MAX_VALUE)); resultSetSerializable.setFirstChunkStringData(Base64.getEncoder().encodeToString(dataBytes)); resultSetSerializable.setFirstChunkByteData(dataBytes); @@ -225,11 +225,11 @@ public void testFirstResponseAndOfflineData() throws Throwable { int dataSize = (int) arrowFile.length(); byte[] dataBytes = new byte[dataSize]; - try (InputStream is = new FileInputStream(arrowFile)) { - is.read(dataBytes, 0, dataSize); - } + try (InputStream is = new FileInputStream(arrowFile)) { + is.read(dataBytes, 0, dataSize); + } - SnowflakeResultSetSerializableV1 resultSetSerializable = new SnowflakeResultSetSerializableV1(); + SnowflakeResultSetSerializableV1 resultSetSerializable = new SnowflakeResultSetSerializableV1(); resultSetSerializable.setFirstChunkStringData(Base64.getEncoder().encodeToString(dataBytes)); resultSetSerializable.setFirstChunkByteData(dataBytes); resultSetSerializable.setChunkFileCount(chunkCount); @@ -277,7 +277,7 @@ private class MockChunkDownloader implements ChunkDownloader { public SnowflakeResultChunk getNextChunkToConsume() throws SnowflakeSQLException { if (currentFileIndex < resultFileNames.size()) { ArrowResultChunk resultChunk = new ArrowResultChunk("", 0, 0, 0, rootAllocator, null); - try (InputStream is = new FileInputStream(resultFileNames.get(currentFileIndex))){ + try (InputStream is = new FileInputStream(resultFileNames.get(currentFileIndex))) { resultChunk.readArrowStream(is); currentFileIndex++; @@ -635,11 +635,11 @@ public void testSortedResultChunkWithStructVectors() throws Throwable { int dataSize = (int) file.length(); byte[] dataBytes = new byte[dataSize]; - try (InputStream is = new FileInputStream(file)) { - is.read(dataBytes, 0, dataSize); - } + try (InputStream is = new FileInputStream(file)) { + is.read(dataBytes, 0, dataSize); + } - resultSetSerializable.setRootAllocator(new RootAllocator(Long.MAX_VALUE)); + resultSetSerializable.setRootAllocator(new RootAllocator(Long.MAX_VALUE)); resultSetSerializable.setFirstChunkStringData( Base64.getEncoder().encodeToString(dataBytes)); resultSetSerializable.setFirstChunkByteData(dataBytes); @@ -723,11 +723,11 @@ public void testSortedResultChunk() throws Throwable { int dataSize = (int) file.length(); byte[] dataBytes = new byte[dataSize]; - try (InputStream is = new FileInputStream(file)) { - is.read(dataBytes, 0, dataSize); - } + try (InputStream is = new FileInputStream(file)) { + is.read(dataBytes, 0, dataSize); + } - resultSetSerializable.setRootAllocator(new RootAllocator(Long.MAX_VALUE)); + resultSetSerializable.setRootAllocator(new RootAllocator(Long.MAX_VALUE)); resultSetSerializable.setFirstChunkStringData( Base64.getEncoder().encodeToString(dataBytes)); resultSetSerializable.setFirstChunkByteData(dataBytes);