-
Notifications
You must be signed in to change notification settings - Fork 33
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Signed-off-by: Heng Qian <[email protected]>
- Loading branch information
1 parent
3ff2ef2
commit 1864f6e
Showing
10 changed files
with
323 additions
and
6 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
122 changes: 122 additions & 0 deletions
122
...-test/src/integration/scala/org/opensearch/flint/spark/ppl/FlintSparkPPLCastITSuite.scala
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,122 @@ | ||
/* | ||
* Copyright OpenSearch Contributors | ||
* SPDX-License-Identifier: Apache-2.0 | ||
*/ | ||
|
||
package org.opensearch.flint.spark.ppl | ||
|
||
import java.sql.Date | ||
import java.sql.Timestamp | ||
|
||
import org.apache.spark.sql.{QueryTest, Row} | ||
import org.apache.spark.sql.streaming.StreamTest | ||
|
||
class FlintSparkPPLCastITSuite | ||
extends QueryTest | ||
with LogicalPlanTestUtils | ||
with FlintPPLSuite | ||
with StreamTest { | ||
|
||
/** Test table and index name */ | ||
private val testTable = "spark_catalog.default.flint_ppl_test" | ||
|
||
override def beforeAll(): Unit = { | ||
super.beforeAll() | ||
// Create test table | ||
createNullableJsonContentTable(testTable) | ||
} | ||
|
||
protected override def afterEach(): Unit = { | ||
super.afterEach() | ||
// Stop all streaming jobs if any | ||
spark.streams.active.foreach { job => | ||
job.stop() | ||
job.awaitTermination() | ||
} | ||
} | ||
|
||
test("test cast number to compatible data types") { | ||
val frame = sql(s""" | ||
| source=$testTable | eval | ||
| id_string = cast(id as string), | ||
| id_double = cast(id as double), | ||
| id_long = cast(id as long), | ||
| id_boolean = cast(id as boolean) | ||
| | fields id, id_string, id_double, id_long, id_boolean | head 1 | ||
| """.stripMargin) | ||
|
||
assert( | ||
frame.dtypes.sameElements( | ||
Array( | ||
("id", "IntegerType"), | ||
("id_string", "StringType"), | ||
("id_double", "DoubleType"), | ||
("id_long", "LongType"), | ||
("id_boolean", "BooleanType")))) | ||
assertSameRows(Seq(Row(1, "1", 1.0, 1L, true)), frame) | ||
} | ||
|
||
test("test cast string to compatible data types") { | ||
val frame = sql(s""" | ||
| source=$testTable | eval | ||
| id_int = cast(cast(id as string) as int), | ||
| cast_true = cast("True" as boolean), | ||
| cast_false = cast("false" as boolean), | ||
| cast_timestamp = cast("2024-11-26 23:39:06" as timestamp), | ||
| cast_date = cast("2024-11-26" as date), | ||
| cast_time = cast("12:34:56" as time) | ||
| | fields id_int, cast_true, cast_false, cast_timestamp, cast_date, cast_time | head 1 | ||
| """.stripMargin) | ||
|
||
// Note: Spark doesn't support data type of `Time`, cast it to StringTypes by default. | ||
assert( | ||
frame.dtypes.sameElements(Array( | ||
("id_int", "IntegerType"), | ||
("cast_true", "BooleanType"), | ||
("cast_false", "BooleanType"), | ||
("cast_timestamp", "TimestampType"), | ||
("cast_date", "DateType"), | ||
("cast_time", "StringType")))) | ||
assertSameRows( | ||
Seq( | ||
Row( | ||
1, | ||
true, | ||
false, | ||
Timestamp.valueOf("2024-11-26 23:39:06"), | ||
Date.valueOf("2024-11-26"), | ||
"12:34:56")), | ||
frame) | ||
} | ||
|
||
test("test cast time related types to compatible data types") { | ||
val frame = sql(s""" | ||
| source=$testTable | eval | ||
| timestamp = cast("2024-11-26 23:39:06" as timestamp), | ||
| ts_str = cast(timestamp as string), | ||
| ts_date = cast(timestamp as date), | ||
| date_str = cast(ts_date as string), | ||
| date_ts = cast(ts_date as timestamp) | ||
| | fields timestamp, ts_str, ts_date, date_str, date_ts | head 1 | ||
| """.stripMargin) | ||
|
||
assert( | ||
frame.dtypes.sameElements( | ||
Array( | ||
("timestamp", "TimestampType"), | ||
("ts_str", "StringType"), | ||
("ts_date", "DateType"), | ||
("date_str", "StringType"), | ||
("date_ts", "TimestampType")))) | ||
assertSameRows( | ||
Seq( | ||
Row( | ||
Timestamp.valueOf("2024-11-26 23:39:06"), | ||
"2024-11-26 23:39:06", | ||
Date.valueOf("2024-11-26"), | ||
"2024-11-26", | ||
Timestamp.valueOf("2024-11-26 00:00:00"))), | ||
frame) | ||
} | ||
|
||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
39 changes: 39 additions & 0 deletions
39
ppl-spark-integration/src/main/java/org/opensearch/sql/ast/expression/Cast.java
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,39 @@ | ||
/* | ||
* Copyright OpenSearch Contributors | ||
* SPDX-License-Identifier: Apache-2.0 | ||
*/ | ||
|
||
package org.opensearch.sql.ast.expression; | ||
|
||
import java.util.Collections; | ||
import java.util.List; | ||
import lombok.EqualsAndHashCode; | ||
import lombok.Getter; | ||
import lombok.RequiredArgsConstructor; | ||
import org.opensearch.sql.ast.AbstractNodeVisitor; | ||
|
||
/** | ||
* Expression node of cast | ||
*/ | ||
@Getter | ||
@EqualsAndHashCode(callSuper = false) | ||
@RequiredArgsConstructor | ||
public class Cast extends UnresolvedExpression { | ||
private final UnresolvedExpression expression; | ||
private final DataType dataType; | ||
|
||
@Override | ||
public List<UnresolvedExpression> getChild() { | ||
return Collections.singletonList(expression); | ||
} | ||
|
||
@Override | ||
public <R, C> R accept(AbstractNodeVisitor<R, C> nodeVisitor, C context) { | ||
return nodeVisitor.visitCast(this, context); | ||
} | ||
|
||
@Override | ||
public String toString() { | ||
return String.format("CAST(%s AS %s)", expression, dataType); | ||
} | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.