v0.6.0
Bug Fixes:
- Properly handle special characters in JDBC connection strings (#132 / #134). This bug affected users whose Redshift passwords contained special characters that were not valid in URLs (e.g. a password containing a percentage-sign (
%
) character). - Restored compatibility with
spark-avro
1.0.0 (#111 / #114). - Fix bugs related to using the PostgreSQL JDBC driver instead of Amazon's official Redshift JDBC driver (#126, #143, #147). If your classpath contains both the PostgreSQL and Amazon drivers, explicitly specifying a JDBC driver class via the
jdbcdriver
parameter will now force that driver class to be used. - Give a better warning message for non-existing S3 buckets when attempting to read their bucket lifecycle configurations (#138 / #142).
- Minor documentation fixes: #119, #120, #123, #137.
Enhancements:
- Redshift queries are now cancelled when thread issuing the query is interrupted (#116 / #117). If you cancel a Databricks notebook shell while it is executing a
spark-redshift
query, the Spark REPL will no longer crash due to interrupts being swallowed. - When writing data back to Redshift, dates are now written in the default Redshift date format (
yyyy-MM-dd
) rather than a timestamp format (#122 / #130). spark-redshift
now implements Spark 1.6's newunhandledFilters
API, which allows Spark to eliminate a duplicate layer of filtering for filters that are pushed down to Redshift (#128).