From dc0f01a1e25d0c7c859a4715b5db5c5feb5ed546 Mon Sep 17 00:00:00 2001 From: Jacek Laskowski Date: Thu, 17 Oct 2024 22:08:38 +0200 Subject: [PATCH] [MINOR] Use uc.version (0.3.0-SNAPSHOT can be built from the repo) --- docs/demo/crud-external-delta-table.md | 4 ++-- ...park-connector-and-external-tables-on-aws-s3.md | 2 +- docs/docker/index.md | 2 +- docs/features/index.md | 14 +++++++------- 4 files changed, 11 insertions(+), 11 deletions(-) diff --git a/docs/demo/crud-external-delta-table.md b/docs/demo/crud-external-delta-table.md index c1f3bd6..9a258e4 100644 --- a/docs/demo/crud-external-delta-table.md +++ b/docs/demo/crud-external-delta-table.md @@ -198,12 +198,12 @@ Table written to successfully at: file:///tmp/delta_table/ ## Spark Connector -=== "Spark 3.5.1 + Delta Lake 3.2.0" +=== "Spark 3.5.1 + Delta Lake {{ delta.version }}" ``` bash ./bin/spark-shell \ --packages \ - io.delta:delta-spark_2.13:3.2.0,io.unitycatalog:unitycatalog-spark:0.2.0-SNAPSHOT \ + io.delta:delta-spark_2.13:{{ delta.version }},io.unitycatalog:unitycatalog-spark:{{ uc.version }} \ --conf spark.sql.extensions=io.delta.sql.DeltaSparkSessionExtension \ --conf spark.sql.catalog.spark_catalog=io.unitycatalog.connectors.spark.UCSingleCatalog \ --conf spark.sql.catalog.spark_catalog.uri=http://localhost:8080 \ diff --git a/docs/demo/spark-connector-and-external-tables-on-aws-s3.md b/docs/demo/spark-connector-and-external-tables-on-aws-s3.md index 79572fd..1a9defe 100644 --- a/docs/demo/spark-connector-and-external-tables-on-aws-s3.md +++ b/docs/demo/spark-connector-and-external-tables-on-aws-s3.md @@ -11,7 +11,7 @@ hide: ``` bash ./bin/spark-sql \ --name "uc-external-tables-s3" \ - --packages "org.apache.hadoop:hadoop-aws:3.3.4,io.delta:delta-spark_{{ scala.version }}:{{ delta.version }},io.unitycatalog:unitycatalog-spark_{{ scala.version }}:0.2.0-SNAPSHOT" \ + --packages "org.apache.hadoop:hadoop-aws:3.3.4,io.delta:delta-spark_{{ scala.version }}:{{ delta.version }},io.unitycatalog:unitycatalog-spark_{{ scala.version }}:{{ uc.version }}" \ --conf "spark.sql.extensions=io.delta.sql.DeltaSparkSessionExtension" \ --conf "spark.sql.catalog.spark_catalog=io.unitycatalog.spark.UCSingleCatalog" \ --conf "spark.hadoop.fs.s3.impl=org.apache.hadoop.fs.s3a.S3AFileSystem" \ diff --git a/docs/docker/index.md b/docs/docker/index.md index a05e809..2b41f11 100644 --- a/docs/docker/index.md +++ b/docs/docker/index.md @@ -17,7 +17,7 @@ As of [this commit]({{ uc.commit }}/7564c7771ed655feb11ba3d5789fd3b7065a99ee), U ```console $ docker image ls --filter=reference='unitycatalog:*' REPOSITORY TAG IMAGE ID CREATED SIZE -unitycatalog 0.2.0-SNAPSHOT c788d0d453b6 8 minutes ago 329MB +unitycatalog {{ uc.version }} c788d0d453b6 8 minutes ago 329MB ``` ```console diff --git a/docs/features/index.md b/docs/features/index.md index dd1d673..8230ec8 100644 --- a/docs/features/index.md +++ b/docs/features/index.md @@ -47,8 +47,8 @@ OpenJDK Runtime Environment Zulu17.52+17-CA (build 17.0.12+7-LTS) OpenJDK 64-Bit Server VM Zulu17.52+17-CA (build 17.0.12+7-LTS, mixed mode, sharing) ``` -```text -$ ./bin/start-uc-server +``` text +❯ ./bin/start-uc-server ################################################################### # _ _ _ _ _____ _ _ # # | | | | (_) | / ____| | | | | # @@ -57,20 +57,20 @@ $ ./bin/start-uc-server # | |__| | | | | | |_| |_| | | |___| (_| | || (_| | | (_) | (_| | # # \____/|_| |_|_|\__|\__, | \_____\__,_|\__\__,_|_|\___/ \__, | # # __/ | __/ | # -# |___/ v0.2.0-SNAPSHOT |___/ # +# |___/ v0.3.0-SNAPSHOT |___/ # ################################################################### ``` -```text -$ ./bin/uc +``` text +❯ ./bin/uc Please provide a entity. Usage: bin/uc [options] -Entities: schema, volume, catalog, function, table +Entities: schema, volume, model_version, auth, catalog, function, permission, registered_model, user, table By default, the client will connect to UC running locally at http://localhost:8080 -To connect to specific UC server, use --server https:// +To connect to specific UC server, use --server https://: Currently, auth using bearer token is supported. Please specify the token via --auth_token