Skip to content

Commit

Permalink
Update CHANGELOG for Spark client (#7959)
Browse files Browse the repository at this point in the history
  • Loading branch information
Isan-Rivkin authored Jul 4, 2024
1 parent a9ec2d7 commit b22e0e1
Show file tree
Hide file tree
Showing 4 changed files with 11 additions and 8 deletions.
3 changes: 3 additions & 0 deletions clients/spark/CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,8 @@
# Changelog

## v0.14.1 - 2024-07-04
* When scanning across all commits, correctly handle lakeFS storage namespaces that do not end in `/` (#7955)

## v0.14.0 - 2024-06-20
* Expose user metadata for commit (#7886)

Expand Down
2 changes: 1 addition & 1 deletion docs/howto/export.md
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ The complete `spark-submit` command would look as follows:
spark-submit --conf spark.hadoop.lakefs.api.url=https://<LAKEFS_ENDPOINT>/api/v1 \
--conf spark.hadoop.lakefs.api.access_key=<LAKEFS_ACCESS_KEY_ID> \
--conf spark.hadoop.lakefs.api.secret_key=<LAKEFS_SECRET_ACCESS_KEY> \
--packages io.lakefs:lakefs-spark-client_2.12:0.14.0 \
--packages io.lakefs:lakefs-spark-client_2.12:0.14.1 \
--class io.treeverse.clients.Main export-app example-repo s3://example-bucket/prefix \
--branch=example-branch
```
Expand Down
8 changes: 4 additions & 4 deletions docs/howto/garbage-collection/gc.md
Original file line number Diff line number Diff line change
Expand Up @@ -121,7 +121,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \
-c spark.hadoop.lakefs.api.secret_key=<LAKEFS_SECRET_KEY> \
-c spark.hadoop.fs.s3a.access.key=<S3_ACCESS_KEY> \
-c spark.hadoop.fs.s3a.secret.key=<S3_SECRET_KEY> \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.1/lakefs-spark-client-assembly-0.14.1.jar \
example-repo us-east-1
```
</div>
Expand All @@ -136,7 +136,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \
-c spark.hadoop.lakefs.api.access_key=<LAKEFS_ACCESS_KEY> \
-c spark.hadoop.lakefs.api.secret_key=<LAKEFS_SECRET_KEY> \
-c spark.hadoop.fs.azure.account.key.<AZURE_STORAGE_ACCOUNT>.dfs.core.windows.net=<AZURE_STORAGE_ACCESS_KEY> \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.1/lakefs-spark-client-assembly-0.14.1.jar \
example-repo
```

Expand All @@ -153,7 +153,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \
-c spark.hadoop.fs.azure.account.oauth2.client.id.<AZURE_STORAGE_ACCOUNT>.dfs.core.windows.net=<application-id> \
-c spark.hadoop.fs.azure.account.oauth2.client.secret.<AZURE_STORAGE_ACCOUNT>.dfs.core.windows.net=<service-credential-key> \
-c spark.hadoop.fs.azure.account.oauth2.client.endpoint.<AZURE_STORAGE_ACCOUNT>.dfs.core.windows.net=https://login.microsoftonline.com/<directory-id>/oauth2/token \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.1/lakefs-spark-client-assembly-0.14.1.jar \
example-repo
```

Expand Down Expand Up @@ -181,7 +181,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \
-c spark.hadoop.fs.gs.impl=com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystem \
-c spark.hadoop.fs.AbstractFileSystem.gs.impl=com.google.cloud.hadoop.fs.gcs.GoogleHadoopFS \
-c spark.hadoop.lakefs.gc.do_sweep=false \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.1/lakefs-spark-client-assembly-0.14.1.jar \
example-repo
```
Expand Down
6 changes: 3 additions & 3 deletions docs/reference/spark-client.md
Original file line number Diff line number Diff line change
Expand Up @@ -31,18 +31,18 @@ can work for other Spark versions and higher Hadoop versions.
Start Spark Shell / PySpark with the `--packages` flag, for instance:

```bash
spark-shell --packages io.lakefs:lakefs-spark-client_2.12:0.14.0
spark-shell --packages io.lakefs:lakefs-spark-client_2.12:0.14.1
```

Alternatively use the assembled jar (an "Überjar") on S3, from
`s3://treeverse-clients-us-east/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar`
`s3://treeverse-clients-us-east/lakefs-spark-client/0.14.1/lakefs-spark-client-assembly-0.14.1.jar`
by passing its path to `--jars`.
The assembled jar is larger but shades several common libraries. Use it if Spark
complains about bad classes or missing methods.
</div>
<div markdown="1" id="databricks">
Include this assembled jar (an "Überjar") from S3, from
`s3://treeverse-clients-us-east/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar`.
`s3://treeverse-clients-us-east/lakefs-spark-client/0.14.1/lakefs-spark-client-assembly-0.14.1.jar`.
</div>
</div>

Expand Down

0 comments on commit b22e0e1

Please sign in to comment.