Skip to content

Commit ea73be9

Browse files
authored
docs:Update README to use --enableSnappy flag to import snappy compre… (#3623)
…ssed snapshots. Thank you for opening a Pull Request! Before submitting your PR, there are a few things you can do to make sure it goes smoothly: - [ ] Make sure to open an issue as a [bug/issue](https://github.com/googleapis/java-bigtable-hbase/issues/new/choose) before writing your code! That way we can discuss the change, evaluate designs, and agree on the general idea - [ ] Ensure the tests and linter pass - [ ] Code coverage does not decrease (if any source code was changed) - [ ] Appropriate docs were updated (if necessary) Fixes #<issue_number_goes_here> ☕️ If you write sample code, please follow the [samples format]( https://github.com/GoogleCloudPlatform/java-docs-samples/blob/main/SAMPLE_FORMAT.md).
1 parent 8c1854d commit ea73be9

File tree

1 file changed

+8
-9
lines changed
  • bigtable-dataflow-parent/bigtable-beam-import

1 file changed

+8
-9
lines changed

bigtable-dataflow-parent/bigtable-beam-import/README.md

Lines changed: 8 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -103,7 +103,7 @@ Exporting HBase snapshots from Bigtable is not supported.
103103
```
104104
1. Run the export.
105105
```
106-
java -jar bigtable-beam-import-2.0.0.jar export \
106+
java -jar bigtable-beam-import-2.3.0.jar export \
107107
--runner=dataflow \
108108
--project=$PROJECT_ID \
109109
--bigtableInstanceId=$INSTANCE_ID \
@@ -143,15 +143,15 @@ Please pay attention to the Cluster CPU usage and adjust the number of Dataflow
143143
144144
1. Run the import.
145145
```
146-
java -jar bigtable-beam-import-2.0.0.jar importsnapshot \
146+
java -jar bigtable-beam-import-2.3.0.jar importsnapshot \
147147
--runner=DataflowRunner \
148148
--project=$PROJECT_ID \
149149
--bigtableInstanceId=$INSTANCE_ID \
150150
--bigtableTableId=$TABLE_NAME \
151151
--hbaseSnapshotSourceDir=$SNAPSHOT_GCS_PATH/data \
152152
--snapshotName=$SNAPSHOT_NAME \
153153
--stagingLocation=$SNAPSHOT_GCS_PATH/staging \
154-
--tempLocation=$SNAPSHOT_GCS_PATH/temp \
154+
--gcpTempLocation=$SNAPSHOT_GCS_PATH/temp \
155155
--maxWorkerNodes=$(expr 3 \* $CLUSTER_NUM_NODES) \
156156
--region=$REGION
157157
```
@@ -171,19 +171,18 @@ Please pay attention to the Cluster CPU usage and adjust the number of Dataflow
171171
172172
1. Run the import.
173173
```
174-
java -jar bigtable-beam-import-2.0.0.jar importsnapshot \
174+
java -jar bigtable-beam-import-2.3.0.jar importsnapshot \
175175
--runner=DataflowRunner \
176176
--project=$PROJECT_ID \
177177
--bigtableInstanceId=$INSTANCE_ID \
178178
--bigtableTableId=$TABLE_NAME \
179179
--hbaseSnapshotSourceDir=$SNAPSHOT_GCS_PATH/data \
180180
--snapshotName=$SNAPSHOT_NAME \
181181
--stagingLocation=$SNAPSHOT_GCS_PATH/staging \
182-
--tempLocation=$SNAPSHOT_GCS_PATH/temp \
182+
--gcpTempLocation=$SNAPSHOT_GCS_PATH/temp \
183183
--maxWorkerNodes=$(expr 3 \* $CLUSTER_NUM_NODES) \
184184
--region=$REGION \
185-
--experiments=use_runner_v2 \
186-
--sdkContainerImage=gcr.io/cloud-bigtable-ecosystem/unified-harness:latest
185+
--enableSnappy=true
187186
```
188187
189188
### Sequence Files
@@ -200,7 +199,7 @@ Please pay attention to the Cluster CPU usage and adjust the number of Dataflow
200199
```
201200
1. Run the import.
202201
```
203-
java -jar bigtable-beam-import-2.0.0.jar import \
202+
java -jar bigtable-beam-import-2.3.0.jar import \
204203
--runner=dataflow \
205204
--project=$PROJECT_ID \
206205
--bigtableInstanceId=$INSTANCE_ID \
@@ -228,7 +227,7 @@ check if there are any rows with mismatched data.
228227
```
229228
1. Run the sync job. It will put the results into `$SNAPSHOT_GCS_PATH/data-verification/output-TIMESTAMP`.
230229
```
231-
java -jar bigtable-beam-import-2.0.0.jar sync-table \
230+
java -jar bigtable-beam-import-2.3.0.jar sync-table \
232231
--runner=dataflow \
233232
--project=$PROJECT_ID \
234233
--bigtableInstanceId=$INSTANCE_ID \

0 commit comments

Comments
 (0)