Nóta
Aðgangur að þessari síðu krefst heimildar. Þú getur prófað aðskrá þig inn eða breyta skráasöfnum.
Aðgangur að þessari síðu krefst heimildar. Þú getur prófað að breyta skráasöfnum.
Partition transform function: A transform for any type that partitions by a hash of the input column. Supports Spark Connect.
Warning
Deprecated in 4.0.0. Use partitioning.bucket instead.
Syntax
from pyspark.sql import functions as dbf
dbf.bucket(numBuckets=<numBuckets>, col=<col>)
Parameters
| Parameter | Type | Description |
|---|---|---|
numBuckets |
pyspark.sql.Column or int |
The number of buckets. |
col |
pyspark.sql.Column or str |
Target date or timestamp column to work on. |
Returns
pyspark.sql.Column: Data partitioned by given columns.
Examples
df.writeTo("catalog.db.table").partitionedBy(
bucket(42, "ts")
).createOrReplace()
Note
This function can be used only in combination with the partitionedBy method of the DataFrameWriterV2.