Catatan
Akses ke halaman ini memerlukan otorisasi. Anda dapat mencoba masuk atau mengubah direktori.
Akses ke halaman ini memerlukan otorisasi. Anda dapat mencoba mengubah direktori.
Periksa apakah nilai kolom berada di antara batas bawah dan atas (inklusif).
Sintaksis
between(lowerBound, upperBound)
Parameter-parameternya
| Parameter | Tipe | Deskripsi |
|---|---|---|
lowerBound |
nilai atau Kolom | Nilai terikat yang lebih rendah |
upperBound |
nilai atau Kolom | Nilai terikat atas |
Pengembalian Barang
Kolom (boolean)
Examples
Menggunakan antara dengan nilai bilangan bulat:
df = spark.createDataFrame([(2, "Alice"), (5, "Bob")], ["age", "name"])
df.select(df.name, df.age.between(2, 4)).show()
# +-----+---------------------------+
# | name|((age >= 2) AND (age <= 4))|
# +-----+---------------------------+
# |Alice| true|
# | Bob| false|
# +-----+---------------------------+
Menggunakan antara dengan nilai string:
df = spark.createDataFrame([("Alice", "A"), ("Bob", "B")], ["name", "initial"])
df.select(df.name, df.initial.between("A", "B")).show()
# +-----+-----------------------------------+
# | name|((initial >= A) AND (initial <= B))|
# +-----+-----------------------------------+
# |Alice| true|
# | Bob| true|
# +-----+-----------------------------------+
Menggunakan antara dengan nilai float:
df = spark.createDataFrame(
[(2.5, "Alice"), (5.5, "Bob")], ["height", "name"])
df.select(df.name, df.height.between(2.0, 5.0)).show()
# +-----+-------------------------------------+
# | name|((height >= 2.0) AND (height <= 5.0))|
# +-----+-------------------------------------+
# |Alice| true|
# | Bob| false|
# +-----+-------------------------------------+
Menggunakan antara dengan nilai tanggal:
import pyspark.sql.functions as sf
df = spark.createDataFrame(
[("Alice", "2023-01-01"), ("Bob", "2023-02-01")], ["name", "date"])
df = df.withColumn("date", sf.to_date(df.date))
df.select(df.name, df.date.between("2023-01-01", "2023-01-15")).show()
# +-----+-----------------------------------------------+
# | name|((date >= 2023-01-01) AND (date <= 2023-01-15))|
# +-----+-----------------------------------------------+
# |Alice| true|
# | Bob| false|
# +-----+-----------------------------------------------+
Menggunakan antara dengan nilai tanda waktu:
import pyspark.sql.functions as sf
df = spark.createDataFrame(
[("Alice", "2023-01-01 10:00:00"), ("Bob", "2023-02-01 10:00:00")],
schema=["name", "timestamp"])
df = df.withColumn("timestamp", sf.to_timestamp(df.timestamp))
df.select(df.name, df.timestamp.between("2023-01-01", "2023-02-01")).show()
# +-----+---------------------------------------------------------+
# | name|((timestamp >= 2023-01-01) AND (timestamp <= 2023-02-01))|
# +-----+---------------------------------------------------------+
# |Alice| true|
# | Bob| false|
# +-----+---------------------------------------------------------+