library(sparklyr)
<- spark_connect(master = "local", version = "3.0.0")
sc <- sdf_len(sc, 1) %>% dplyr::mutate(m = map(1, 0, 2, 2, 3, -1))
sdf <- sdf %>% hof_map_filter(~ .x > .y)
filtered_sdf
Filters a map
R/dplyr_hof.R
hof_map_filter
Description
Filters entries in a map using the function specified (this is essentially a dplyr wrapper to the map_filter(expr, func)
higher- order function, which is supported since Spark 3.0)
Usage
hof_map_filter(x, func, expr = NULL, dest_col = NULL, ...)
Arguments
Arguments | Description |
---|---|
x | The Spark data frame to be processed |
func | The filter function to apply (it should take (key, value) as arguments and return a boolean value, with FALSE indicating the key-value pair should be discarded and TRUE otherwise) |
expr | The map being filtered, could be any SQL expression evaluating to a map (default: the last column of the Spark data frame) |
dest_col | Column to store the filtered result (default: expr) |
… | Additional params to dplyr::mutate |