Create a list of tokens
Usage
as_tokens(
tbl,
token_field = "token",
pos_field = get_dict_features()[1],
nm = NULL
)
Arguments
- tbl
A tibble of tokens out of
tokenize()
.- token_field
<
data-masked
> Column containing tokens.- pos_field
Column containing features that will be kept as the names of tokens. If you don't need them, give a
NULL
for this argument.- nm
Names of returned list. If left with
NULL
, "doc_id" field oftbl
is used instead.
Examples
if (FALSE) { # \dontrun{
tokenize(
data.frame(
doc_id = seq_along(5:8),
text = ginga[5:8]
)
) |>
prettify(col_select = "POS1") |>
as_tokens()
} # }