Tokenize a file/string.
Turns input into a character vector. Usually the tokenization is done purely in C++, and never exposed to R (because that requires a copy). This function is useful for testing, or when a file doesn't parse correctly and you want to see the underlying tokens.
tokenize(file, tokenizer = tokenizer_csv(), skip = 0, n_max = -1L)
file |
Either a path to a file, a connection, or literal data (either a single string or a raw vector). Files ending in Literal data is most useful for examples and tests. It must contain at least one new line to be recognised as data (instead of a path) or be a vector of greater than length 1. Using a value of |
tokenizer |
A tokenizer specification. |
skip |
Number of lines to skip before reading data. |
n_max |
Optionally, maximum number of rows to tokenize. |
tokenize("1,2\n3,4,5\n\n6") # Only tokenize first two lines tokenize("1,2\n3,4,5\n\n6", n = 2)
Please choose more modern alternatives, such as Google Chrome or Mozilla Firefox.