Control matching behaviour with modifier functions.
Compare literal bytes in the string. This is very fast, but not usually what you want for non-ASCII character sets.
Compare strings respecting standard collation rules.
The default. Uses ICU regular expressions.
Match boundaries between things.
fixed(pattern, ignore_case = FALSE) coll(pattern, ignore_case = FALSE, locale = "en", ...) regex(pattern, ignore_case = FALSE, multiline = FALSE, comments = FALSE, dotall = FALSE, ...) boundary(type = c("character", "line_break", "sentence", "word"), skip_word_none = NA, ...)
pattern |
Pattern to modify behaviour. |
ignore_case |
Should case differences be ignored in the match? |
locale |
Locale to use for comparisons. See
|
... |
Other less frequently used arguments passed on to
|
multiline |
If |
comments |
If |
dotall |
If |
type |
Boundary type to detect.
|
skip_word_none |
Ignore "words" that don't contain any characters
or numbers - i.e. punctuation. Default |
str_wrap()
for breaking text to form paragraphs
stringi::stringi-search-boundaries
for more detail on the
various boundaries
pattern <- "a.b" strings <- c("abb", "a.b") str_detect(strings, pattern) str_detect(strings, fixed(pattern)) str_detect(strings, coll(pattern)) # coll() is useful for locale-aware case-insensitive matching i <- c("I", "\u0130", "i") i str_detect(i, fixed("i", TRUE)) str_detect(i, coll("i", TRUE)) str_detect(i, coll("i", TRUE, locale = "tr")) # Word boundaries words <- c("These are some words.") str_count(words, boundary("word")) str_split(words, " ")[[1]] str_split(words, boundary("word"))[[1]] # Regular expression variations str_extract_all("The Cat in the Hat", "[a-z]+") str_extract_all("The Cat in the Hat", regex("[a-z]+", TRUE)) str_extract_all("a\nb\nc", "^.") str_extract_all("a\nb\nc", regex("^.", multiline = TRUE)) str_extract_all("a\nb\nc", "a.") str_extract_all("a\nb\nc", regex("a.", dotall = TRUE))
Please choose more modern alternatives, such as Google Chrome or Mozilla Firefox.