870 lines
35 KiB
Plaintext
870 lines
35 KiB
Plaintext
# Regular expressions {#sec-regular-expressions}
|
||
|
||
```{r}
|
||
#| echo: false
|
||
|
||
source("_common.R")
|
||
```
|
||
|
||
## Introduction
|
||
|
||
In @sec-strings, you learned a whole bunch of useful functions for working with strings.
|
||
This chapter will focus on functions that use **regular expressions**, a concise and powerful language for describing patterns within strings.
|
||
The term "regular expression" is a bit of a mouthful, so most people abbreviate it to "regex"[^regexps-1] or "regexp".
|
||
|
||
[^regexps-1]: You can pronounce it with either a hard-g (reg-x) or a soft-g (rej-x).
|
||
|
||
The chapter starts with the basics of regular expressions and the most useful stringr functions for data analysis.
|
||
We'll then expand your knowledge of patterns and cover seven important new topics (escaping, anchoring, character classes, shorthand classes, quantifiers, precedence, and grouping).
|
||
Next, we'll talk about some of the other types of patterns that stringr functions can work with and the various "flags" that allow you to tweak the operation of regular expressions.
|
||
We'll finish with a survey of other places in the tidyverse and base R where you might use regexes.
|
||
|
||
### Prerequisites
|
||
|
||
In this chapter, we'll use regular expression functions from stringr and tidyr, both core members of the tidyverse, as well as data from the babynames package.
|
||
|
||
```{r}
|
||
#| label: setup
|
||
#| message: false
|
||
|
||
library(tidyverse)
|
||
library(babynames)
|
||
```
|
||
|
||
Through this chapter, we'll use a mix of very simple inline examples so you can get the basic idea, the baby names data, and three character vectors from stringr:
|
||
|
||
- `fruit` contains the names of 80 fruits.
|
||
- `words` contains 980 common English words.
|
||
- `sentences` contains 720 short sentences.
|
||
|
||
## Pattern basics {#sec-reg-basics}
|
||
|
||
We'll use `str_view()` to learn how regex patterns work.
|
||
We used `str_view()` in the last chapter to better understand a string vs. its printed representation, and now we'll use it with its second argument, a regular expression.
|
||
When this is supplied, `str_view()` will show only the elements of the string vector that match, surrounding each match with `<>`, and, where possible, highlighting the match in blue.
|
||
|
||
The simplest patterns consist of letters and numbers which match those characters exactly:
|
||
|
||
```{r}
|
||
str_view(fruit, "berry")
|
||
```
|
||
|
||
Letters and numbers match exactly and are called **literal characters**.
|
||
Most punctuation characters, like `.`, `+`, `*`, `[`, `]`, and `?`, have special meanings[^regexps-2] and are called **metacharacters**. For example, `.`
|
||
will match any character[^regexps-3], so `"a."` will match any string that contains an "a" followed by another character
|
||
:
|
||
|
||
[^regexps-2]: You'll learn how to escape these special meanings in @sec-regexp-escaping.
|
||
|
||
[^regexps-3]: Well, any character apart from `\n`.
|
||
|
||
```{r}
|
||
str_view(c("a", "ab", "ae", "bd", "ea", "eab"), "a.")
|
||
```
|
||
|
||
Or we could find all the fruits that contain an "a", followed by three letters, followed by an "e":
|
||
|
||
```{r}
|
||
str_view(fruit, "a...e")
|
||
```
|
||
|
||
**Quantifiers** control how many times a pattern can match:
|
||
|
||
- `?` makes a pattern optional (i.e. it matches 0 or 1 times)
|
||
- `+` lets a pattern repeat (i.e. it matches at least once)
|
||
- `*` lets a pattern be optional or repeat (i.e. it matches any number of times, including 0).
|
||
|
||
```{r}
|
||
# ab? matches an "a", optionally followed by a "b".
|
||
str_view(c("a", "ab", "abb"), "ab?")
|
||
|
||
# ab+ matches an "a", followed by at least one "b".
|
||
str_view(c("a", "ab", "abb"), "ab+")
|
||
|
||
# ab* matches an "a", followed by any number of "b"s.
|
||
str_view(c("a", "ab", "abb"), "ab*")
|
||
```
|
||
|
||
**Character classes** are defined by `[]` and let you match a set of characters, e.g., `[abcd]` matches "a", "b", "c", or "d".
|
||
You can also invert the match by starting with `^`: `[^abcd]` matches anything **except** "a", "b", "c", or "d".
|
||
We can use this idea to find the words containing an "x" surrounded by vowels, or a "y" surrounded by consonants:
|
||
|
||
```{r}
|
||
str_view(words, "[aeiou]x[aeiou]")
|
||
str_view(words, "[^aeiou]y[^aeiou]")
|
||
```
|
||
|
||
You can use **alternation**, `|`, to pick between one or more alternative patterns.
|
||
For example, the following patterns look for fruits containing "apple", "melon", or "nut", or a repeated vowel.
|
||
|
||
```{r}
|
||
str_view(fruit, "apple|melon|nut")
|
||
str_view(fruit, "aa|ee|ii|oo|uu")
|
||
```
|
||
|
||
Regular expressions are very compact and use a lot of punctuation characters, so they can seem overwhelming and hard to read at first.
|
||
Don't worry; you'll get better with practice, and simple patterns will soon become second nature.
|
||
Let's kick off that process by practicing with some useful stringr functions.
|
||
|
||
## Key functions {#sec-stringr-regex-funs}
|
||
|
||
Now that you've got the basics of regular expressions under your belt, let's use them with some stringr and tidyr functions.
|
||
In the following section, you'll learn how to detect the presence or absence of a match, how to count the number of matches, how to replace a match with fixed text, and how to extract text using a pattern.
|
||
|
||
### Detect matches
|
||
|
||
`str_detect()` returns a logical vector that is `TRUE` if the pattern matches an element of the character vector and `FALSE` otherwise:
|
||
|
||
```{r}
|
||
str_detect(c("a", "b", "c"), "[aeiou]")
|
||
```
|
||
|
||
Since `str_detect()` returns a logical vector of the same length as the initial vector, it pairs well with `filter()`.
|
||
For example, this code finds all the most popular names containing a lower-case "x":
|
||
|
||
```{r}
|
||
babynames |>
|
||
filter(str_detect(name, "x")) |>
|
||
count(name, wt = n, sort = TRUE)
|
||
```
|
||
|
||
We can also use `str_detect()` with `summarize()` by pairing it with `sum()` or `mean()`: `sum(str_detect(x, pattern))` tells you the number of observations that match and `mean(str_detect(x, pattern))` tells you the proportion that match.
|
||
For example, the following snippet computes and visualizes the proportion of baby names[^regexps-4] that contain "x", broken down by year.
|
||
It looks like they've radically increased in popularity lately!
|
||
|
||
[^regexps-4]: This gives us the proportion of **names** that contain an "x"; if you wanted the proportion of babies with a name containing an x, you'd need to perform a weighted mean.
|
||
|
||
```{r}
|
||
#| fig-alt: |
|
||
#| A time series showing the proportion of baby names that contain the letter x.
|
||
#| The proportion declines gradually from 8 per 1000 in 1880 to 4 per 1000 in
|
||
#| 1980, then increases rapidly to 16 per 1000 in 2019.
|
||
|
||
babynames |>
|
||
group_by(year) |>
|
||
summarize(prop_x = mean(str_detect(name, "x"))) |>
|
||
ggplot(aes(x = year, y = prop_x)) +
|
||
geom_line()
|
||
```
|
||
|
||
There are two functions that are closely related to `str_detect()`: `str_subset()` and `str_which()`.
|
||
`str_subset()` returns a character vector containing only the strings that match.
|
||
`str_which()` returns an integer vector giving the positions of the strings that match.
|
||
|
||
### Count matches
|
||
|
||
The next step up in complexity from `str_detect()` is `str_count()`: rather than a true or false, it tells you how many matches there are in each string.
|
||
|
||
```{r}
|
||
x <- c("apple", "banana", "pear")
|
||
str_count(x, "p")
|
||
```
|
||
|
||
Note that each match starts at the end of the previous match, i.e. regex matches never overlap.
|
||
For example, in `"abababa"`, how many times will the pattern `"aba"` match?
|
||
Regular expressions say two, not three:
|
||
|
||
```{r}
|
||
str_count("abababa", "aba")
|
||
str_view("abababa", "aba")
|
||
```
|
||
|
||
It's natural to use `str_count()` with `mutate()`.
|
||
The following example uses `str_count()` with character classes to count the number of vowels and consonants in each name.
|
||
|
||
```{r}
|
||
babynames |>
|
||
count(name) |>
|
||
mutate(
|
||
vowels = str_count(name, "[aeiou]"),
|
||
consonants = str_count(name, "[^aeiou]")
|
||
)
|
||
```
|
||
|
||
If you look closely, you'll notice that there's something off with our calculations: "Aaban" contains three "a"s, but our summary reports only two vowels.
|
||
That's because regular expressions are case sensitive.
|
||
There are three ways we could fix this:
|
||
|
||
- Add the upper case vowels to the character class: `str_count(name, "[aeiouAEIOU]")`.
|
||
- Tell the regular expression to ignore case: `str_count(name, regex("[aeiou]", ignore_case = TRUE))`. We'll talk about more in @sec-flags.
|
||
- Use `str_to_lower()` to convert the names to lower case: `str_count(str_to_lower(name), "[aeiou]")`.
|
||
|
||
This variety of approaches is pretty typical when working with strings --- there are often multiple ways to reach your goal, either by making your pattern more complicated or by doing some preprocessing on your string.
|
||
If you get stuck trying one approach, it can often be useful to switch gears and tackle the problem from a different perspective.
|
||
|
||
In this case, since we're applying two functions to the name, I think it's easier to transform it first:
|
||
|
||
```{r}
|
||
babynames |>
|
||
count(name) |>
|
||
mutate(
|
||
name = str_to_lower(name),
|
||
vowels = str_count(name, "[aeiou]"),
|
||
consonants = str_count(name, "[^aeiou]")
|
||
)
|
||
```
|
||
|
||
### Replace values
|
||
|
||
As well as detecting and counting matches, we can also modify them with `str_replace()` and `str_replace_all()`.
|
||
`str_replace()` replaces the first match, and as the name suggests, `str_replace_all()` replaces all matches.
|
||
|
||
```{r}
|
||
x <- c("apple", "pear", "banana")
|
||
str_replace_all(x, "[aeiou]", "-")
|
||
```
|
||
|
||
`str_remove()` and `str_remove_all()` are handy shortcuts for `str_replace(x, pattern, "")`:
|
||
|
||
```{r}
|
||
x <- c("apple", "pear", "banana")
|
||
str_remove_all(x, "[aeiou]")
|
||
```
|
||
|
||
These functions are naturally paired with `mutate()` when doing data cleaning, and you'll often apply them repeatedly to peel off layers of inconsistent formatting.
|
||
|
||
### Extract variables {#sec-extract-variables}
|
||
|
||
The last function we'll discuss uses regular expressions to extract data out of one column into one or more new columns: `separate_wider_regex()`.
|
||
It's a peer of the `separate_wider_position()` and `separate_wider_delim()` functions that you learned about in @sec-string-columns.
|
||
These functions live in tidyr because they operate on (columns of) data frames, rather than individual vectors.
|
||
|
||
Let's create a simple dataset to show how it works.
|
||
Here we have some data derived from `babynames` where we have the name, gender, and age of a bunch of people in a rather weird format[^regexps-5]:
|
||
|
||
[^regexps-5]: We wish we could reassure you that you'd never see something this weird in real life, but unfortunately over the course of your career you're likely to see much weirder!
|
||
|
||
```{r}
|
||
df <- tribble(
|
||
~str,
|
||
"<Sheryl>-F_34",
|
||
"<Kisha>-F_45",
|
||
"<Brandon>-N_33",
|
||
"<Sharon>-F_38",
|
||
"<Penny>-F_58",
|
||
"<Justin>-M_41",
|
||
"<Patricia>-F_84",
|
||
)
|
||
```
|
||
|
||
To extract this data using `separate_wider_regex()` we just need to construct a sequence of regular expressions that match each piece.
|
||
If we want the contents of that piece to appear in the output, we give it a name:
|
||
|
||
```{r}
|
||
df |>
|
||
separate_wider_regex(
|
||
str,
|
||
patterns = c(
|
||
"<",
|
||
name = "[A-Za-z]+",
|
||
">-",
|
||
gender = ".",
|
||
"_",
|
||
age = "[0-9]+"
|
||
)
|
||
)
|
||
```
|
||
|
||
If the match fails, you can use `too_few = "debug"` to figure out what went wrong, just like `separate_wider_delim()` and `separate_wider_position()`.
|
||
|
||
### Exercises
|
||
|
||
1. What baby name has the most vowels?
|
||
What name has the highest proportion of vowels?
|
||
(Hint: what is the denominator?)
|
||
|
||
2. Replace all forward slashes in `"a/b/c/d/e"` with backslashes.
|
||
What happens if you attempt to undo the transformation by replacing all backslashes with forward slashes?
|
||
(We'll discuss the problem very soon.)
|
||
|
||
3. Implement a simple version of `str_to_lower()` using `str_replace_all()`.
|
||
|
||
4. Create a regular expression that will match telephone numbers as commonly written in your country.
|
||
|
||
## Pattern details
|
||
|
||
Now that you understand the basics of the pattern language and how to use it with some stringr and tidyr functions, it's time to dig into more of the details.
|
||
First, we'll start with **escaping**, which allows you to match metacharacters that would otherwise be treated specially.
|
||
Next, you'll learn about **anchors** which allow you to match the start or end of the string.
|
||
Then, you'll learn more about **character classes** and their shortcuts which allow you to match any character from a set.
|
||
Next, you'll learn the final details of **quantifiers** which control how many times a pattern can match.
|
||
Then, we have to cover the important (but complex) topic of **operator precedence** and parentheses.
|
||
And we'll finish off with some details of **grouping** components of the pattern.
|
||
|
||
The terms we use here are the technical names for each component.
|
||
They're not always the most evocative of their purpose, but it's very helpful to know the correct terms if you later want to Google for more details.
|
||
|
||
### Escaping {#sec-regexp-escaping}
|
||
|
||
In order to match a literal `.`, you need an **escape** which tells the regular expression to match metacharacters[^regexps-6] literally.
|
||
Like strings, regexps use the backslash for escaping.
|
||
So, to match a `.`, you need the regexp `\.`. Unfortunately this creates a problem.
|
||
We use strings to represent regular expressions, and `\` is also used as an escape symbol in strings.
|
||
So to create the regular expression `\.` we need the string `"\\."`, as the following example shows.
|
||
|
||
[^regexps-6]: The complete set of metacharacters is `.^$\|*+?{}[]()`
|
||
|
||
```{r}
|
||
# To create the regular expression \., we need to use \\.
|
||
dot <- "\\."
|
||
|
||
# But the expression itself only contains one \
|
||
str_view(dot)
|
||
|
||
# And this tells R to look for an explicit .
|
||
str_view(c("abc", "a.c", "bef"), "a\\.c")
|
||
```
|
||
|
||
In this book, we'll usually write regular expression without quotes, like `\.`.
|
||
If we need to emphasize what you'll actually type, we'll surround it with quotes and add extra escapes, like `"\\."`.
|
||
|
||
If `\` is used as an escape character in regular expressions, how do you match a literal `\`?
|
||
Well, you need to escape it, creating the regular expression `\\`.
|
||
To create that regular expression, you need to use a string, which also needs to escape `\`.
|
||
That means to match a literal `\` you need to write `"\\\\"` --- you need four backslashes to match one!
|
||
|
||
```{r}
|
||
x <- "a\\b"
|
||
str_view(x)
|
||
str_view(x, "\\\\")
|
||
```
|
||
|
||
Alternatively, you might find it easier to use the raw strings you learned about in @sec-raw-strings).
|
||
That lets you avoid one layer of escaping:
|
||
|
||
```{r}
|
||
str_view(x, r"{\\}")
|
||
```
|
||
|
||
If you're trying to match a literal `.`, `$`, `|`, `*`, `+`, `?`, `{`, `}`, `(`, `)`, there's an alternative to using a backslash escape: you can use a character class: `[.]`, `[$]`, `[|]`, ...
|
||
all match the literal values.
|
||
|
||
```{r}
|
||
str_view(c("abc", "a.c", "a*c", "a c"), "a[.]c")
|
||
str_view(c("abc", "a.c", "a*c", "a c"), ".[*]c")
|
||
```
|
||
|
||
### Anchors
|
||
|
||
By default, regular expressions will match any part of a string.
|
||
If you want to match at the start or end you need to **anchor** the regular expression using `^` to match the start or `$` to match the end:
|
||
|
||
```{r}
|
||
str_view(fruit, "^a")
|
||
str_view(fruit, "a$")
|
||
```
|
||
|
||
It's tempting to think that `$` should match the start of a string, because that's how we write dollar amounts, but that's not what regular expressions want.
|
||
|
||
To force a regular expression to match only the full string, anchor it with both `^` and `$`:
|
||
|
||
```{r}
|
||
str_view(fruit, "apple")
|
||
str_view(fruit, "^apple$")
|
||
```
|
||
|
||
You can also match the boundary between words (i.e. the start or end of a word) with `\b`.
|
||
This can be particularly useful when using RStudio's find and replace tool.
|
||
For example, if to find all uses of `sum()`, you can search for `\bsum\b` to avoid matching `summarize`, `summary`, `rowsum` and so on:
|
||
|
||
```{r}
|
||
x <- c("summary(x)", "summarize(df)", "rowsum(x)", "sum(x)")
|
||
str_view(x, "sum")
|
||
str_view(x, "\\bsum\\b")
|
||
```
|
||
|
||
When used alone, anchors will produce a zero-width match:
|
||
|
||
```{r}
|
||
str_view("abc", c("$", "^", "\\b"))
|
||
```
|
||
|
||
This helps you understand what happens when you replace a standalone anchor:
|
||
|
||
```{r}
|
||
str_replace_all("abc", c("$", "^", "\\b"), "--")
|
||
```
|
||
|
||
### Character classes
|
||
|
||
A **character class**, or character **set**, allows you to match any character in a set.
|
||
As we discussed above, you can construct your own sets with `[]`, where `[abc]` matches "a", "b", or "c" and `[^abc]` matches any character except "a", "b", or "c".
|
||
Apart from `^` there are two other characters that have special meaning inside of `[]:`
|
||
|
||
- `-` defines a range, e.g., `[a-z]` matches any lower case letter and `[0-9]` matches any number.
|
||
- `\` escapes special characters, so `[\^\-\]]` matches `^`, `-`, or `]`.
|
||
|
||
Here are few examples:
|
||
|
||
```{r}
|
||
x <- "abcd ABCD 12345 -!@#%."
|
||
str_view(x, "[abc]+")
|
||
str_view(x, "[a-z]+")
|
||
str_view(x, "[^a-z0-9]+")
|
||
|
||
# You need an escape to match characters that are otherwise
|
||
# special inside of []
|
||
str_view("a-b-c", "[a-c]")
|
||
str_view("a-b-c", "[a\\-c]")
|
||
```
|
||
|
||
Some character classes are used so commonly that they get their own shortcut.
|
||
You've already seen `.`, which matches any character apart from a newline.
|
||
There are three other particularly useful pairs[^regexps-7]:
|
||
|
||
[^regexps-7]: Remember, to create a regular expression containing `\d` or `\s`, you'll need to escape the `\` for the string, so you'll type `"\\d"` or `"\\s"`.
|
||
|
||
- `\d` matches any digit;\
|
||
`\D` matches anything that isn't a digit.
|
||
- `\s` matches any whitespace (e.g., space, tab, newline);\
|
||
`\S` matches anything that isn't whitespace.
|
||
- `\w` matches any "word" character, i.e. letters and numbers;\
|
||
`\W` matches any "non-word" character.
|
||
|
||
The following code demonstrates the six shortcuts with a selection of letters, numbers, and punctuation characters.
|
||
|
||
```{r}
|
||
x <- "abcd ABCD 12345 -!@#%."
|
||
str_view(x, "\\d+")
|
||
str_view(x, "\\D+")
|
||
str_view(x, "\\s+")
|
||
str_view(x, "\\S+")
|
||
str_view(x, "\\w+")
|
||
str_view(x, "\\W+")
|
||
```
|
||
|
||
### Quantifiers {#sec-quantifiers}
|
||
|
||
**Quantifiers** control how many times a pattern matches.
|
||
In @sec-reg-basics you learned about `?` (0 or 1 matches), `+` (1 or more matches), and `*` (0 or more matches).
|
||
For example, `colou?r` will match American or British spelling, `\d+` will match one or more digits, and `\s?` will optionally match a single item of whitespace.
|
||
You can also specify the number of matches precisely with `{}`:
|
||
|
||
- `{n}` matches exactly n times.
|
||
- `{n,}` matches at least n times.
|
||
- `{n,m}` matches between n and m times.
|
||
|
||
### Operator precedence and parentheses
|
||
|
||
What does `ab+` match?
|
||
Does it match "a" followed by one or more "b"s, or does it match "ab" repeated any number of times?
|
||
What does `^a|b$` match?
|
||
Does it match the complete string a or the complete string b, or does it match a string starting with a or a string ending with b?
|
||
|
||
The answer to these questions is determined by operator precedence, similar to the PEMDAS or BEDMAS rules you might have learned in school.
|
||
You know that `a + b * c` is equivalent to `a + (b * c)` not `(a + b) * c` because `*` has higher precedence and `+` has lower precedence: you compute `*` before `+`.
|
||
|
||
Similarly, regular expressions have their own precedence rules: quantifiers have high precedence and alternation has low precedence which means that `ab+` is equivalent to `a(b+)`, and `^a|b$` is equivalent to `(^a)|(b$)`.
|
||
Just like with algebra, you can use parentheses to override the usual order.
|
||
But unlike algebra you're unlikely to remember the precedence rules for regexes, so feel free to use parentheses liberally.
|
||
|
||
### Grouping and capturing
|
||
|
||
As well as overriding operator precedence, parentheses have another important effect: they create **capturing groups** that allow you to use sub-components of the match.
|
||
|
||
The first way to use a capturing group is to refer back to it within a match with **back reference**: `\1` refers to the match contained in the first parenthesis, `\2` in the second parenthesis, and so on.
|
||
For example, the following pattern finds all fruits that have a repeated pair of letters:
|
||
|
||
```{r}
|
||
str_view(fruit, "(..)\\1")
|
||
```
|
||
|
||
And this one finds all words that start and end with the same pair of letters:
|
||
|
||
```{r}
|
||
str_view(words, "^(..).*\\1$")
|
||
```
|
||
|
||
You can also use back references in `str_replace()`.
|
||
For example, this code switches the order of the second and third words in `sentences`:
|
||
|
||
```{r}
|
||
sentences |>
|
||
str_replace("(\\w+) (\\w+) (\\w+)", "\\1 \\3 \\2") |>
|
||
str_view()
|
||
```
|
||
|
||
If you want to extract the matches for each group you can use `str_match()`.
|
||
But `str_match()` returns a matrix, so it's not particularly easy to work with[^regexps-8]:
|
||
|
||
[^regexps-8]: Mostly because we never discuss matrices in this book!
|
||
|
||
```{r}
|
||
sentences |>
|
||
str_match("the (\\w+) (\\w+)") |>
|
||
head()
|
||
```
|
||
|
||
You could convert to a tibble and name the columns:
|
||
|
||
```{r}
|
||
sentences |>
|
||
str_match("the (\\w+) (\\w+)") |>
|
||
as_tibble(.name_repair = "minimal") |>
|
||
set_names("match", "word1", "word2")
|
||
```
|
||
|
||
But then you've basically recreated your own version of `separate_wider_regex()`.
|
||
Indeed, behind the scenes, `separate_wider_regex()` converts your vector of patterns to a single regex that uses grouping to capture the named components.
|
||
|
||
Occasionally, you'll want to use parentheses without creating matching groups.
|
||
You can create a non-capturing group with `(?:)`.
|
||
|
||
```{r}
|
||
x <- c("a gray cat", "a grey dog")
|
||
str_match(x, "gr(e|a)y")
|
||
str_match(x, "gr(?:e|a)y")
|
||
```
|
||
|
||
### Exercises
|
||
|
||
1. How would you match the literal string `"'\`? How about `"$^$"`?
|
||
|
||
2. Explain why each of these patterns don't match a `\`: `"\"`, `"\\"`, `"\\\"`.
|
||
|
||
3. Given the corpus of common words in `stringr::words`, create regular expressions that find all words that:
|
||
|
||
a. Start with "y".
|
||
b. Don't start with "y".
|
||
c. End with "x".
|
||
d. Are exactly three letters long. (Don't cheat by using `str_length()`!)
|
||
e. Have seven letters or more.
|
||
f. Contain a vowel-consonant pair.
|
||
g. Contain at least two vowel-consonant pairs in a row.
|
||
h. Only consist of repeated vowel-consonant pairs.
|
||
|
||
4. Create 11 regular expressions that match the British or American spellings for each of the following words: airplane/aeroplane, aluminum/aluminium, analog/analogue, ass/arse, center/centre, defense/defence, donut/doughnut, gray/grey, modeling/modelling, skeptic/sceptic, summarize/summarise.
|
||
Try and make the shortest possible regex!
|
||
|
||
5. Switch the first and last letters in `words`.
|
||
Which of those strings are still `words`?
|
||
|
||
6. Describe in words what these regular expressions match: (read carefully to see if each entry is a regular expression or a string that defines a regular expression.)
|
||
|
||
a. `^.*$`
|
||
b. `"\\{.+\\}"`
|
||
c. `\d{4}-\d{2}-\d{2}`
|
||
d. `"\\\\{4}"`
|
||
e. `\..\..\..`
|
||
f. `(.)\1\1`
|
||
g. `"(..)\\1"`
|
||
|
||
7. Solve the beginner regexp crosswords at <https://regexcrossword.com/challenges/beginner>.
|
||
|
||
## Pattern control
|
||
|
||
It's possible to exercise extra control over the details of the match by using a pattern object instead of just a string.
|
||
This allows you to control the so called regex flags and match various types of fixed strings, as described below.
|
||
|
||
### Regex flags {#sec-flags}
|
||
|
||
There are a number of settings that can be used to control the details of the regexp.
|
||
These settings are often called **flags** in other programming languages.
|
||
In stringr, you can use these by wrapping the pattern in a call to `regex()`.
|
||
The most useful flag is probably `ignore_case = TRUE` because it allows characters to match either their uppercase or lowercase forms:
|
||
|
||
```{r}
|
||
bananas <- c("banana", "Banana", "BANANA")
|
||
str_view(bananas, "banana")
|
||
str_view(bananas, regex("banana", ignore_case = TRUE))
|
||
```
|
||
|
||
If you're doing a lot of work with multiline strings (i.e. strings that contain `\n`), `dotall`and `multiline` may also be useful:
|
||
|
||
- `dotall = TRUE` lets `.` match everything, including `\n`:
|
||
|
||
```{r}
|
||
x <- "Line 1\nLine 2\nLine 3"
|
||
str_view(x, ".Line")
|
||
str_view(x, regex(".Line", dotall = TRUE))
|
||
```
|
||
|
||
- `multiline = TRUE` makes `^` and `$` match the start and end of each line rather than the start and end of the complete string:
|
||
|
||
```{r}
|
||
x <- "Line 1\nLine 2\nLine 3"
|
||
str_view(x, "^Line")
|
||
str_view(x, regex("^Line", multiline = TRUE))
|
||
```
|
||
|
||
Finally, if you're writing a complicated regular expression and you're worried you might not understand it in the future, you might try `comments = TRUE`.
|
||
It tweaks the pattern language to ignore spaces and new lines, as well as everything after `#`.
|
||
This allows you to use comments and whitespace to make complex regular expressions more understandable[^regexps-9], as in the following example:
|
||
|
||
[^regexps-9]: `comments = TRUE` is particularly effective in combination with a raw string, as we use here.
|
||
|
||
```{r}
|
||
phone <- regex(
|
||
r"(
|
||
\(? # optional opening parens
|
||
(\d{3}) # area code
|
||
[)\-]? # optional closing parens or dash
|
||
\ ? # optional space
|
||
(\d{3}) # another three numbers
|
||
[\ -]? # optional space or dash
|
||
(\d{4}) # four more numbers
|
||
)",
|
||
comments = TRUE
|
||
)
|
||
|
||
str_extract(c("514-791-8141", "(123) 456 7890", "123456"), phone)
|
||
```
|
||
|
||
If you're using comments and want to match a space, newline, or `#`, you'll need to escape it with `\`.
|
||
|
||
### Fixed matches
|
||
|
||
You can opt-out of the regular expression rules by using `fixed()`:
|
||
|
||
```{r}
|
||
str_view(c("", "a", "."), fixed("."))
|
||
```
|
||
|
||
`fixed()` also gives you the ability to ignore case:
|
||
|
||
```{r}
|
||
str_view("x X", "X")
|
||
str_view("x X", fixed("X", ignore_case = TRUE))
|
||
```
|
||
|
||
If you're working with non-English text, you will probably want `coll()` instead of `fixed()`, as it implements the full rules for capitalization as used by the `locale` you specify.
|
||
See @sec-other-languages for more details on locales.
|
||
|
||
```{r}
|
||
str_view("i İ ı I", fixed("İ", ignore_case = TRUE))
|
||
str_view("i İ ı I", coll("İ", ignore_case = TRUE, locale = "tr"))
|
||
```
|
||
|
||
## Practice
|
||
|
||
To put these ideas into practice we'll solve a few semi-authentic problems next.
|
||
We'll discuss three general techniques:
|
||
|
||
1. checking your work by creating simple positive and negative controls
|
||
2. combining regular expressions with Boolean algebra
|
||
3. creating complex patterns using string manipulation
|
||
|
||
### Check your work
|
||
|
||
First, let's find all sentences that start with "The".
|
||
Using the `^` anchor alone is not enough:
|
||
|
||
```{r}
|
||
str_view(sentences, "^The")
|
||
```
|
||
|
||
Because that pattern also matches sentences starting with words like `They` or `These`.
|
||
We need to make sure that the "e" is the last letter in the word, which we can do by adding a word boundary:
|
||
|
||
```{r}
|
||
str_view(sentences, "^The\\b")
|
||
```
|
||
|
||
What about finding all sentences that begin with a pronoun?
|
||
|
||
```{r}
|
||
str_view(sentences, "^She|He|It|They\\b")
|
||
```
|
||
|
||
A quick inspection of the results shows that we're getting some spurious matches.
|
||
That's because we've forgotten to use parentheses:
|
||
|
||
```{r}
|
||
str_view(sentences, "^(She|He|It|They)\\b")
|
||
```
|
||
|
||
You might wonder how you might spot such a mistake if it didn't occur in the first few matches.
|
||
A good technique is to create a few positive and negative matches and use them to test that your pattern works as expected:
|
||
|
||
```{r}
|
||
pos <- c("He is a boy", "She had a good time")
|
||
neg <- c("Shells come from the sea", "Hadley said 'It's a great day'")
|
||
|
||
pattern <- "^(She|He|It|They)\\b"
|
||
str_detect(pos, pattern)
|
||
str_detect(neg, pattern)
|
||
```
|
||
|
||
It's typically much easier to come up with good positive examples than negative examples, because it takes a while before you're good enough with regular expressions to predict where your weaknesses are.
|
||
Nevertheless, they're still useful: as you work on the problem you can slowly accumulate a collection of your mistakes, ensuring that you never make the same mistake twice.
|
||
|
||
### Boolean operations {#sec-boolean-operations}
|
||
|
||
Imagine we want to find words that only contain consonants.
|
||
One technique is to create a character class that contains all letters except for the vowels (`[^aeiou]`), then allow that to match any number of letters (`[^aeiou]+`), then force it to match the whole string by anchoring to the beginning and the end (`^[^aeiou]+$`):
|
||
|
||
```{r}
|
||
str_view(words, "^[^aeiou]+$")
|
||
```
|
||
|
||
But you can make this problem a bit easier by flipping the problem around.
|
||
Instead of looking for words that contain only consonants, we could look for words that don't contain any vowels:
|
||
|
||
```{r}
|
||
str_view(words[!str_detect(words, "[aeiou]")])
|
||
```
|
||
|
||
This is a useful technique whenever you're dealing with logical combinations, particularly those involving "and" or "not".
|
||
For example, imagine if you want to find all words that contain "a" and "b".
|
||
There's no "and" operator built in to regular expressions so we have to tackle it by looking for all words that contain an "a" followed by a "b", or a "b" followed by an "a":
|
||
|
||
```{r}
|
||
str_view(words, "a.*b|b.*a")
|
||
```
|
||
|
||
It's simpler to combine the results of two calls to `str_detect()`:
|
||
|
||
```{r}
|
||
words[str_detect(words, "a") & str_detect(words, "b")]
|
||
```
|
||
|
||
What if we wanted to see if there was a word that contains all vowels?
|
||
If we did it with patterns we'd need to generate 5!
|
||
(120) different patterns:
|
||
|
||
```{r}
|
||
#| results: false
|
||
words[str_detect(words, "a.*e.*i.*o.*u")]
|
||
# ...
|
||
words[str_detect(words, "u.*o.*i.*e.*a")]
|
||
```
|
||
|
||
It's much simpler to combine five calls to `str_detect()`:
|
||
|
||
```{r}
|
||
words[
|
||
str_detect(words, "a") &
|
||
str_detect(words, "e") &
|
||
str_detect(words, "i") &
|
||
str_detect(words, "o") &
|
||
str_detect(words, "u")
|
||
]
|
||
```
|
||
|
||
In general, if you get stuck trying to create a single regexp that solves your problem, take a step back and think if you could break the problem down into smaller pieces, solving each challenge before moving onto the next one.
|
||
|
||
### Creating a pattern with code
|
||
|
||
What if we wanted to find all `sentences` that mention a color?
|
||
The basic idea is simple: we just combine alternation with word boundaries.
|
||
|
||
```{r}
|
||
str_view(sentences, "\\b(red|green|blue)\\b")
|
||
```
|
||
|
||
But as the number of colors grows, it would quickly get tedious to construct this pattern by hand.
|
||
Wouldn't it be nice if we could store the colors in a vector?
|
||
|
||
```{r}
|
||
rgb <- c("red", "green", "blue")
|
||
```
|
||
|
||
Well, we can!
|
||
We'd just need to create the pattern from the vector using `str_c()` and `str_flatten()`:
|
||
|
||
```{r}
|
||
str_c("\\b(", str_flatten(rgb, "|"), ")\\b")
|
||
```
|
||
|
||
We could make this pattern more comprehensive if we had a good list of colors.
|
||
One place we could start from is the list of built-in colors that R can use for plots:
|
||
|
||
```{r}
|
||
str_view(colors())
|
||
```
|
||
|
||
But lets first eliminate the numbered variants:
|
||
|
||
```{r}
|
||
cols <- colors()
|
||
cols <- cols[!str_detect(cols, "\\d")]
|
||
str_view(cols)
|
||
```
|
||
|
||
Then we can turn this into one giant pattern.
|
||
We won't show the pattern here because it's huge, but you can see it working:
|
||
|
||
```{r}
|
||
pattern <- str_c("\\b(", str_flatten(cols, "|"), ")\\b")
|
||
str_view(sentences, pattern)
|
||
```
|
||
|
||
In this example, `cols` only contains numbers and letters so you don't need to worry about metacharacters.
|
||
But in general, whenever you create patterns from existing strings it's wise to run them through `str_escape()` to ensure they match literally.
|
||
|
||
### Exercises
|
||
|
||
1. For each of the following challenges, try solving it by using both a single regular expression, and a combination of multiple `str_detect()` calls.
|
||
|
||
a. Find all `words` that start or end with `x`.
|
||
b. Find all `words` that start with a vowel and end with a consonant.
|
||
c. Are there any `words` that contain at least one of each different vowel?
|
||
|
||
2. Construct patterns to find evidence for and against the rule "i before e except after c"?
|
||
|
||
3. `colors()` contains a number of modifiers like "lightgray" and "darkblue".
|
||
How could you automatically identify these modifiers?
|
||
(Think about how you might detect and then remove the colors that are modified).
|
||
|
||
4. Create a regular expression that finds any base R dataset.
|
||
You can get a list of these datasets via a special use of the `data()` function: `data(package = "datasets")$results[, "Item"]`.
|
||
Note that a number of old datasets are individual vectors; these contain the name of the grouping "data frame" in parentheses, so you'll need to strip those off.
|
||
|
||
## Regular expressions in other places
|
||
|
||
Just like in the stringr and tidyr functions, there are many other places in R where you can use regular expressions.
|
||
The following sections describe some other useful functions in the wider tidyverse and base R.
|
||
|
||
### tidyverse
|
||
|
||
There are three other particularly useful places where you might want to use a regular expressions
|
||
|
||
- `matches(pattern)` will select all variables whose name matches the supplied pattern.
|
||
It's a "tidyselect" function that you can use anywhere in any tidyverse function that selects variables (e.g., `select()`, `rename_with()` and `across()`).
|
||
|
||
- `pivot_longer()'s` `names_pattern` argument takes a vector of regular expressions, just like `separate_wider_regex()`.
|
||
It's useful when extracting data out of variable names with a complex structure
|
||
|
||
- The `delim` argument in `separate_longer_delim()` and `separate_wider_delim()` usually matches a fixed string, but you can use `regex()` to make it match a pattern.
|
||
This is useful, for example, if you want to match a comma that is optionally followed by a space, i.e. `regex(", ?")`.
|
||
|
||
### Base R
|
||
|
||
`apropos(pattern)` searches all objects available from the global environment that match the given pattern.
|
||
This is useful if you can't quite remember the name of a function:
|
||
|
||
```{r}
|
||
apropos("replace")
|
||
```
|
||
|
||
`list.files(path, pattern)` lists all files in `path` that match a regular expression `pattern`.
|
||
For example, you can find all the R Markdown files in the current directory with:
|
||
|
||
```{r}
|
||
head(list.files(pattern = "\\.Rmd$"))
|
||
```
|
||
|
||
It's worth noting that the pattern language used by base R is very slightly different to that used by stringr.
|
||
That's because stringr is built on top of the [stringi package](https://stringi.gagolewski.com), which is in turn built on top of the [ICU engine](https://unicode-org.github.io/icu/userguide/strings/regexp.html), whereas base R functions use either the [TRE engine](https://github.com/laurikari/tre) or the [PCRE engine](https://www.pcre.org), depending on whether or not you've set `perl = TRUE`.
|
||
Fortunately, the basics of regular expressions are so well established that you'll encounter few variations when working with the patterns you'll learn in this book.
|
||
You only need to be aware of the difference when you start to rely on advanced features like complex Unicode character ranges or special features that use the `(?…)` syntax.
|
||
|
||
## Summary
|
||
|
||
With every punctuation character potentially overloaded with meaning, regular expressions are one of the most compact languages out there.
|
||
They're definitely confusing at first but as you train your eyes to read them and your brain to understand them, you unlock a powerful skill that you can use in R and in many other places.
|
||
|
||
In this chapter, you've started your journey to become a regular expression master by learning the most useful stringr functions and the most important components of the regular expression language.
|
||
And there are plenty of resources to learn more.
|
||
|
||
A good place to start is `vignette("regular-expressions", package = "stringr")`: it documents the full set of syntax supported by stringr.
|
||
Another useful reference is [https://www.regular-expressions.info/](https://www.regular-expressions.info/tutorial.html).
|
||
It's not R specific, but you can use it to learn about the most advanced features of regexes and how they work under the hood.
|
||
|
||
It's also good to know that stringr is implemented on top of the stringi package by Marek Gagolewski.
|
||
If you're struggling to find a function that does what you need in stringr, don't be afraid to look in stringi.
|
||
You'll find stringi very easy to pick up because it follows many of the the same conventions as stringr.
|
||
|
||
In the next chapter, we'll talk about a data structure closely related to strings: factors.
|
||
Factors are used to represent categorical data in R, i.e. data with a fixed and known set of possible values identified by a vector of strings.
|