r4ds/relational-data.Rmd

644 lines
31 KiB
Plaintext
Raw Normal View History

2016-04-23 02:12:09 +08:00
# Relational data
2016-07-19 21:01:50 +08:00
## Introduction
It's rare that a data analysis involves only a single data frame.
Typically you have many data frames, and you must combine them to answer the questions that you're interested in.
Collectively, multiple data frames are called **relational data** because it is the relations, not just the individual datasets, that are important.
Relations are always defined between a pair of data frames.
All other relations are built up from this simple idea: the relations of three or more data frames are always a property of the relations between each pair.
Sometimes both elements of a pair can be the same data frame!
This is needed if, for example, you have a data frame of people, and each person has a reference to their parents.
2016-01-15 21:05:45 +08:00
To work with relational data you need verbs that work with pairs of data frames.
There are three families of verbs designed to work with relational data:
- **Mutating joins**, which add new variables to one data frame from matching observations in another.
- **Filtering joins**, which filter observations from one data frame based on whether or not they match an observation in the other data frame.
- **Set operations**, which treat observations as if they were set elements.
The most common place to find relational data is in a *relational* database management system (or RDBMS), a term that encompasses almost all modern databases.
If you've used a database before, you've almost certainly used SQL.
If so, you should find the concepts in this chapter familiar, although their expression in dplyr is a little different.
One other major terminology difference between databases and R is that what we generally refer to as data frames in R while the same concept is referred to as "table" in databases.
Hence you'll see references to one-table and two-table verbs in dplyr documentation.
Generally, dplyr is a little easier to use than SQL because dplyr is specialised to do data analysis: it makes common data analysis operations easier, at the expense of making it more difficult to do other things that aren't commonly needed for data analysis.
2016-07-19 21:01:50 +08:00
### Prerequisites
We will explore relational data from `nycflights13` using the two-table verbs from dplyr.
2016-07-27 05:31:26 +08:00
```{r setup, message = FALSE}
2016-10-04 01:30:24 +08:00
library(tidyverse)
2016-07-19 21:01:50 +08:00
library(nycflights13)
```
## nycflights13 {#nycflights13-relational}
We will use the nycflights13 package to learn about relational data.
nycflights13 contains five tibbles : `airlines`, `airports`, `weather` and `planes` which are all related to the `flights` data frame that you used in Chapter \@ref(data-transform) on data transformation:
- `airlines` lets you look up the full carrier name from its abbreviated code:
2016-01-30 21:13:46 +08:00
```{r}
airlines
```
- `airports` gives information about each airport, identified by the `faa` airport code:
2016-01-30 21:13:46 +08:00
```{r}
airports
```
2016-01-13 23:06:43 +08:00
- `planes` gives information about each plane, identified by its `tailnum`:
2016-01-13 23:06:43 +08:00
```{r}
planes
```
2016-01-30 21:13:46 +08:00
- `weather` gives the weather at each NYC airport for each hour:
```{r}
weather
```
One way to show the relationships between the different data frames is with a diagram:
```{r, echo = FALSE, fig.alt = "Diagram of the relationships between airports, planes, flights, weather, and airlines datasets from the nycflights13 package. The faa variable in the airports data frame is connected to the origin and dest variables in the flights data frame. The tailnum variable in the planes data frame is connected to the tailnum variable in flights. The year, month, day, hour, and origin variables are connected to the variables with the same name in the flights data frame. And finally the carrier variables in the airlines data frame is connected to the carrier variable in the flights data frame. There are no direct connections between airports, planes, airlines, and weather data frames."}
knitr::include_graphics("diagrams/relational-nycflights.png")
```
This diagram is a little overwhelming, but it's simple compared to some you'll see in the wild!
The key to understanding diagrams like this is to remember each relation always concerns a pair of data frames.
You don't need to understand the whole thing; you just need to understand the chain of relations between the data frames that you are interested in.
For nycflights13:
- `flights` connects to `planes` via a single variable, `tailnum`.
- `flights` connects to `airlines` through the `carrier` variable.
2016-07-27 05:31:26 +08:00
- `flights` connects to `airports` in two ways: via the `origin` and `dest` variables.
2016-01-30 21:13:46 +08:00
- `flights` connects to `weather` via `origin` (the location), and `year`, `month`, `day` and `hour` (the time).
### Exercises
1. Imagine you wanted to draw (approximately) the route each plane flies from its origin to its destination.
What variables would you need?
What data frames would you need to combine?
2. I forgot to draw the relationship between `weather` and `airports`.
2016-08-12 21:58:32 +08:00
What is the relationship and how should it appear in the diagram?
2016-01-30 21:13:46 +08:00
3. `weather` only contains information for the origin (NYC) airports.
If it contained weather records for all airports in the USA, what additional relation would it define with `flights`?
2016-01-13 23:06:43 +08:00
2016-01-15 21:05:45 +08:00
## Keys
The variables used to connect each pair of data frames are called **keys**.
A key is a variable (or set of variables) that uniquely identifies an observation.
In simple cases, a single variable is sufficient to identify an observation.
For example, each plane is uniquely identified by its `tailnum`.
In other cases, multiple variables may be needed.
For example, to identify an observation in `weather` you need five variables: `year`, `month`, `day`, `hour`, and `origin`.
2016-01-15 21:05:45 +08:00
There are two types of keys:
- A **primary key** uniquely identifies an observation in its own data frame.
For example, `planes$tailnum` is a primary key because it uniquely identifies each plane in the `planes` data frame.
2016-01-15 21:05:45 +08:00
- A **foreign key** uniquely identifies an observation in another data frame.
For example, `flights$tailnum` is a foreign key because it appears in the `flights` data frame where it matches each flight to a unique plane.
2016-01-15 21:05:45 +08:00
A variable can be both a primary key *and* a foreign key.
For example, `origin` is part of the `weather` primary key, and is also a foreign key for the `airports` data frame.
2016-01-15 21:05:45 +08:00
Once you've identified the primary keys in your data frames, it's good practice to verify that they do indeed uniquely identify each observation.
One way to do that is to `count()` the primary keys and look for entries where `n` is greater than one:
2016-01-15 21:05:45 +08:00
```{r}
2022-02-24 03:15:52 +08:00
planes |>
count(tailnum) |>
2016-08-01 00:32:16 +08:00
filter(n > 1)
2022-02-24 03:15:52 +08:00
weather |>
count(year, month, day, hour, origin) |>
2016-08-01 00:32:16 +08:00
filter(n > 1)
2016-01-15 21:05:45 +08:00
```
Sometimes a data frame doesn't have an explicit primary key: each row is an observation, but no combination of variables reliably identifies it.
For example, what's the primary key in the `flights` data frame?
You might think it would be the date plus the flight or tail number, but neither of those are unique:
2016-01-15 21:05:45 +08:00
```{r}
2022-02-24 03:15:52 +08:00
flights |>
count(year, month, day, flight) |>
2016-08-01 00:32:16 +08:00
filter(n > 1)
2022-02-24 03:15:52 +08:00
flights |>
count(year, month, day, tailnum) |>
2016-08-01 00:32:16 +08:00
filter(n > 1)
2016-01-15 21:05:45 +08:00
```
When starting to work with this data, I had naively assumed that each flight number would be only used once per day: that would make it much easier to communicate problems with a specific flight.
Unfortunately that is not the case!
If a data frame lacks a primary key, it's sometimes useful to add one with `mutate()` and `row_number()`.
That makes it easier to match observations if you've done some filtering and want to check back in with the original data.
This is called a **surrogate key**.
2016-01-15 21:05:45 +08:00
A primary key and the corresponding foreign key in another data frame form a **relation**.
Relations are typically one-to-many.
For example, each flight has one plane, but each plane has many flights.
In other data, you'll occasionally see a 1-to-1 relationship.
You can think of this as a special case of 1-to-many.
You can model many-to-many relations with a many-to-1 relation plus a 1-to-many relation.
For example, in this data there's a many-to-many relationship between airlines and airports: each airline flies to many airports; each airport hosts many airlines.
2016-01-15 21:05:45 +08:00
### Exercises
2016-08-12 21:58:32 +08:00
1. Add a surrogate key to `flights`.
2. We know that some days of the year are "special", and fewer people than usual fly on them.
How might you represent that data as a data frame?
What would be the primary keys of that data frame?
How would it connect to the existing data frames?
3. Identify the keys in the following datasets
2016-01-15 21:05:45 +08:00
a. `Lahman::Batting`
b. `babynames::babynames`
c. `nasaweather::atmos`
d. `fueleconomy::vehicles`
e. `ggplot2::diamonds`
2016-07-27 05:31:26 +08:00
(You might need to install some packages and read some documentation.)
2016-01-30 21:13:46 +08:00
4. Draw a diagram illustrating the connections between the `Batting`, `People`, and `Salaries` data frames in the Lahman package.
Draw another diagram that shows the relationship between `People`, `Managers`, `AwardsManagers`.
2016-01-30 21:13:46 +08:00
How would you characterise the relationship between the `Batting`, `Pitching`, and `Fielding` data frames?
2016-01-15 21:05:45 +08:00
## Mutating joins {#mutating-joins}
The first tool we'll look at for combining a pair of data frames is the **mutating join**.
A mutating join allows you to combine variables from two data frames.
It first matches observations by their keys, then copies across variables from one data frame to the other.
Like `mutate()`, the join functions add variables to the right, so if you have a lot of variables already, the new variables won't get printed out.
For these examples, we'll make it easier to see what's going on in the examples by creating a narrower dataset:
```{r}
2022-02-24 03:15:52 +08:00
flights2 <- flights |>
2016-08-01 00:32:16 +08:00
select(year:day, hour, origin, dest, tailnum, carrier)
flights2
```
2016-08-15 20:33:36 +08:00
(Remember, when you're in RStudio, you can also use `View()` to avoid this problem.)
Imagine you want to add the full airline name to the `flights2` data.
You can combine the `airlines` and `flights2` data frames with `left_join()`:
```{r}
2022-02-24 03:15:52 +08:00
flights2 |>
select(-origin, -dest) |>
left_join(airlines, by = "carrier")
```
The result of joining airlines to flights2 is an additional variable: `name`.
This is why I call this type of join a mutating join.
In this case, you could have got to the same place using `mutate()` and R's base subsetting:
```{r}
2022-02-24 03:15:52 +08:00
flights2 |>
select(-origin, -dest) |>
2016-08-12 21:58:32 +08:00
mutate(name = airlines$name[match(carrier, airlines$carrier)])
```
2016-01-13 23:06:43 +08:00
But this is hard to generalise when you need to match multiple variables, and takes close reading to figure out the overall intent.
The following sections explain, in detail, how mutating joins work.
You'll start by learning a useful visual representation of joins.
We'll then use that to explain the four mutating join functions: the inner join, and the three outer joins.
When working with real data, keys don't always uniquely identify observations, so next we'll talk about what happens when there isn't a unique match.
Finally, you'll learn how to tell dplyr which variables are the keys for a given join.
### Understanding joins
2016-07-27 05:31:26 +08:00
To help you learn how joins work, I'm going to use a visual representation:
```{r, echo = FALSE, out.width = NULL, fig.alt = "x and y are two data frames with 2 columns and 3 rows each. The first column in each is the key and the second is the value. The contents of these data frames are given in the subsequent code chunk."}
knitr::include_graphics("diagrams/join-setup.png")
```
```{r}
2016-10-07 21:16:09 +08:00
x <- tribble(
~key, ~val_x,
1, "x1",
2, "x2",
3, "x3"
)
y <- tribble(
~key, ~val_y,
1, "y1",
2, "y2",
4, "y3"
)
```
The coloured column represents the "key" variable: these are used to match the rows between the data frames.
The grey column represents the "value" column that is carried along for the ride.
In these examples I'll show a single key variable, but the idea generalises in a straightforward way to multiple keys and multiple values.
A join is a way of connecting each row in `x` to zero, one, or more rows in `y`.
The following diagram shows each potential match as an intersection of a pair of lines.
```{r, echo = FALSE, out.width = NULL, fig.alt = "x and y data frames placed next to each other. with the key variable moved up front in y so that the key variable in x and key variable in y appear next to each other."}
knitr::include_graphics("diagrams/join-setup2.png")
```
2016-07-27 05:31:26 +08:00
(If you look closely, you might notice that we've switched the order of the key and value columns in `x`. This is to emphasise that joins match based on the key; the value is just carried along for the ride.)
In an actual join, matches will be indicated with dots.
The number of dots = the number of matches = the number of rows in the output.
2016-01-14 23:51:33 +08:00
```{r join-inner, echo = FALSE, out.width = NULL, fig.alt = "Keys 1 and 2 in x and y data frames are matched and indicated with lines joining these rows with dot in the middle. Hence, there are two dots in this diagram. The resulting joined data frame has two rows and 3 columns: key, val_x, and val_y. Values in the key column are 1 and 2 (the matched values)."}
2016-01-14 23:51:33 +08:00
knitr::include_graphics("diagrams/join-inner.png")
```
### Inner join {#inner-join}
The simplest type of join is the **inner join**.
An inner join matches pairs of observations whenever their keys are equal:
```{r, echo = FALSE, ref.label = "join-inner", opts.label = TRUE}
knitr::include_graphics("diagrams/join-inner.png")
```
(To be precise, this is an inner **equijoin** because the keys are matched using the equality operator. Since most joins are equijoins we usually drop that specification.)
The output of an inner join is a new data frame that contains the key, the x values, and the y values.
We use `by` to tell dplyr which variable is the key:
```{r}
2022-02-24 03:15:52 +08:00
x |>
2016-08-01 00:32:16 +08:00
inner_join(y, by = "key")
```
The most important property of an inner join is that unmatched rows are not included in the result.
This means that generally inner joins are usually not appropriate for use in analysis because it's too easy to lose observations.
### Outer joins {#outer-join}
An inner join keeps observations that appear in both data frames.
An **outer join** keeps observations that appear in at least one of the data frames.
There are three types of outer joins:
- A **left join** keeps all observations in `x`.
- A **right join** keeps all observations in `y`.
- A **full join** keeps all observations in `x` and `y`.
These joins work by adding an additional "virtual" observation to each data frame.
This observation has a key that always matches (if no other key matches), and a value filled with `NA`.
2016-01-13 23:06:43 +08:00
Graphically, that looks like:
```{r, echo = FALSE, out.width = NULL, fig.alt = "Three diagrams for left, right, and full joins. In each diagram data frame x is on the left and y is on the right. The result of the join is always a data frame with three columns (key, val_x, and val_y). Left join: keys 1 and 2 from x are matched to those in y, key 3 is also carried along to the joined result since it's on the left data frame, but key 4 from y is not carried along since it's on the right but not on the left. The result is a data frame with 3 rows: keys 1, 2, and 3, all values from val_x, and the corresponding values from val_y for keys 1 and 2 with an NA for key 3, val_y. Right join: keys 1 and 2 from x are matched to those in y, key 4 is also carried along to the joined result since it's on the right data frame, but key 3 from x is not carried along since it's on the left but not on the right. The result is a data frame with 3 rows: keys 1, 2, and 4, all values from val_y, and the corresponding values from val_x for keys 1 and 2 with an NA for key 4, val_x. Full join: The resulting data frame has 4 rows: keys 1, 2, 3, and 4 with all values from val_x and val_y, however key 2, val_y and key 4, val_x are NAs since those keys aren't present in their respective data frames."}
2016-01-13 23:06:43 +08:00
knitr::include_graphics("diagrams/join-outer.png")
```
The most commonly used join is the left join: you use this whenever you look up additional data from another data frame, because it preserves the original observations even when there isn't a match.
The left join should be your default join: use it unless you have a strong reason to prefer one of the others.
2016-01-13 23:06:43 +08:00
Another way to depict the different types of joins is with a Venn diagram:
```{r, echo = FALSE, out.width = NULL, fig.alt = "Venn diagrams for inner, full, left, and right joins. Each join represented with two intersecting circles representing data frames x and y, with x on the right and y on the left. Shading indicates the result of the join. Inner join: Only intersection is shaded. Full join: Everything is shaded. Left join: Only x is shaded, but not the area in y that doesn't intersect with x. Right join: Only y is shaded, but not the area in x that doesn't intersect with y."}
knitr::include_graphics("diagrams/join-venn.png")
```
However, this is not a great representation.
It might jog your memory about which join preserves the observations in which data frame, but it suffers from a major limitation: a Venn diagram can't show what happens when keys don't uniquely identify an observation.
### Duplicate keys {#join-matches}
So far all the diagrams have assumed that the keys are unique.
But that's not always the case.
This section explains what happens when the keys are not unique.
There are two possibilities:
1. One data frame has duplicate keys.
This is useful when you want to add in additional information as there is typically a one-to-many relationship.
2016-01-30 21:13:46 +08:00
```{r, echo = FALSE, out.width = NULL, fig.alt = "Diagram describing a left join where one of the data frames (x) has duplicate keys. Data frame x is on the left, has 4 rows and 2 columns (key, val_x), and has the keys 1, 2, 2, and 1. Data frame y is on the right, has 2 rows and 2 columns (key, val_y), and has the keys 1 and 2. Left joining these two data frames yields a data frame with 4 rows (keys 1, 2, 2, and 1) and 3 columns (val_x, key, val_y). All values from x$val_x are carried along, values in y for key 1 and 2 are duplicated."}
knitr::include_graphics("diagrams/join-one-to-many.png")
```
Note that I've put the key column in a slightly different position in the output.
This reflects that the key is a primary key in `y` and a foreign key in `x`.
```{r}
2016-10-07 21:16:09 +08:00
x <- tribble(
~key, ~val_x,
1, "x1",
2, "x2",
2, "x3",
1, "x4"
)
y <- tribble(
~key, ~val_y,
1, "y1",
2, "y2"
)
left_join(x, y, by = "key")
```
2. Both data frames have duplicate keys.
This is usually an error because in neither data frame do the keys uniquely identify an observation.
When you join duplicated keys, you get all possible combinations, the Cartesian product:
2016-01-30 21:13:46 +08:00
```{r, echo = FALSE, out.width = NULL, fig.alt = "Diagram describing a left join where both data frames (x and y) have duplicate keys. Data frame x is on the left, has 4 rows and 2 columns (key, val_x), and has the keys 1, 2, 2, and 3. Data frame y is on the right, has 4 rows and 2 columns (key, val_y), and has the keys 1, 2, 2, and 3 as well. Left joining these two data frames yields a data frame with 6 rows (keys 1, 2, 2, 2, 2, and 3) and 3 columns (key, val_x, val_y). All values from both datasets are included."}
knitr::include_graphics("diagrams/join-many-to-many.png")
```
2016-01-30 21:13:46 +08:00
```{r}
2016-10-07 21:16:09 +08:00
x <- tribble(
~key, ~val_x,
1, "x1",
2, "x2",
2, "x3",
3, "x4"
)
y <- tribble(
~key, ~val_y,
1, "y1",
2, "y2",
2, "y3",
3, "y4"
)
left_join(x, y, by = "key")
```
### Defining the key columns {#join-by}
So far, the pairs of data frames have always been joined by a single variable, and that variable has the same name in both data frames.
That constraint was encoded by `by = "key"`.
You can use other values for `by` to connect the data frames in other ways:
- The default, `by = NULL`, uses all variables that appear in both data frames, the so called **natural** join.
For example, the flights and weather data frames match on their common variables: `year`, `month`, `day`, `hour` and `origin`.
2016-01-30 21:13:46 +08:00
```{r}
2022-02-24 03:15:52 +08:00
flights2 |>
2016-08-01 00:32:16 +08:00
left_join(weather)
```
2016-01-30 21:13:46 +08:00
- A character vector, `by = "x"`.
This is like a natural join, but uses only some of the common variables.
For example, `flights` and `planes` have `year` variables, but they mean different things so we only want to join by `tailnum`.
2016-01-30 21:13:46 +08:00
```{r}
2022-02-24 03:15:52 +08:00
flights2 |>
2016-08-01 00:32:16 +08:00
left_join(planes, by = "tailnum")
```
2016-01-30 21:13:46 +08:00
Note that the `year` variables (which appear in both input data frames, but are not constrained to be equal) are disambiguated in the output with a suffix.
- A named character vector: `by = c("a" = "b")`.
This will match variable `a` in data frame `x` to variable `b` in data frame `y`.
The variables from `x` will be used in the output.
2016-01-30 21:13:46 +08:00
For example, if we want to draw a map we need to combine the flights data with the airports data which contains the location (`lat` and `lon`) of each airport.
Each flight has an origin and destination `airport`, so we need to specify which one we want to join to:
2016-01-30 21:13:46 +08:00
```{r}
2022-02-24 03:15:52 +08:00
flights2 |>
2016-08-01 00:32:16 +08:00
left_join(airports, c("dest" = "faa"))
2022-02-24 03:15:52 +08:00
flights2 |>
2016-08-01 00:32:16 +08:00
left_join(airports, c("origin" = "faa"))
```
### Exercises
1. Compute the average delay by destination, then join on the `airports` data frame so you can show the spatial distribution of delays.
Here's an easy way to draw a map of the United States:
2016-01-30 21:13:46 +08:00
2016-07-27 05:31:26 +08:00
```{r, eval = FALSE}
2022-02-24 03:15:52 +08:00
airports |>
semi_join(flights, c("faa" = "dest")) |>
2016-01-30 21:13:46 +08:00
ggplot(aes(lon, lat)) +
borders("state") +
geom_point() +
coord_quickmap()
```
2016-01-30 21:13:46 +08:00
(Don't worry if you don't understand what `semi_join()` does --- you'll learn about it next.)
2016-08-12 21:58:32 +08:00
You might want to use the `size` or `colour` of the points to display the average delay for each airport.
2. Add the location of the origin *and* destination (i.e. the `lat` and `lon`) to `flights`.
2016-08-12 21:58:32 +08:00
3. Is there a relationship between the age of a plane and its delays?
4. What weather conditions make it more likely to see a delay?
2016-01-30 21:13:46 +08:00
5. What happened on June 13 2013?
Display the spatial pattern of delays, and then use Google to cross-reference with the weather.
2016-01-30 21:13:46 +08:00
```{r, eval = FALSE, include = FALSE}
2016-07-27 05:31:26 +08:00
worst <- filter(flights, !is.na(dep_time), month == 6, day == 13)
2022-02-24 03:15:52 +08:00
worst |>
group_by(dest) |>
summarise(delay = mean(arr_delay), n = n()) |>
filter(n > 5) |>
inner_join(airports, by = c("dest" = "faa")) |>
ggplot(aes(lon, lat)) +
borders("state") +
geom_point(aes(size = n, colour = delay)) +
coord_quickmap()
```
2016-01-13 23:06:43 +08:00
### Other implementations
2016-01-30 21:13:46 +08:00
`base::merge()` can perform all four types of mutating join:
| dplyr | merge |
|--------------------|-------------------------------------------|
| `inner_join(x, y)` | `merge(x, y)` |
| `left_join(x, y)` | `merge(x, y, all.x = TRUE)` |
| `right_join(x, y)` | `merge(x, y, all.y = TRUE)`, |
| `full_join(x, y)` | `merge(x, y, all.x = TRUE, all.y = TRUE)` |
The advantages of the specific dplyr verbs is that they more clearly convey the intent of your code: the difference between the joins is really important but concealed in the arguments of `merge()`.
dplyr's joins are considerably faster and don't mess with the order of the rows.
SQL is the inspiration for dplyr's conventions, so the translation is straightforward:
| dplyr | SQL |
|------------------------------|------------------------------------------------|
| `inner_join(x, y, by = "z")` | `SELECT * FROM x INNER JOIN y USING (z)` |
| `left_join(x, y, by = "z")` | `SELECT * FROM x LEFT OUTER JOIN y USING (z)` |
| `right_join(x, y, by = "z")` | `SELECT * FROM x RIGHT OUTER JOIN y USING (z)` |
| `full_join(x, y, by = "z")` | `SELECT * FROM x FULL OUTER JOIN y USING (z)` |
2016-01-30 21:13:46 +08:00
Note that "INNER" and "OUTER" are optional, and often omitted.
Joining different variables between the data frames, e.g. `inner_join(x, y, by = c("a" = "b"))` uses a slightly different syntax in SQL: `SELECT * FROM x INNER JOIN y ON x.a = y.b`.
As this syntax suggests, SQL supports a wider range of join types than dplyr because you can connect the data frames using constraints other than equality (sometimes called non-equijoins).
## Filtering joins {#filtering-joins}
Filtering joins match observations in the same way as mutating joins, but affect the observations, not the variables.
There are two types:
- `semi_join(x, y)` **keeps** all observations in `x` that have a match in `y`.
- `anti_join(x, y)` **drops** all observations in `x` that have a match in `y`.
Semi-joins are useful for matching filtered summary data frames back to the original rows.
For example, imagine you've found the top ten most popular destinations:
```{r}
2022-02-24 03:15:52 +08:00
top_dest <- flights |>
count(dest, sort = TRUE) |>
head(10)
top_dest
```
Now you want to find each flight that went to one of those destinations.
You could construct a filter yourself:
```{r}
2022-02-24 03:15:52 +08:00
flights |>
2016-08-01 00:32:16 +08:00
filter(dest %in% top_dest$dest)
```
But it's difficult to extend that approach to multiple variables.
For example, imagine that you'd found the 10 days with highest average delays.
How would you construct the filter statement that used `year`, `month`, and `day` to match it back to `flights`?
Instead you can use a semi-join, which connects the two data frames like a mutating join, but instead of adding new columns, only keeps the rows in `x` that have a match in `y`:
```{r}
2022-02-24 03:15:52 +08:00
flights |>
2016-08-01 00:32:16 +08:00
semi_join(top_dest)
```
Graphically, a semi-join looks like this:
```{r, echo = FALSE, out.width = NULL, fig.alt = "Diagram of a semi join. Data frame x is on the left and has two columns (key and val_x) with keys 1, 2, and 3. Diagram y is on the right and also has two columns (key and val_y) with keys 1, 2, and 4. Semi joining these two results in a data frame with two rows and two columns (key and val_x), with keys 1 and 2 (the only keys that match between the two data frames)."}
knitr::include_graphics("diagrams/join-semi.png")
```
Only the existence of a match is important; it doesn't matter which observation is matched.
This means that filtering joins never duplicate rows like mutating joins do:
```{r, echo = FALSE, out.width = NULL, fig.alt = "Diagram of a semi join with data frames with duplicated keys. Data frame x is on the left and has two columns (key and val_x) with keys 1, 2, 2, and 3. Diagram y is on the right and also has two columns (key and val_y) with keys 1, 2, 2, and 3 as well. Semi joining these two results in a data frame with four rows and two columns (key and val_x), with keys 1, 2, 2, and 3 (the matching keys, each appearing as many times as they do in x)."}
knitr::include_graphics("diagrams/join-semi-many.png")
```
The inverse of a semi-join is an anti-join.
An anti-join keeps the rows that *don't* have a match:
```{r, echo = FALSE, out.width = NULL, fig.alt = "Diagram of an anti join. Data frame x is on the left and has two columns (key and val_x) with keys 1, 2, and 3. Diagram y is on the right and also has two columns (key and val_y) with keys 1, 2, and 4. Anti joining these two results in a data frame with one row and two columns (key and val_x), with keys 3 only (the only key in x that is not in y)."}
knitr::include_graphics("diagrams/join-anti.png")
```
Anti-joins are useful for diagnosing join mismatches.
For example, when connecting `flights` and `planes`, you might be interested to know that there are many `flights` that don't have a match in `planes`:
```{r}
2022-02-24 03:15:52 +08:00
flights |>
anti_join(planes, by = "tailnum") |>
count(tailnum, sort = TRUE)
```
### Exercises
1. What does it mean for a flight to have a missing `tailnum`?
What do the tail numbers that don't have a matching record in `planes` have in common?
(Hint: one variable explains \~90% of the problems.)
2. Filter flights to only show flights with planes that have flown at least 100 flights.
2016-08-12 21:58:32 +08:00
3. Combine `fueleconomy::vehicles` and `fueleconomy::common` to find only the records for the most common models.
2016-08-12 21:58:32 +08:00
4. Find the 48 hours (over the course of the whole year) that have the worst delays.
Cross-reference it with the `weather` data.
Can you see any patterns?
2016-01-30 21:13:46 +08:00
5. What does `anti_join(flights, airports, by = c("dest" = "faa"))` tell you?
2016-03-25 01:25:56 +08:00
What does `anti_join(airports, flights, by = c("faa" = "dest"))` tell you?
6. You might expect that there's an implicit relationship between plane and airline, because each plane is flown by a single airline.
Confirm or reject this hypothesis using the tools you've learned above.
2016-10-07 21:16:20 +08:00
2016-01-18 22:21:22 +08:00
## Join problems
The data you've been working with in this chapter has been cleaned up so that you'll have as few problems as possible.
Your own data is unlikely to be so nice, so there are a few things that you should do with your own data to make your joins go smoothly.
2016-01-18 22:21:22 +08:00
1. Start by identifying the variables that form the primary key in each data frame.
You should usually do this based on your understanding of the data, not empirically by looking for a combination of variables that give a unique identifier.
If you just look for variables without thinking about what they mean, you might get (un)lucky and find a combination that's unique in your current data but the relationship might not be true in general.
2016-01-30 21:13:46 +08:00
For example, the altitude and longitude uniquely identify each airport, but they are not good identifiers!
2016-07-27 05:31:26 +08:00
2016-01-18 22:21:22 +08:00
```{r}
2022-02-24 03:15:52 +08:00
airports |> count(alt, lon) |> filter(n > 1)
2016-01-18 22:21:22 +08:00
```
2. Check that none of the variables in the primary key are missing.
If a value is missing then it can't identify an observation!
2016-01-30 21:13:46 +08:00
3. Check that your foreign keys match primary keys in another data frame.
The best way to do this is with an `anti_join()`.
It's common for keys not to match because of data entry errors.
Fixing these is often a lot of work.
2016-01-30 21:13:46 +08:00
If you do have missing keys, you'll need to be thoughtful about your use of inner vs. outer joins, carefully considering whether or not you want to drop rows that don't have a match.
2016-01-18 22:21:22 +08:00
Be aware that simply checking the number of rows before and after the join is not sufficient to ensure that your join has gone smoothly.
If you have an inner join with duplicate keys in both data frames, you might get unlucky as the number of dropped rows might exactly equal the number of duplicated rows!
2016-01-18 22:21:22 +08:00
## Set operations {#set-operations}
The final type of two-table verb are the set operations.
Generally, I use these the least frequently, but they are occasionally useful when you want to break a single complex filter into simpler pieces.
All these operations work with a complete row, comparing the values of every variable.
These expect the `x` and `y` inputs to have the same variables, and treat the observations like sets:
- `intersect(x, y)`: return only observations in both `x` and `y`.
- `union(x, y)`: return unique observations in `x` and `y`.
- `setdiff(x, y)`: return observations in `x`, but not in `y`.
Given this simple data:
```{r}
2016-10-07 21:16:09 +08:00
df1 <- tribble(
~x, ~y,
1, 1,
2, 1
)
df2 <- tribble(
~x, ~y,
1, 1,
1, 2
)
```
The four possibilities are:
```{r}
intersect(df1, df2)
2016-08-12 21:58:32 +08:00
# Note that we get 3 rows, not 4
union(df1, df2)
2016-08-12 21:58:32 +08:00
setdiff(df1, df2)
2016-08-12 21:58:32 +08:00
setdiff(df2, df1)
```