The frequency of early fatality and the transient nature of emergency medical care mean that a single database will rarely suffice for population based injury research. Linking records from multiple data sources is therefore a promising method for injury surveillance or trauma system evaluation. The purpose of this article is to review the historical development of record linkage, provide a basic mathematical foundation, discuss some practical issues, and consider some ethical concerns.
Clerical or computer assisted deterministic record linkage methods may suffice for some applications, but probabilistic methods are particularly useful for larger studies. The probabilistic method attempts to simulate human reasoning by comparing each of several elements from the two records. The basic mathematical specifications are derived algebraically from fundamental concepts of probability, although the theory can be extended to include more advanced mathematics.
Probabilistic, deterministic, and clerical techniques may be combined in different ways depending upon the goal of the record linkage project. If a population parameter is being estimated for a purely statistical study, a completely probabilistic approach may be most efficient; for other applications, where the purpose is to make inferences about specific individuals based upon their data contained in two or more files, the need for a high positive predictive value would favor a deterministic method or a probabilistic method with careful clerical review. Whatever techniques are used, researchers must realize that the combination of data sources entails additional ethical obligations beyond the use of each source alone.
Statistics from Altmetric.com
- CODES, Crash Outcome Data Evaluation System
- EMS, Emergency Medical Services
- NPV, negative predictive value
- PPV, positive predictive value
The frequency of early fatality and transient nature of trauma care mean that a single database will rarely suffice for population based injury research. Emergency Medical Services (EMS) and vital statistics data have been combined to determine outcomes of cardiac arrest,1 and a similar approach is warranted for victims of severe injuries, who often die without entering an EMS system or require transfer from one hospital to another. Record linkage methods have therefore been advocated for studies of injury outcomes.2–4
For small applications, enough information is usually present to allow an accurate human judgment about whether a record from one source refers to the same case as a record from another source. However, this “manual” or “clerical” method becomes impractical with large numbers. A natural solution is to use a computer for “matching” or “linking” records; for simplicity, these terms will be used interchangeably, although some have reserved the former for the true relationship and the latter for the decision to accept that two records from different sources refer to the same case.5
The easiest computer assisted method is to link cases that have the same identification number, or some other element or group of elements that uniquely identify a given person or episode. This approach may be referred to as deterministic (or “exact” or “all-or-none”) matching, and is effective in many cases. However, the necessary information may be absent, may have different formats or variations in different sources, or may be inaccurately entered or missing. Most of the interest in large scale record linkage research has therefore focused on probabilistic methods that simulate human pattern recognition when deciding that a record from one source refers to the same person or event as a record from another source. Despite the sophistication of some computer methods, the only “gold standard” for whether two records truly match is still the judgment of a human reviewer, and a combination of deterministic and probabilistic computer methods, along with human judgment, will often be the best approach.6,7
Much information about record linkage is available, and there have been previous reviews of the subject,8–10 but references are in diverse locations mostly irrelevant to the field of injury control. The purpose of this article is to review the historical development of record linkage, provide a basic mathematical foundation, discuss some practical issues, and consider some ethical concerns arising from linking multiple databases. This is not an exhaustive review, but an outline of the main principles. More detailed information is available in the references, including proceedings of the United States Federal Committee on Statistical Methodology workshops from 1985 and 1997, which contain reprints of some classic articles.11,12
The potential benefits of linking medical and vital statistics records were recognized even before computers became widely available.13 By 1959, Newcombe and colleagues in Canada reported the ability to link such records contained on punch cards at a rate of about 10 per minute, and hoped that technology would increase this rate by a factor of at least 20.14 Twenty years later, Newcombe was able to demonstrate the superiority of his computer methods over clerical methods for a large record linkage project, and the processing rate had increased to about 14 000 records per minute.15 This processing speed has now also been vastly exceeded, along with further improvements in programming and data storage, and reductions in the size and cost of computer hardware.
Increasingly practical applications have been developed, largely based on Newcombe’s methodology (see table 1).6,12,14,16–30 Advances in probabilistic record linkage methodology have begun to see applications in injury epidemiology, which had previously been limited to cumbersome manual methods31–34 or ad hoc deterministic procedures.35–38 A “road injury database” was constructed in Western Australia by linking medical, police, and traffic data using the Canadian Generalized Iterative Record Linkage System39; later the methods were adapted to use Automatch and hospital trauma registries were added.40,41 Hospital trauma registries in Maine were linked to death certificates, hospital discharge abstracts, and EMS data initially using a deterministic computer program,37 but when Automatch became available, the latter was found easier to specify and generalize.42 Hospital readmissions for injury in New Zealand were more easily identified by probabilistic methods, which allowed for more variables to be used for linkage than a deterministic method, even when some values were missing or erroneous.43
The Crash Outcome Data Evaluation System (CODES) project has been carried out in the past decade under the direction of the United States National Highway Traffic Safety Administration. This project has used probabilistic methods to link crash data with EMS, hospitalization, and death certificate data in several states. Many of the results from this project are available only as government documents,28,44,45 although limited results from some states are accessible in the medical literature.46–51 Despite some criticism,52 CODES has produced a major increase in the experience and understanding of record linkage methods within the injury control community. Building on this experience, probabilistic linkage of other injury data has been successfully accomplished in Maine53 and Utah.54 The latest CODES projects have used new software, with an easier user interface.
PRACTICAL AND THEORETICAL ISSUES
Although the mathematics and computer matching procedures are very interesting (see Appendix), the most difficult and time consuming part of a record linkage project is the preprocessing.14,20,28 Missing or miscoded data, duplicate records, etc must be dealt with, and files must be put into standard formats for dates, locations, etc. Indeed, the success of record linkage is much more dependent on data quality than on software.
Special problems arise if names are available for linkage.16 Although this may allow greater accuracy, the relative frequency or infrequency of different names, changes due to marriage, potential variations in spelling, nicknames, abbreviations, etc, greatly increase the complexity of matching. Numerous clever approaches can be programmed,55 but human pattern recognition is particularly hard to replicate in this area.5 In practice, confidentiality restrictions usually do not allow the use of names in large medical databases.
With the probabilistic approach, the number of possible comparisons increases with the product of the file sizes, which becomes impractical when the files are large. The usual remedy is to stratify the procedure by restricting the comparisons to “blocks” or “pockets” of cases where one or more variables match exactly. This essentially utilizes a deterministic approach to assist the probabilistic method, but can be further modified by “blocking” sequentially using different variables.
In epidemiologic studies using record linkage, the probability of falsely matching records that should not have been matched must be balanced against the probability of failing to match records that should have been matched. Records that are falsely matched (“mismatches” or “homonym errors”) will lead to misidentification of the outcome for specific cases as well as underestimation of the total number of cases; records that are falsely unmatched (“false non-matches”, “erroneous non-matches”, “failures to match”, or “synonym errors”) will lead to missing data from one or the other source and overestimation of the total number of cases. The theoretical magnitude of these errors can be estimated algebraically after certain assumptions.56
The frequency of false positives and false negatives can be expressed in familiar terms of sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV),57 as depicted in table 2. In practice, the number of records truly unmatched is generally so large that specificity and NPV are not useful measurements. Furthermore, for any real application, it may be difficult to specify the “gold standard” against which matched or unmatched records are considered “true” or “false”; because of this, a method to estimate the PPV based on the frequency of duplicate links has been proposed.58
The basic mathematical concepts are described in the appendix; more advanced mathematical implications of automated record linkage have attracted the interest of some famous statisticians over the past half century,59–63 and the status of current research in this area has been summarized by Winkler.63 Fellegi and Sunter presented the formal theoretical structure for record linkage most often cited today,64 and showed that the approach based on likelihood ratios (developed empirically by Newcombe14) was in accordance with classical hypothesis testing theory. Newcombe, in one of his last publications,65 acknowledged that this approach can also be derived from Bayes’ Theorem as in the Appendix.
The mathematical approach to record linkage theory becomes more complicated when allowing for blocking or missing data.61,64 Other theoretical complications result if one allows partial credit for “near matches”.16 For very large samples, sophisticated mathematical research has gone into the problem of minimizing the need for human review by estimating error using models based upon past human experience.62
A certain degree of “art”6,66 or “fiddling around”67 with the linkages will be necessary despite mathematical and technological advances. As mentioned above, the “blocking” strategy essentially combines deterministic and probabilistic approaches, and human review of preliminary results is certainly part of the validation of any computer program.
The best method for a given linkage project depends in part on its purpose.68 If a population parameter is being estimated for a purely statistical study (for example, the effect of wearing safety belts on mortality), a completely probabilistic approach may be most efficient. To some extent, the numbers of records falsely matched and records falsely unmatched will cancel each other as match cut offs are varied.56,57,69 The sensitivity and PPV can be estimated and used to develop confidence limits on the parameter estimate.62
For other applications, where the purpose is to make inferences about specific individuals based upon their data contained in two or more files (for example, flow of patients through multiple phases of care in a trauma system), a completely probabilistic approach would not be likely to give acceptable results. In this case, we must be quite sure that records from different sources truly refer to the same person (high PPV), and might favor a deterministic method.70 However, probabilistic methods with careful clerical review may also be useful.41
A few studies have compared deterministic and probabilistic methods, using human review or artificially withheld identifying information as a “gold standard”. Roos et al23 and Jamieson et al7 both found that a probabilistic method identified more matches; however the latter study found that only their deterministic method was free from falsely matched records and suggested that a combination of methods might be valuable. Gomatam et al have compared Automatch and a “stepwise deterministic strategy” using two files for which the true relationships were known from other data71; they also found that the sensitivity of the probabilistic method was better, but the PPV for the deterministic method was nearly 100%.
ETHICS OF RECORD LINKAGE
In 1946, the Chief of the United States Public Health Service’s Office of Vital Statistics proposed that hospital, insurance, and other records for an individual be linked to provide statistical information for research.13 Noting that registration systems developed in Europe under police authority “will find disfavor in the United States”, he admired the decentralized Canadian system in which vital records were kept “in their proper place, i.e., under the control of public health and statistical agencies”, but linked to a federal index with a personal identification number. As predicted, American concerns for privacy have led to a more cautious approach to record linkage than in Canada.72,73
Ethical issues were not on the program of a symposium on record linkage techniques held in 1985, but the editors recognized that this was an important area for further research.11 Privacy issues were prominently addressed at a subsequent symposium held in 1997,12 where some of the leading theoreticians carefully analyzed the social implications of their scientific work.74,75 Citizens in the United States have a healthy mistrust of government, especially the huge federal bureaucracy. While there is broad support for the use of statistical information in public health research, this support depends upon the trust of the public that information accumulated for the general good will not be used against individual citizens.68
While the risk to patients may seem small, linkage of one database to another does create not only new generalizable knowledge about cause-and-effect relationships but also more specific knowledge about some individuals. Even if permission has been obtained to use separate databases, combining them adds a new level of obligation to the researcher and should only be done with the approval of the owners of the original data sets and an institutional review board; this does not necessarily mean that informed consent has to be obtained from each person whose records may be included (which would generally be impractical), but an impartial evaluation should show that the research is of good quality, that the risks are minimal, and that confidentiality of individual information will be maintained.10,70,76
In the United States, privacy considerations are even more important since the Health Insurance Portability and Accountability Act took effect in 2003; these regulations specifically prohibit the use of names, social security numbers, or vehicle identification numbers, and mandate informed consent for research using medical records unless waived by an institutional review board. The effect of this new legislation on clinical research is still being debated,77,78 although it should be noted that special provisions are made for public health authorities, including “an individual or entity acting under a grant of authority from or contract with such public agency”.79
APPENDIX: MATHEMATICAL BACKGROUND
PROBABILITY AND ODDS
Let us define the probability of A, signified by P(A), to mean your degree of belief that A is true, expressed as a fraction ranging from slightly more than 0 (impossible) to slightly less than 1 (certain). We can define:
where P(Ā) means the probability that A is not true. Note that when P(A) is very small, there is not much difference between the probability and the odds. Also, equation 1 can be rewritten as:
JOINT AND CONDITIONAL PROBABILITY; INDEPENDENCE
Let us define the joint probability of A and B to be the probability that both A and B are true, written symbolically as P(A,B). Let us also define the conditional probability that A is true, given that B is true, written symbolically as:
A and B may be defined as independent if:
BAYES’ THEOREM; WEIGHTS
From the definition of conditional probability, we get:
Further algebra gives us:
If we assume A1|B…An|B are independent, then with repeated applications of Bayes’ Theorem we get:
Now, consider P(B) to mean “the probability that two records on different lists refer to the same person” and A1 (for example) meaning “element 1 (age, sex, or whatever) is the same on both lists”. Record linkage terminology refers to P(A1|B) as an M probability (the probability that element 1 is the same if the records truly match), and refers to P(A1|B̄) as a U probability (the probability that element 1 is the same, just by chance, when the records truly should be unmatched). If a given element is not the same on both lists, the likelihood ratio becomes (1−M)/(1−U).
Newcombe introduced logarithms in his explanation of record linkage methods, but later was concerned that they might be more confusing than helpful.8 If we take the logarithm of both sides of equation 4, we obtain:
In other words, the posterior log odds (or overall weight) that the two records refer to the same person equals some constant (the prior log odds) plus the sum of the log likelihood ratios (agreement or disagreement weights) for each element.
ESTIMATING POSTERIOR PROBABILITIES
If we can demonstrate that our linking variables are nearly independent, then equation 4 will be approximately valid. If you have reason to believe (from other knowledge) that the number of matching records is about NX, the number of records in file A is NA, and the number of records in file B is NB, then you can estimate the prior probability that a randomly selected record from file A matches with a randomly selected record from file B as:
This will generally be a very small number, so the prior odds will be similar. If you choose to work with logarithms, the log odds will be a very negative number, to which the agreement weights (minus the disagreement weights) will be added to obtain the posterior log odds (equation 5). With or without logarithms, by reversing our previous transformations (equation 4 and equation 1) you can obtain a posterior probability (or absolute probability) that two records match.
This approach can also evaluate the feasibility of a proposed record linkage project.8,81,84 If the file sizes are known, and the number of expected links between them can be estimated, and the M and U probabilities can be approximated as described earlier, then equations 6, 4, and 1 can be used to see whether two truly matching records will be assigned a very high (for example, 95% or 99%) posterior probability of being correct.84 If not, the project may be impractical.
Suppose you have the data presented in table 3, and need to decide which ambulance cases correspond to which emergency department cases. For this small number, you could match them using your own inspection and judgment (based on past experience with these kinds of patients and records), but let us employ the probabilistic method (and the assumptions given in table 3) to simulate this reasoning.
Record linkage methods are important for injury research or surveillance, because any single database is often inadequate.
Computer assisted methods, simulating the human judgment that two records from different sources actually refer to the same event or person, are only superior when a large number of records must be processed.
The basic mathematical theory behind probabilistic record linkage is not difficult to explain, and accords well with human intuition.
Despite the speed and sophistication of modern record linkage software, deficiencies in data quality are the greatest obstacles to successful record linkage.
Deterministic (exact) methods or careful human review of probabilistic results are required if record linkage is used to make inferences about individual cases.
Linking two or more databases entails ethical obligations beyond the use of each separate database.
We can calculate posterior probabilities for each pair of records from the ambulance list and the emergency department list. The highest score would be for the pair A10-E19, with posterior odds (equation 4) of about:
and therefore a posterior probability (from equation 1) much greater than 0.9999. Also scoring very high would be the other exact matches A01-E01, A05-E09, A07-E13, and A08-E15. These are easily identified in a sorted list (like table 3), and would also be found by deterministic computer methods.
Scoring not quite so high would be those pairs where one or more elements did not match, for example A03-E12, with posterior odds calculated as about:
and a posterior probability of 0.9996. Here, the admission year and sex were different, so that the likelihood ratio for these terms is (1−M)/(1−U). Our assumption that nearly all the ambulance records should match to an emergency department record resulted in a relatively large prior probability; the relatively large posterior probability thus reflects our judgment that the discrepancies are likely due to data entry errors. We would probably also accept the pairs A02-E07 and A04-E08, with posterior probabilities of 0.9990. Notice that it would be difficult for a human to find these probable matches on a longer list, and not simple to develop a deterministic computer strategy to identify them.
A09 presents a problem because it might be matched either to E17 or to E18, with posterior probabilities of 0.9805 or 0.9921, respectively. The pair A06-E19, with posterior probability of 0.9507, is also uncertain. Human judgment might help resolve such cases, but error is still possible. All other pairs of records not yet mentioned have much lower posterior probabilities, and would probably not be considered as potential matches.
This process could be made more sophisticated by allowing dates to differ by one day, separating month from day, penalizing missing data less than erroneous data, etc. We might also modify the M or U probabilities, or the prior odds, after reviewing initial results. This human/machine interaction should produce results that accord with human intuition, but can be expanded to manage thousands of records in each file.
Supported by Grant #R49/CCR119798-01 from the National Center for Injury Prevention and Control.
If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.