Data Descriptor | Open

Disambiguation of patent inventors and assignees using high-resolution geolocation data

Published online:


Patent data represent a significant source of information on innovation, knowledge production, and the evolution of technology through networks of citations, co-invention and co-assignment. A major obstacle to extracting useful information from this data is the problem of name disambiguation: linking alternate spellings of individuals or institutions to a single identifier to uniquely determine the parties involved in knowledge production and diffusion. In this paper, we describe a new algorithm that uses high-resolution geolocation to disambiguate both inventors and assignees on about 8.5 million patents found in the European Patent Office (EPO), under the Patent Cooperation Treaty (PCT), and in the US Patent and Trademark Office (USPTO). We show this disambiguation is consistent with a number of ground-truth benchmarks of both assignees and inventors, significantly outperforming the use of undisambiguated names to identify unique entities. A significant benefit of this work is the high quality assignee disambiguation with coverage across the world coupled with an inventor disambiguation (that is competitive with other state of the art approaches) in multiple patent offices.

Design Type(s)
  • data integration objective
Measurement Type(s)
  • text disambiguation
Technology Type(s)
  • Computational Technique
Factor Type(s)
    Sample Characteristic(s)
    • United States of America
    • Europe
    • global

    Background & Summary

    In many contexts, technological progress and innovation is essential to national or regional economic growth and output1. One way of measuring innovation is the production of patents, which represent a technological advancement produced by individuals (generally, these are inventors on the patents) working at research institutions (generally, these are the assignees on the patent). The analysis of bibliographic databases has provided techniques for evaluating information spillovers2,​3,​4, mobility between regions5,​6,​7, interregional and international collaborations6,8,​9,​10,​11, and the emergence of new technologies12 in the context of patent and publication activities. Among many others, these studies have provided an in-depth picture of the dynamics of regional and institutional talents and quantify the success of various inter-organizational and inter-regional collaboration, of great use to policy makers on the institutional, regional, and international level.

    A major problem in the use of patent data (or any bibliometric database, such as for scholarly publications13,14) is the disambiguation of authors and institutions. There are a wide range of alternate spellings of a person’s or institution’s name, where, for example, ‘The National Institutes of Health’ and ‘NIH’ may refer to the same institution. Typos and misspellings of names are also common in bibliographic data (e.g., ‘National Institute of Health,’ missing an ‘s’ in the second word). The goal of disambiguation is to link all of these alternate spellings of institutional or individual names without incorrectly linking similar names referring to distinct entities. This is a difficult task, as there are millions of names to disambiguate (making pairwise comparisons of the full dataset computationally expensive) and an evaluation of how likely two names on patents are to be the same entity is not known a priori and often relies on machine learning techniques13,15,​16,​17,​18. This disambiguation problem has been effectively approached in recent years in the context of patent data using Baysian methods15, Markov Chain Monte Carlo approaches19,20, structural equivalence or other network similarity properties17,21, and supervised machine learning22, with a significant effort continuing at the USPTO18 with a focus on inventor disambiguation using an efficient hierarchical clustering approach19,20. Each of these approaches has a differing coverage (generally focused on inventors in the USPTO), and a single approach that effectively disambiguates both assignees and inventors in multiple national patent offices is of great value in disambiguating the names of individuals and institutions globally.

    In this paper, we describe a straightforward but accurate approach to the disambiguation problem using high precision geolocation of assignee and inventor addresses, depicted in Fig. 1 for the Boston and Paris areas. Two inventors (or two assignees) that provide exactly the same high-resolution address and also have ‘similar’ names are very likely to refer to the same entity. Thus, knowing that two entities have exactly the same address allows a great deal of flexibility in name matching, and we design two simple string matching approaches to link similar names that share a high-resolution geolocation. Inventors and assignees that have addresses with low resolution, which are extremely common in the US Patent and Trademark Office (USPTO) data, are linked to exact name matches nearby, greatly increasing the coverage of the disambiguation. Additional linking can be accomplished by searching for similar names that share other characteristics in common (such as co-inventors or co-citations). We show that this approach provides a complete disambiguation of inventors and assignees on 8.47 million patents, and that the precision and recall of the resulting disambiguation is superior to or competitive with other well known disambiguation methods.

    Figure 1: High-resolution geolocations of assignees in the Boston (left) and Paris (right) regions.
    Figure 1

    All high-resolution geolocations in Boston are included (~1.1 K within 10 km of the city center), but due to the density of assignees in Paris, only lat/longs that are seen on at least three patents are shown (~5.2 K within 10 km of the city center). Clusters of geolocations are observed near the Massachusetts Institute of Technology, the Massachusetts General Hospital, Boston University, and downtown Boston.

    Our algorithm is one of many approaches to the problem of disambiguation of inventors and/or assignees in patent databases, and differs in many ways from previous methods. Our approach combines a flexible but simplistic name matching algorithm at high resolution geolocations with a stricter name matching at low resolution geolocations applied to both assignee and inventor names, but other previous works use more complex Bayesian or supervised machine learning algorithms. In Table 1, we highlight some of the primary differences between our approach and previous methods. Li et al.15 used a Bayesian algorithm with an iterative blocking scheme applied only to USPTO inventors, which may outperform our approach in more difficult-to-disambiguate cases. Pezzoni et al.17 performed a disambiguation of EPO inventors matching similar names that share additional characteristics in common. Ventura22 used a large and verified database of pre-disambiguated inventors in the OptoElectronics field to design a supervised learning technique that performs extremely accurately on that subset of the data. The OECD provides the HAN database23 that corrects for a variety of alternate spellings and legal distinctions of assignees for patents in the USPTO, EPO, and PCT. Each of these algorithms has been shown to be accurate and useful in certain contexts, but none provide a broad and unified approach to the disambiguation of assignee and inventor names in multiple patent offices.

    Table 1: A comparison between different existing disambiguation algorithms covering patent data.

    Our results are a significant improvement over existing disambiguation techniques in a number of ways. This is the first simultaneous disambiguation of names in the EPO, PCT, and USPTO, providing a bridge between two existing but often non-overlapping streams of research focused on European4,8,10 and US12,15,24,25 patents. This combined disambiguation is of particular use in minimizing bias due to European or American institutions tending to apply for patents in their own domestic offices (the so-called home bias), and ensuring fair coverage of research hubs on both sides of the Atlantic. A disambiguation of both the assignee and inventor names is an aspect lacking in most freely available disambiguations of patent databases. This is of particular importance for understanding the mobility of inventors between institutions (whether in the same region or in different countries), as well as identifying the institutional research hubs that are major players in each country and region. Finally, we emphasize that our disambiguation also incorporates high-resolution geolocations of all inventor and assignee, providing geospatial information that can go beyond traditional regional or national aggregations based on (often arbitrary) administrative boundaries26. We believe the breadth and level of detail in the database produced by this work will be of great value to researchers, and will make the data freely available for noncommercial use.


    Patent data sources

    In this paper, we combine a number of distinct databases covering different patent offices: the US Patent and Trademark Office (USPTO), the European Patent Office (EPO), and patents filed under the Patent Cooperation Treaty (PCT). The USPTO and EPO generally provide protection of intellectual property within the US and European Union (EU) respectively, while the PCT provides a method of international protection for a patent filed domestically. In general, US firms tend to be most represented in the USPTO, European entities in the EPO, and other nations (particularly developing nations) in the PCT database.

    Information about USPTO patents is extracted from Harvard’s Dataverse project15 from the work of Ref. 15 (covering patents granted between 1975 and 2010), while information about EPO and PCT patents is extracted combining the Organization for Economic Cooperation and Development’s (OECD) RegPat27 and Citation28 databases (the 2014 version, covering patents applied for between 1977 and 2011). We also incorporate the OECD’s Triadic Patent Families29 database in this work, which identifies patents that cover the same technology filed in the USPTO, EPO, and Japanese Patent Offices (JPO). This combination of the USPTO, EPO, and PCT databases provides the application year for each patent, the names and addresses of all inventors and assignees, citations between patents (including those filed in different offices), and patents that are found in the same triadic family. Note there may be no legal requirements or error checking on these addresses. Inventors are free to use their home or office addresses, and assignees can use a PO Box if desired, leading to great heterogeneity in the address fields.

    Our method will be compared with two other freely available methods of name cleaning and disambiguation that exist in the literature: for inventor disambiguation we will compare our results to the output of Ref. 15, and for assignee disambiguation we will compare our results to the name cleaning provided by the OECD Harmonized Applicant Name23 (HAN) database. This paper will focus heavily on direct collaborations of assignees and inventors, direct citations, or exact family correspondence, and will not use additional information available in the databases such as the technological classes of each patent. Technological classifications have multiple potential levels of aggregation, ranging from very general topics like ‘Pharmaceuticals’ to specific topics like the full International Patent Classification (IPC) code ‘A61K003/121: Medicinal preparations containing acyclic ketones,’ and it is not clear which classification scheme is best for the disambiguation process. Future work may incorporate this additional information (with patent classifications playing a major role in other disambiguation methods15), but the high fidelity of our results with benchmarks suggest the incorporation of additional bibliographic data will provide a relatively moderate improvement.

    Geolocation and disambiguation of assignees and inventors

    The fundamental difficulty that must be overcome in name disambiguation is the possibility of alternate or error-ridden spellings of names or addresses in the database. Two examples of names requiring disambiguation are depicted in Fig. 2: on the left are some of the names associated with the National Institutes of Health (NIH) in Bethesda MD (assignee on ~4,900 patents), and some of the names associated with a prolific inventor, Rosa Maria Cuberes-Altisent, in the Barcelona area (inventor on 90 patents). In both cases, the subsets of names shown in the white boxes of Fig. 2 indicate the extreme heterogeneity in some inventor or assignee names, with ~90 unique alternate spellings for the NIH and ~10 unique alternate spellings for Rosa Maria Cuberes-Altisent (removing spacing, punctuation, and capitalization). Disambiguation of these names requires not only matching all of the possible variations in the spelling of the institution or individual, but also not matching other names that refer to different entities with similar names. A wide range of methods of varying complexity have been generated to solve this disambiguation problem for authors of publications13,14,30,​31,​32,​33, patent assignees16,34,35, and patent inventor names15,17 (with the disambiguation of Li et al.15 a recent and comprehensive result for the USPTO). In the case of inventor disambiguation, these methods will generally compare pairs of names using the similarity of the text of the names as well as data regarding the assignees, patent citations, patent classes, and geographical information.

    Figure 2: Geolocations of assignee addresses (left) or inventor addresses (right) for two examples.
    Figure 2

    The NIH in Bethesda Maryland, and an inventor in the Barcelona area with many names and addresses. In both cases, there is a great deal of heterogeneity of names at some geolocations, but many of the names are ‘similar’ at precise addresses. There are also exact name matches nearby (within 20 km in this paper), highlighted in bold text.

    The geographical information found in the USPTO typically suffers from low quality addresses, where below 5% of USPTO inventors complete the street field in their address on the patent (city- or zipcode-level information is the highest resolution available). At this resolution, geolocation can be used as one of many rough indicators of the similarity between two names when comparing them for disambiguation (compare the small number of locations in Fig. 3 to the dense coverage of high-resolution locations in Fig. 1). However, patents in the EPO or PCT databases are found to contain higher resolution addresses in a far greater fraction of cases (where a street number, street, city, state, and zip code are often all provided), which can provide much greater specificity when comparing inventors: if two inventors have ‘similar’ names and live at exactly the same address, it is far more likely they refer to the same person than if they had ‘similar’ names and lived in the same general area. The same state of affairs exists for assignee names, with the EPO and PCT addresses often having street-level information and the USPTO addresses tending to be of low quality.

    Figure 3: The handful of geolocations of low-resolution addresses in the Boston area.
    Figure 3

    Note that there are ~40 K patents found at these ~180 points, in comparison to the ~15 K patents found at the ~1,100 points in Fig. 1. The loss of specificity in these geolocations would make flexible name matching unreliable at these low resolution geolocations.

    Our strategy for disambiguation will be to leverage these high precision addresses in the EPO and PCT by flexibly matching names that are simultaneously found at exact, high-precision geolocations in any patent office, then to match ‘nearby’ names that are exactly the same. ‘Nearby’ is taken as distance of ≤20 km in this paper, which is a rough upper bound on the mean commute distance in major metropolitan regions in the US36. Small variations in this threshold are not expected to significantly alter our results, but a significant decrease or increase in the threshold may cause significantly fewer or more (respectively) links between names. The general idea is sketched in Fig. 2 for specific assignee and inventor names, and a schematic of the methodology is diagrammed in Fig. 4. A detailed summary of each step can be found in the appendix. We first geolocate the assignees and inventor addresses for every patent in the three databases (~4 million unique addresses) using Yahoo’s YQL API26, converting the text into likely latitude/longitude (lat/long) pairs and the quality of that geolocation (step 1 in Fig. 4). The quality returned by YQL generally indicates if the geolocation was resolved at the level of a point (e.g., ‘55 Fruit Street, Boston, MA, USA’), line (e.g., ‘Fruit Street, Boston, MA, USA’), zip code (e.g., ‘02114 MA USA’), city (e.g., ‘Boston, MA, USA’), state (e.g., ‘MA, USA’), or country (e.g., ‘USA’). Low quality geolocations are far less informative than high resolution geolocations, with a far greater number and diversity of inventor and assignee names at each low res point than at at a high resolution location (compare Fig. 3 and Fig. 1). In early 2016, Yahoo discontinued the use of the particular API used in this paper, and no longer appears to produce the same quality indicator (see Other geolocation APIs often provide quality indicators (e.g., the Google Maps API labels geolocations as ‘rooftop’ for the highest resolution, with other indicators for lower resolution). While the results of this paper are strongly dependent on the division between high and low resolution geolocations, we expect our results to be robust to changes of the specific quality indicator used.

    Figure 4: Schematic diagram of the method.
    Figure 4

    Names and addresses are first geolocated (discussed in Sec. 1 of the SI), after which there is a search for ‘similar’ names at exact high resolution lat/longs (the meaning of ‘similar’ is discussed in Sec. 2 of the SI for assignees and Sec. 3 of the SI for inventors). Once similar names are clustered at each lat/long, ‘nearby’ exact name matches are linked (described in Sec. 4 of the SI). Assignees and inventors names that could not be geolocated are linked to disambiguated names with shared coinventorships (described in Sec. 5 of the SI). Finally, similar inventor names for inventors that are not found at the same address but share other characteristics in common before are linked, creating identifiers for mobile inventors (described in Sec. 6 of the SI).

    For every high-quality geolocation in the data (those better than line-resolution) we attempt to flexibly match similar strings (step 2 in Fig. 4), with specific examples in Fig. 2: assignees with names involving ‘institute’ or ‘health’ at identical lat/longs are likely to be referring to the NIH, and inventors with names like ‘cuberes’ and ‘rosa’ at identical lat/longs are likely referring to Rosa Maria Cuberes-Altisent. The specifics of the matching can be found in Sec. 3 of the SI, but generally speaking we match pairs of assignee names that share either one rare or two common words (in any order, up to one spelling error) and match pairs of inventor names for which the first two words in either name occurs in the other (in any order, up to one spelling error), so long as both names share a high-resolution geolocation. After this first round of disambiguation, which only compares pairs of names found at exactly the same high-resolution geolocations, we search for exact name matches that were geolocated to a lat/long pair of low quality (having a YQL quality indicator on the level of city-resolution) in step 3 of Fig. 4. For example, we link the names at the indicated high-resolution geolocations due to the simultaneous name matchings of ‘National Institutes of Health’ and ‘Cuberes Altisent, Rosa’ occurring in Fig. 2. We also search for inventor name prefixes that clash (where an inventor may leave his or her middle name out on some patents) to prevent incorrect linking of low resolution names. The process is fully described in Sec. 4 of the SI. Steps 2–3 in Fig. 4 are run in parallel to one another, with assignee and inventor disambiguation independent of each other up to this point.

    Patents that have assignees that could not be geolocated at the precision of city-level or better are common, particularly in the USPTO data, and it is important to include patents with extremely poor geolocations in the disambiguation as well. Matching assignee names to the disambiguation not straightforward if multiple disambiguated IDs have the same name as the un-geolocated assignee: for example, one must determine whether the assignee ‘General Electric Corporation’ without an address provided (occurring over 18,500 times in the USPTO) is referring the same institution as the ‘General Electric Corporation’ found in New York, Ohio, Massachusetts, or many additional potential geolocations found for that name. We identify the most likely disambiguated ID for linking an unlocated patent as that with a similar name and shared inventor or an inventor in close physical proximity to the geolocated assignee (a detailed description of this method is found in Sec. 5 of the SI).

    The results of the disambiguation after step 4 of Fig. 4 produces assignee and inventor IDs that are highly localized, linking names only if their geolocations are within a 20 km radius of one another. For assignees this localized partitioning is reasonable, with different corporate offices, research labs, or subsidiaries of the same company in different cities treated as independent entities (for example, ‘General Electric Corporation’ in Schenectady NY is not the same as ‘General Electric Corporation’ in Cincinnati OH). Inventors are fundamentally different in that they can be mobile, moving between different cities and countries. We therefore perform a final round of linking on the disambiguated inventor IDs in step 5 of Fig. 4 based on characteristics in common that indicate a potential relationship: a shared co-inventor, shared assignee, shared triadic family, or citation. Note that we do not include the technological classes of the patents in our measure of similarity, because they are less personal (working on a research topic is far less informative than collaborating with an individual) and the ‘correct’ level of aggregation is unclear. A link between disambiguated IDs is performed for inventors with identical names and one characteristic or similar names and two characteristics in common, providing a method to identify the movement of unique individuals between different cities.

    This procedure produces a disambiguation of ~9.3 M patents in the three patent offices. A total of ~804 K unique geolocated assignee identifiers are found, as well as ~443 K unlocated assignee IDs (due to missing addresses that could not be linked to the disambiguation). We find ~3.8 M unique geolocated inventor IDs that do not move (i.e., are only found in the same 20 km radius), ~425 K mobile inventors (IDs that are linked to geolocations more than 20 km apart), and ~439 K inventor IDs that could not be linked to a location.

    The algorithm uses geospatial information throughout, so the unlocated IDs are likely to be error prone and unreliable. Because correction of errors in spelling occurs only in steps 1 and 5 of Fig. 4, it may be important to further distinguish between ‘high-quality’ disambiguated IDs (those involving both geolocation and spelling correction), and ‘low-quality’ disambiguated IDs (those using only geolocations to link nearby identical names). High-quality inventor or assignee IDs can be linked to a high-resolution geolocation, while high-quality inventor IDs also include those linked in step 5 through additional shared characteristics. We find a total of ~290 K high quality assignees and ~1.8 M high quality inventor IDs using this definition, with the best coverage in the EPO and worst coverage in the USPTO (see Table 2). Our method produces a complete disambiguation of all assignees and inventors on ~8.5 M patents, and complete high-quality disambiguations on ~4 M patents worldwide (see Table 3). The disambiguation thus completely geolocates and disambiguates all names on over 90% of the patents in these offices, with nearly 50% of the those patent disambiguations incorporating high resolution addresses and error correction in the names.

    Table 2: Summary of the assignee and inventor coverage of the disambiguation in the three patent offices (patent count in millions).
    Table 3: Summary of the total coverage of the disambiguation in the three patent offices (patent count in millions).

    Data records

    The results of our algorithm are being made freely available, with the full release found in Data Citation 1: Figshare The content and format of each individual data file in the release is described below.

    Patent descriptor file

    Main descriptor file. File Name: all_disambiguated_patents_withLocal.txt (Data Citation 1: Figshare

    This is a list of every patent with a disambiguated assignee or inventor in our disambiguation. There are 7 columns and 9.3 M lines, ‘|’ delimited; the columns are:

    1. pat: the patent publication number. Each can be linked exactly to either the OECD or USPTO databases for further processing if needed.

    2. invs: the inventors on the patent. If multiple inventors are found, they are comma separated.

    3. localInvs: the local IDs of the inventors. If no mobile inventor is found, this is identical to the invs column. If multiple inventors are found, they are comma separated.

    4. apps: the assignees (applicants) on the patent. If multiple assignees are found, they are comma separated.

    5. yr: the application year of the patent.

    6. classes: the WIPO industrial fields of the patent. If there were more than one field, they are comma separated. These are integers between 1 and 36, the meaning of which is in the file wipo_class_ID.txt (described in Sec. 3.1.2).

    7. wasComplete: a flag to indicate whether too many or too few inventors or assignees were found. A ‘1’ indicates no problem was detected with this patent, a ‘0’ indicates a count difference between disambiguated IDs and names found on the patent. There are 134 K errors of this type in the data.

    In the paper, we describe dividing disambiguation into high- and low-quality, based on whether geolocation and name cleaning was performed for that ID. These are indicated by the first two characters in the IDs themselves:

    High quality IDs:

    • HA: High-resolution assignee (linked in step 1)

    • HI: High-resolution inventor (linked in step 1)

    • HX: High-resolution cross-linked inventor (linked in step 5)

    • HM: High-resolution mobile inventor (linked in step 5)

    • LX: Low-resolution cross-linked inventor (linked in step 5)

    • LM: Low-resolution mobile inventor (linked in step 5)

    Low quality IDs:

    • LA: Low-resolution assignee

    • HS: High-resolution split inventor

    • LI: Low-resolution inventor

    • LS: Low-resolution split inventor

    • UI: Unlocated inventor

    • UX: Unlocated cross-linked inventor

    • US: Unlocated split inventor

    Each of these two-letter codes are followed by a number:

    • For assignees (where the second character is ‘A’), the number indicates that assignee’s rank by total number of patents. In the case of a tie, the rank is alphabetical. For example, HA1 is a high resolution assignee with the most patents in the world, and LA4 is a low resolution assignee with the fourth most patents in the world.

    • For local inventor IDs (where the second character is NOT ‘M’ or ‘A’), they indicate the rank of that local ID by total number of patents. In the case of a tie, the rank is alphabetical. For example, HI1 is the most prolific high-resolution inventor in a single 20 km region, while LI2 is the second-most prolific inventor in a single 20 km region.

    • For mobile inventor IDs (where the second character is an ‘M’), they indicate the rank among other mobile IDs of that Inventor’s most prolific local ID. For example, LM1 is the mobile inventor with the greatest number of patents in one 20 km region, and all geolocatons for that inventor were low resolution (indicated by the ‘L’). HM2 is the mobile inventor with the second greatest number of patents in one 20 km region, and at least one geolocation was high resolution (indicated by the ‘H’).

    Top-level patent classifications in the patent descriptor file

    File Name: wipo_class_ID.txt (Data Citation 1: Figshare

    This is a list of the link between WIPO fields37 and the indicator found in

    all_disambiguated_patents_withLocal.txt. There are two columns, 37 lines,‘,’ delimited; the columns are:

    1. wipoClass: the top level field

    2. ID: the number between 1–36 corresponding to that class.

    Mobile inventor links. File Name: mobilityLinks.txt (Data Citation 1: Figshare

    This is a list of local inventor IDs (IDs found within 20 km of one another) to the mobile IDs detected in step 5. 2 columns, 1 M rows,‘,’ delimited; the columns are:

    1. localID: local ID of the inventor.

    2. mobileID: mobile ID of the inventor.

    Disambiguated IDs to names and locations

    Inventors. File Name: LinkedInventorNameLocData.txt (Data Citation 1: Figshare

    This is a list of the patents, geolocations, and names of every inventor. There are 6 columns and 25.3 M lines,‘|’ delimited; the columns are

    1. ID: the local ID of the inventor

    2. pat: the patent publication ID.

    3. name: the raw, undisambiguated name of this inventor on this patent.

    4. loc: the geolocation of the inventors address listed on this patent.

    5. qual: the yahoo quality of the geolocation.

    6. loctype: the division of the geolocation into high and low resolution. If one wishes to work with individual data at a high resolution, do not use ‘low.’

    Assignees. File Name: LinkedAssigneeNameLocData.txt (Data Citation 1: Figshare

    This is a list of the patents, geolocations, and names of every inventor. There are 6 columns and 8.2 M lines, ‘|’ delimited; the columns are

    1. ID: the ID of the assignee

    2. pat: the patent publication ID.

    3. name: the raw, undisambiguated name of this assignee on this patent.

    4. loc: the geolocation of the assignee’s address listed on this patent.

    5. qual: the yahoo quality of the geolocation.

    6. loctype: the division of the geolocation into high and low resolution. If one wishes to work with data at a high resolution, do not use ‘low.’

    Patent citations

    File Name: Citation_App+Exa.txt (Data Citation 1: Figshare

    This is a list of patent citations found in the OECD or Dataverse15 databases. Non-patent citations (e.g., to journals) are not included. Patent citations can be added by the applicant or the examiner in the office (the latter suggesting it is relevant but the inventor may not have been aware of it). A typical USPTO patent has significantly more outgoing citations than a typical EPO/PCT patent, and depending on the research topic under study, one may need to carefully consider the impact of this difference between US and European offices when looking at citation statistics. 4 columns, 42.5 M lines, ‘|’ delimited; the columns are:

    1. citing: the patent on which the citation is found

    2. cited: the patent that is cited.

    3. cite_origin: whether the citation was supplied by the applicant (labeled ‘a’) or examiner (labeled ‘e’).

    4. data_source: whether the citation information originated in the USPTO database (‘u’), in the EPO’s database (‘e’), or in the PCT database (‘w’).

    Patent classifications

    Main classification file. File Name: Patent_classes.txt (Data Citation 1: Figshare

    This is a harmonized list of patent classifications on all levels of resolution using the data we had available. USPTO patents in the Dataverse15 databases had US Patent Classifications (USPCs), while EPO and PCT had International Patent Classifications (IPCs) and Cooperative Patent Classifications (CPCs). The concordance between the two classification systems is provided in the USPC-IPC correspondence file (described below). 3 columns, 79.6 M lines, ‘|’ delimited; the columns are:

    1. Pub_number: the patent publication number.

    2. class: the patent classification number.

    3. class_type: the type of classification this represents. ‘u’ indicates USPC, ‘i’ indicates IPC, ‘c’ indicates CPC, ‘m’ indicates an inferred IPC from the USPC-to-IPC concordance (described below), and ‘w’ indicates the WIPO field (described in the Main classification file (described above)).

    USPC-IPC correspondence. File Name: USPC_to_IPC_fullList.txt (Data Citation 1: Figshare

    The USPC and IPC systems were harmonized using the concordance provided by ReedTech at In order to simplify the classification of patents, we compared all IPCs found in the OECD’s database to the USPTO classes liked in ReedTech’s concordance table and assigned them to the appropriate USPTO label. 2 columns, 183 K lines, ‘|’ delimited; the columns are:

    1. USPC: the USPC

    2. IPC: the IPCs that are found in our database that are equivalent to that USPC. If more than one IPC is equivalent, they are separated by a ‘,’.

    IPC-field correspondence. File Name: IPC_to_Industry_WIPOJan2013.txt (Data Citation 1: Figshare

    A linking between IPCs and coarse Sector and Field data, as defined in the Schmoch report37. 3 columns, 754 lines, tab delimited; the columns are

    1. IPC_code: the IPC code

    2. Sector_en: the top level sector classification.

    3. Field_en: the top level field classification.

    Triadic patent families

    File Name: patent_to_triadic_family.txt (Data Citation 1: Figshare

    A linking between all EPO, USPTO, and PCT patents and the triadic families indicated by the OECD data. Two columns, 3.2 M lines, ‘|’ delimited; the columns are

    1. Pub_number: the patent’s publication number

    2. Triadic_Family: the triadic family in which the patent is included.

    New benchmarks

    Harvard inventors. File Name: combined_labels_HarvardInv.txt (Data Citation 1: Figshare

    The manual Harvard disambiguation used as one of the benchmarks in the paper. In all columns, if more than one ID is found, the elements are comma separated. 6 columns, 587 rows, ‘|’ delimited; the columns are:

    1. Pub_number: the publication number of the patent

    2. ManualIDs: The manually disambiguated IDs of each inventor on the patent.

    3. OurIDs: the output of our algorithm on this patent.

    4. LiLow: The ‘low’ disambiguation of Li et al. on this patent.

    5. LiHigh: the ‘high’ disambiguation of Li et al. on this patent.

    6. RawNames: the undisambiguated names on the patent, with case and punctuation dropped.

    Boston-area assignee benchmark

    File Name: combined_labels_boston.txt (Data Citation 1: Figshare

    The manual disambiguation of EPO/PCT assignees in the Boston area, used as one of the benchmarks in the paper. In all columns, if more than one ID is found, the elements are comma separated. 5 columns, 22528 rows, ‘|’ delimited; the columns are

    1. Pub_number: the publication number of the patent

    2. ManualIDs: The manually disambiguated IDs of each inventor on the patent. Each ID has the following structure: (1) a leading character of ‘r’, indicating within-region, or ‘e’, indicating an external collaborator; (2) a number with no specific meaning; (3) an ‘_’; and (4) a manual classification of the type of institution. The institution can have the values ‘university’, ‘hospital’, ‘company’, ‘lab’, and ‘other’.

    3. OurIDs: the output of our algorithm on this patent.

    4. HanIDs: The IDs for this patent from the HAN databases.

    5. RawNames: the undisambiguated names on the patent, with case, punctuation, and spacing dropped.

    Note that there is no correspondence between the boston-area IDs and the paris-area IDs (described in the Paris-area assignee benchmark section (described below)). Do not attempt to mix the two disambiguations together as incorrect ID-clashes are possible.

    Paris-area assignee benchmark

    File Name: combined_labels_paris.txt (Data Citation 1: Figshare

    The manual disambiguation of EPO/PCT assignees in the Paris area, used as one of the benchmarks in the paper. In all columns, if more than one ID is found, the elements are comma separated. 5 columns, 18877 rows, ‘|’ delimited; the columns are:

    1. Pub_number: the publication number of the patent

    2. ManualIDs: The manually disambiguated IDs of each inventor on the patent. Each ID has the following structure: (1) a leading character of ‘r’, indicating within-region, or ‘e’, indicating an external collaborator; (2) a number with specific meaning; (3) an ‘_’; and (4) a manual classification of the type of institution. The institution can have the values ‘university’, ‘hospital’, ‘company’, ‘lab’, and ‘other’.

    3. OurIDs: the output of our algorithm on this patent.

    4. HanIDs: The IDs for this patent from the HAN databases.

    5. RawNames: the undisambiguated names on the patent, with case, punctuation, and spacing dropped.

    Note that there is no correpsondence between the boston-area IDs (described in 3.6.2) and the paris-area IDs. Do not attempt to mix the two disambiguations together as incorrect ID-clashes are possible.

    Technical Validation

    Inventor and assignee benchmarks

    We will quantify the compatibility of a trial disambiguation to the ‘correct’ ground truth partition using four different indicators: splitting, lumping, precision, and recall. Splitting and lumping are two related statistics to measure the similarity between two partitions that have been previously used for inventors in the UPSTO15, which respectively estimate the number of patents that are missing in the trial disambiguation in comparison to the benchmark (i.e., a true inventor is split into multiple IDs in the trial) and the number of patents that are added in the trial disambiguation compared to the benchmark (i.e., multiple trial IDs point to the same true inventor). The splitting associated with trial ID i is determined by identifying the true ID that shares the most patents, mi; with Pit the set of patents invented by i in the trial partition and PmiT the set of patents invented by mi in the true disambiguation, the splitting is si=|Pit\PmiT| (i.e., the number of patents in the trial that are associated with i but not with mi in the true disambiguation). Lumping is similarly defined for the true ID j by identifying the trial ID Mj with which it shares the most patents, with the lumping of j given by lj=|PjT\PMjt| (the number of patents in the trial that are associated with Mj but not with j). To perform this matching, we identify the best link for the ID with the largest number of patents (i or j, depending on if we are computing splitting or lumping), then progressively search for the best matches for smaller IDs that have not been previously assigned. Rarely occurring IDs may not be assigned a best match using this procedure (if all potential matches were previously assigned), and in this case all occurrences of that ID will be treated as erroneous. The total splitting and lumping of the trial partition are given by (1)split=isii|Pit|lump=jljj|PjT|, which estimates the fraction of patent that suffer from a splitting or lumping error.

    Splitting and lumping in equation (1) link trial IDs to true IDs that share the most patents in common, and treat patents assigned to a different ID as an error. While this is a useful statistic, it ignores the degree of splitting or lumping that has occurred: splitting remains unaltered by the number of IDs a true identity is associated with in the trial (and similarly for the lumping statistic), suggesting that some features of the similarity between partitions may be overlooked. We therefore also examine a pairwise measure of precision and recall of the trial and true disambiguations, schematically diagrammed in Fig. 5. The disambiguated name-to-ID’s nT(p)={niT(p)} for the ith name on patent p that can be compared to the disambiguated names in the trial disambiguation. For each pair of patents p1 and p2, we can determine the number of true positives, false positives, and false negatives between the trial and benchmark IDs by comparing the number of overlapping identifiers, as diagrammed in Fig. 5. Defining xT(p1,p2)=|{nkT(p1)nkT(p2)| as the size of the intersection of the IDs in the benchmark and xt(p1,p2)=|{nkt(p1)nkt(p2)| the size of the intersection in the trial disambiguation, there are at most TP(p1, p2)=min(xT, xt) IDs that agree in both partitions, FN(p1, p2)=max(0, xTxt) matches in the benchmark not seen in the trial, and FP(p1, p2)=max(0, xtxT) matches in the trial partition that don’t match in the benchmark. An estimate of the precision and recall for the trial is then (2)prec=p1p2TP(p1,p2)p1p2[TP(p1,p2)+FP(p1,p2)]rec=p1p2TP(p1,p2)p1p2[TP(p1,p2)+FN(p1,p2)] Note that this method does not compare the positions of the name matches (only the number of matches), and thus neglects the possibility of a transposition of the IDs in the trial (e.g., if names A and E are incorrectly linked together and simultaneously names B and D are incorrectly linked together). Due to the large number of IDs for both assignees and inventors, this type of error is expected to have a negligible effect. This approximation can be relaxed, at an increased computational cost. Using the definitions in equation (2), if all matches found in the trial are also found in the benchmark, the trial disambiguation will have high precision; and if all matches in the benchmark are found in the trial, the trial disambiguation will have high recall. These definitions of precision and recall are independent of a matching between the two partitions (that is, associating each trial ID with a ‘best’ true ID), comparing all assignments on an equal footing (but at greater computational cost due to the pairwise comparisons required).

    Figure 5: Schematic of the pairwise measurements of precision and recall.
    Figure 5

    Each name is assigned a unique ID in the benchmark (indicated by the color of the circles) and a ID in the trial disambiguation (indicated by nt). Two true positives occur if nAt=nDt and nBt=nEt. Two false negatives would occur if nAtnDt and nBtnEt. A false positive occurs if nCt=nFt.

    In order to evaluate the accuracy of our inventor disambiguation, it is necessary to find a set of ground truth identification of patent inventors: a manually curated subset of the patent data for which a correct disambiguation of the names has been performed. Such a benchmark exists for ~100 USPTO inventors of ~1300 patents in the area of engineering and biochemistry that has been used as a golden standard in the literature15. This is a complete disambiguation of all patents invented by a specific set of inventors. We also generated a manual disambiguation of USPTO patents filed by assignees containing the phrases ‘harvard college’ and ‘harvard university’ (case insensitive) in the name. This second benchmark is a complete disambiguation of all inventors on a specific set of patents, with 587 patents with 1,000 name/address pairs and 827 disambiguated inventor IDs (using our approach). In Tables 4 and 5, we compare the splitting, lumping, precision, and recall of four potential disambiguations: ‘none’ (meaning that case and punctuation are ignored but no further disambiguation is performed), ‘lower’ and ‘upper’ (two outputs of the disambiguation of Ref 15), and our approach. For the golden standard benchmark (Table 4), all patents linked to a trial ID are included when computing the statistics (including incorrectly-assigned patents, representing false positives), but un-disambiguated collaborators of these inventors are ignored. For the Harvard inventor disambiguation shown in Table 5, only the disambiguated patents are included (and no additional patents that may represent false positives are added) because we do not have a complete list of the ‘correct’ patents for each inventor. The exclusion of such false positives increase the apparent precision of each method in Table 5, since exact name matches for inventors affiliated with Harvard are very likely referring to the same person.

    Table 4: Benchmarking of inventor disambiguations on the Golden Standard benchmark.
    Table 5: Benchmarking of inventor disambiguations on patents assigned to Harvard.

    In the left side of Tables 4 and 5, we see that all disambiguation methods (‘lower,’, ‘upper,’ and ‘ours’) provide an improvement over the un-disambiguated results in all statistics, and in general the results of Li et al.15 produce a higher improvement in all statistics in comparison to our approach. The quality of our disambiguation is comparable to that of Li’s in the case of the Harvard dataset, but in the case of the Golden Standard benchmark our results have about half the lumping of the un-disambiguated name and major improvements in both precision and recall. The disambiguation of Li is a more significant improvement over the raw names in lumping, splitting, and recall, but is comparable to our approach in precision. A higher recall can be found by removing all IDs associated with low-quality disambiguations in our approach (those IDs that did not incorporate correction of spelling errors in steps 1 and 5 of Fig. 4). On the right sides of Tables 4 and 5, we discard all patents having at least one low-quality ID, as well as all names or Li IDs occurring on those patents. We find a significant improvement in all quality statistics using our approach in both benchmarks, although still marginally underperforming Li et al. in both datasets. These results indicate that our inventor disambiguation is expected to have high precision in all cases (i.e., will have very few false positives), but higher recall (very few false negatives) for high-quality disambiguated names. Low-quality disambiguations are expected to underestimate the number of patents invented by the person, but not significantly overestimate them.

    We are not aware of a freely available ground truth disambiguation of assignees, and in order to test the accuracy of our methods we manually generated our own benchmark from a by-hand disambiguation of a subset of patent assignees in the EPO and PCT data. To create a benchmark of manageable size, we focused on assignees in the EPO and PCT in specific regions active in specific fields of research. The OECD REGPAT database27 provides geolocation information on the level of NUTS3, and we generated a list of all assignees on patents assigned to names in Boston or Paris (the selection of assignees was made where Boston assignees have addresses in the NUTS3’s US25017, US25025, or US25021 and Paris have addresses in NUTS3’s beginning with ‘FR10’ in the OECD RegPat database27. We note that the OECD’s geolocation process is independent from our own, so this benchmark is not affected by any potential errors or incompleteness in our geolocation procedure.). These names include both regional assignees as well as external collaborators of those regional institutions, and in order to reduce the number of assignee names to disambiguate, we retained only assignees names found on at least ten biopharma patents (biopharma patents are defined as having an IPC classification which falls under the fields ‘Pharmaceuticals’ or ‘Biotechnology’ using the WIPO field-level aggregation of patent classes37). Exclusion of less-common names greatly reduced the number of manual matches that were required, but may introduce a bias against small startups, individuals that were assignees, or large institutions that rarely work in the biopharma fields. This produced a final set of ~700 disambiguated assignees (from ~900 raw names) on ~23 K biopharma patents in the Boston area and 640 disambiguated assignees (from ~1,100 raw names) on ~19 K biopharma patents in the Paris area.

    In Table 6, we determine the error statistics for undisambiguated names (removing spacing and punctuation); the HAN name harmonization, which accounts for synonyms such as ‘Co’ and ‘Company’ in assignee names (with unique HAN IDs extracted directly from the OECD’s HAN database23); and the method in this paper (generated using the OECD’s REGPAT database27). We are not aware of another freely available assignee disambiguation of EPO and PCT patents that can be used for comparison. As was the case in Table 5, false positives from patents not found in the manual disambiguation are excluded from this trial, and the precision is likely to be inflated (because our approach permits a maximum distance between geolocations of 20 km on the disambiguation of assignees, it is unlikely that the precision using our approach is significantly higher than what would be observed if all patents were included.). In all cases the HAN harmonization is a marginal improvement over the raw names, but our method produces a significant improvement over both in every statistic. The lower overall recall in the Paris area is due to many alternate spellings from the presence or absence of accents on words, a complication not typically present in the Boston area. Note that the names in the Paris area involve terms in French (e.g., ‘societé aononyme’ is a commonly occurring substring of assignee names), and the high precision of our disambiguation in the Paris area suggests that the algorithm is robust to variations in language. Over 95% of the patents in Table 6 are assigned to solely high-quality assignee disambiguations, and there is little benefit in dividing the results into differing qualities of disambiguation (as was beneficial in Tables 4 and 5). The error statistics in Table 6 suggest that our assignee disambiguation is likely reliable in many major global research hubs.

    Table 6: Benchmarking of assignee disambiguations in the EPO and under the PCT for assignees geolocated in the Boston (left) or Paris (right) areas.

    Leading inventors and assignees

    Bibliometric data has been implemented in a variety of contexts in both publications7,11,38 and patents2,4,10,24,25,39 to track individual careers, institutional collaborations, and information flows between regions. In almost all studies, major players represented by top authors, inventors, or institutions tend to be emphasized (as measured by the number of patents or publications, number of citations, or other quality indicators), due to their high productivity and output. It is therefore worthwhile to determine how accurately our approach is able to identify the top players in the world, and what advantages our method may have over existing datasets.

    In Table 7, we list the top ten players as determined by Li’s ‘lower’ inventor disambiguation of USPTO inventors (the ‘upper’ disambiguation does not differ significantly on these names) compared to the output of our methods. We find a negligible difference (typically less than 2% variation in the patent counts) between the approaches in the majority of cases, but significant differences do sometimes occur. In the case of Gurtej Sandhu, we find our algorithm detects an additional ~100 patents, an improvement over Li’s approach sufficient to change his ranking from 7th to 5th place in our ordering. Our method sometimes does underestimate the patents invented by top players, particularly for Japanese inventors. This is due to the low-quality of geolocations provided by YQL in Japan: of the ~100 K address containing the word ‘Tokyo,’ 0.6% produced a high-resolution geolocation (meaning that virtually no Japanese inventor will have a high-resolution geolocation) but 82% are found on the neighborhood level or better (meaning they contained more information than simply a city name). It may therefore be reasonable to expect a higher number of splitting errors for Japanese inventors using our data. Table 7 indicates that, with the exception of these splitting errors of Japanese inventors, our method is able to reliably detect prolific inventors consistent with Ref. 15.

    Table 7: The top ten patent inventors by number of patents in the USPTO (1980–2010), ranked using the disambiguation of Li et al.15 and ours (matching IDs between the two algorithms were based on the number of shared patents).

    A surprising aspect of Table 7 is that not a single European is found in the list of the top 20 most prolific inventors. This is due to an intrinsic bias in the choice of USPTO patents: Europeans may be more likely to apply for patents in the EPO (which grants protection in Europe where they live and work) than in the USPTO. Non-US and non-European inventors and assignees may likewise be more likely to submit patents to under the PCT than in either the USPTO or EPO, to ensure international and domestic protection of their inventions. A significant advantage of our method is that these three major offices are linked in the disambiguation, permitting a more even accounting of highly prolific inventors. In Table 8, we see that while the top five inventors worldwide all focus their patenting activity most heavily in the USPTO, many prolific inventors focus their patent applications in the EPO or under the PCT. We are able to identify a number of top inventors from China (filing their patents almost exclusively under the PCT) and Germany (filing their patents in both the EPO and USPTO, typically), many of whom would not be seen in the top hundred inventors when focused solely on the USPTO. When these disambiguated IDs are low resolution, we continue to expect most errors to be splitting (thus underestimating the total number of patents invented by that person) instead of lumping (which would overestimate that person’s patent count). Table 8 suggests that our inventor disambiguation may be more appropriate in studying the effects of global knowledge production and international knowledge spillovers through collaborations across the Atlantic as well as in emerging economies.

    Table 8: The twenty most prolific inventors in the combined USPTO, EPO, and PCT according to our disambiguation (recall that Table 7 included only USPTO patents), along with the specific counts per office.

    It is also useful to compare our disambiguation to that of Li et al.15 on names that are in some sense difficult, rather than solely on the most prolific 20 inventors. To do so, we also extract the top 20 disambiguated inventors in the USPTO using our algorithm with a two- or three-character last name. These last names tend to be of Asian descent and to be more common globally than long names are, and we might naturally expect a greater level of lumping errors in these names than the (generally more European) names in the top 20 globally prolific inventors. In Table 9, we compare our IDs to those of of Li et al., and examine the level of overlap between the two disambiguations. In 75% of the cases, there is a negligible difference between the two disambiguations, and a ≥95% overlap between our disambiguation and the most common shared ID in Li’s disambiguation.

    Table 9: Comparison of our disambiguation to that of Li et al.15, restricted the 20 most prolific USPTO inventors with at most three letters in their last name.

    The column ‘diff’ in Table 9 indicates the apparent meaning of the difference between our disambiguation in the 4 inventors for which our disambiguation differs significantly from Li’s. It takes on one of three values, denoting the apparent correctness of differences between our disambiguation and Li’s: ‘−’ denoting a negligible difference (16 of the 20 cases), ‘G’ denoting an apparent gain of information by connecting distinct IDs in Li’s disambiguation that likely refer to the same person (3 of 20 cases), and ‘L’ an apparent loss of information by apparently erroneously connecting IDs in Li’s disambiguation that are probably correctly separated (1 of 20 cases). In two cases, the gains are due to the flexibility of our name matching algorithm, which overcomes difficulties in linking identities with multiple and variable middle names (Louis Lu Chen Hsu and Chen Hua Douglas Yu). The two other differences between our disambiguation and that of Li’s are due to our linking of exact name matches with other characteristics in common (citations, coinventorships, or assignees), leading to the beneficial merging in the case of Jerry Wu (working with assignee ‘Hon Hai Precision Ind. Co. Ltd.’ in both locations), and to the apparently detrimental clustering in the case of Hiroshi Ito (simultaneously working in multiple cities and with multiple different assignees). Lumping errors are thus possible using our algorithm (as they are in any algorithm), but we see no evidence of a systematic inflation of the patenting of short, commonly occurring names. Rather, Table 9 indicates that our approach is consistent with existing disambiguations and in some cases may improve the name disambiguation due to improvements with the flexible name matching.

    Another significant advantage of our methodology is the simultaneous disambiguation of assignees, allowing us to determine the major research institutions in addition to prolific inventors. In Table 10, we see the top patent assignees are dominated by computer and communications companies, with all members of the list firms with highly recognizable names. We also list the number of unique names that were involved in the disambiguation (spacing and punctuation ignored), highlighting the extreme variability of names that may occur. US and Japanese firms tend to patent in the USPTO (as was the case for inventors), and most European countries tend to patent more heavily in the EPO. The exceptions of Nokia Siemens in Finland and Ericsson in Sweden, who tend to evenly split their applications between the EPO and under the PCT, may be explained by these countries joining the European Union in 1995 (note that while Sweden joined the EU in 1995, it was a signatory of the European Patent Convention in 1978.). Note also that Nokia Siemens is found twice in this list: once in Germany and once in Finland, with the appropriate designations of a limited liability corporation (‘GMBH’ and ‘OY’, respectively) in their countries. Our method correctly treats these as distinct entities despite their similar names, due to the 20 km geolocation restriction on assignees. Note that in order to determine the number of patents held by a multinational corporation (in the case of Nokia Siemens, the NOKIA group), an ownership tree must be determined from some alternate datasource. All of the European firms in Table 10 would see their ranking drop significantly if only the USPTO were taken into account, with Nokia-Siemens’ German office the only EU research lab that would be included in the top 20. As was the case for inventors, by including the USPTO, EPO, and PCT, a broader, more international picture of major research institutions can be extracted than if only one were included.

    Table 10: The twenty largest assignee headquarters as measured by the total number of patents worldwide, 1980–2010, along with the nation of the geolocation and the number of names for that assignee that were used in the disambiguation.

    Our assignee disambiguation algorithm searches for similar names at identical positions and identical names at nearby positions, linking regional elements together that represent the same institution (as illustrated in Fig. 2). The accuracy of our approach and the difficulty in such a disambiguation is shown in Table 11, where the commonest names associated with the NIH are shown (and correctly linked to one another). While many have a similar structure between them (i.e., text indicating US government, followed by text indicating department), there is significant heterogeneity in the names that would make simple string matching difficult. Linking the name ‘National Institutes of Health’ to the other names in this list would likely be unfeasible unless a high-resolution geolocation approach were used, as the only word shared between the names is ‘health,’ commonly occurring in the region. We note that there are also a large number of other government offices within 20 km having a similar naming structure to the NIH (and to each other), and our algorithm correctly distinguishes between them.

    Table 11: Names appearing on at least 100 patents that are aggregated into a single entity for the NIH in Bethesda MD using our algorithm.

    Government labs and universities tend to have a relatively simple structure, operating in a single region. Multinational corporations have a more complex and non-regional structure, and it is worth emphasizing what our disambiguation is actually identifying in the case of companies. A disambiguation of assignees for heterogeneous entities like IBM can potentially be performed on two levels: a disambiguation of the individual elements on a regional level, or a disambiguation of the full legal structure of the entity regardless of proximity. Our algorithm performs the former: we have have disambiguated extremely heterogeneous names that refer to the same regional entity, with global coverage. Linking these regional entities under their correct global corporate umbrella is a valuable exercise23, but not attempted here. Further work can be employed using this local disambiguation coupled with other databases, including ownership databases or wikipedia, to link regional entities to their corporate owners. Examples of the meaning of the disambiguated assignees for two major corporations (IBM and Nokia Siemens) are presented in Table 12, and highlight the fact that the individual disambiguated entities are meaningful, referring to different research labs or production facilities.

    Table 12: The five most prolific assignees containing the string ‘international business machines’ (top) and either ‘nokia’ or ‘siemens’ (bottom) in our disambiguation.

    The assignee name ‘International Business Machines Corporation’ is the second most commonly occurring in the patent database, and while it is almost certainly the case that each of these patents were ultimately developed under the direction of IBM, the location where the work was actually done may vary. IBM’s main offices and labs are in Armonk NY, but semiconductor design was worked on in Fishkill NY and its first manufacturing plant was in Endicott NY. Table 12 highlights the fact that, rather than focused on a single legal entity regardless of geography, our algorithm identifies productive sub-units of those legal entities. Rather than aggregating each ‘international business machines’ lab into a single corporate entity, our algorithm is identifying the regional players, agnostic to the corporate structure that may be directing the research activities of those labs. We note the algorithm is note immune from errors: HA5998 is not aggregated into HA1, the main IBM ID, due to a spelling error in the city. The address fields for HA5998 read ‘Armond NY’ rather than ‘Armonk NY’, which yahoo’s geolocation placed more than 20 km distant. This erroneous disambiguation represents a splitting error well below 1%, showing the algorithm is still quite robust.

    Nokia Siemens has the second most prolific regional lab in the world (using our algorithm), with the main lab in Munich having over 95% of its patents under the original name of Siemens AG. Importantly, our algorithm is able to correctly link patents produced at this location after the merger of Nokia and Siemens: every occurrence of the name ‘Nokia Siemens Networks GMBH’ is also (correctly) linked to the Nokia Siemens disambiguated ID HA2. Likewise, the names associated with HA15, the main Nokia Siemens lab in Helsinki, are composed of ~90% ‘Nokia Corp’ and ~10% ‘Nokia Siemens Networks OY.’ This implies that patents are linked to the disambiguated regional institution both before and after the two companies merged. Abrupt name changes in a company (as may happen due to a merger) will prevent our algorithm from preserving the links between that institution’s patents after the change. Drastic name change events can only be fully accounted for using external data sources that track these legal events.

    Additional Information

    How to cite this article: Morrison, G. et al. Disambiguation of patent inventors and assignees using high-resolution geolocation data. Sci. Data 4:170064 doi: 10.1038/sdata.2017.64 (2017).

    Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.


    1. 1.

      Patent statistics as economic indicators: a survey. Tech. Rep. (National Bureau of Economic Research, 1990).

    2. 2.

      & Knowledge networks as channels and conduits: The effects of spillovers in the boston biotechnology community. Org Sci 15, 5 (2004).

    3. 3.

      & Knowledge flows and the absorptive capacity of regions. Research Policy 44, 4 (2015).

    4. 4.

      , , & Border sensitive centrality in global patent citation networks. Journal of Complex Networks 2, 518–536 (2014).

    5. 5.

      , & . Noncompetes and inventor mobility: Specialists, stars, and the Michigan experiment (Division of Research, Harvard Business School, 2007).

    6. 6.

      , & Foreign born scientists: Mobility patterns for sixteen countries. Nature Biotech 30, 1250 (2012).

    7. 7.

      et al. Career on the move: Geography, stratification, and scientific impact. Sci Rep 4, 4770 (2014).

    8. 8.

      et al. Is europe evolving toward an integrated research area? Science 339, 650 (2013).

    9. 9.

      , , & Public research and industrial innovation: A comparison of us and european innovation systems in the life sciences. Science and Innovation. Rethinking the Rationales for Funding and Governance. Cheltenham: Edward Elgar 169–201 (2003).

    10. 10.

      , , , & The evolution of networks of innovators within and across borders: Evidence from patent data. Research Policy 44, 651–668 (2015).

    11. 11.

      , & The increasing dominance of teams in production of knowledge. Science 316, 1036 (2007).

    12. 12.

      , , & Invention as a combinatorial process: evidence from us patents. J of The Royal Soc Interface 12, 20150272 (2015).

    13. 13.

      & Author name disambiguation in medline. ACM Trans Knowl Discov Data 3, 11 (2009).

    14. 14.

      & Bibliometric fingerprints: name disambiguation based on approximate structure equivalence of cognitive maps. Scienometrics 84, 763 (2010).

    15. 15.

      et al. Disambiguation and co-authorship networks of the us patent inventor database (1975-2010). Research Policy 43, 941 (2014).

    16. 16.

      , & Efficient supervised and semi-supervised approaches for affiliations disambiguation. Scientometrics 97, 47 (2013).

    17. 17.

      , & How to kill inventors: testing the massacrator© algorithm for inventor disambiguation. Scientometrics 101, 477–504 (2014).

    18. 18.

      Patentsview inventor disambiguation workshop. US Patent and Trademark Office (2015).

    19. 19.

      , & in Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1, 379-388 (2012).

    20. 20.

      , & in Proceedings of the 30th International Conference on Machine Learning (2013).

    21. 21.

      & How to play the ‘names game’: Patent retrieval comparing different heuristics. Res Pol 38, 1617 (2009).

    22. 22.

      , & Seeing the non-stars:(some) sources of bias in past disambiguation approaches and a new public tool leveraging labeled records. Res Pol 44, 1672 (2015).

    23. 23.

      OECD HAN database (July, 2014).

    24. 24.

      & Flows of knowledge from universities and federal laboratories: Modeling the flow of patent citations over time and across institutional and geographical ‰boundaries. Proc Natl Acad Sci 93, 12671 (1996).

    25. 25.

      & International knowledge flows: Evidence from patent citations. Econ Innov and New Tech 8, 105 (1999).

    26. 26.

      , , & Identifying geographic clusters: A network analytic approach. Research Policy 44, 1749–1762 (2015).

    27. 27.

      OECD REGPAT database (July, 2014).

    28. 28.

      OECD Citation database (July, 2014).

    29. 29.

      OECD triadic patents families database (July, 2014).

    30. 30.

      , , , & Two supervised learning approaches for name disambiguation in author citations. In Proceedings of the 2004 Joint ACM/IEEE Conference on Digital Libraries, 2004, 296 (2004).

    31. 31.

      & Bibliometric fingerprints: name disambiguation based on approximate structure equivalence of cognitive maps. Scientometrics 84, 763 (2010).

    32. 32.

      et al. On co-authorship for author disambiguation. Inf. Process. and Manag 45, 84 (2009).

    33. 33.

      , , & Citation-based bootstrapping for large-scale author disambiguation. J Am Soc Info Sci and Tech 63, 1030 (2012).

    34. 34.

      et al. Automated disambiguation of us patent grants and applications (2015).

    35. 35.

      , & The nber patent citation data file: Lessons, insights and methodological tools Tech. Rep. (National Bureau of Economic Research, 2001).

    36. 36.

      & The growing distance between people and jobs in metropolitan America (2015).

    37. 37.

      Concept of a technology classification for country comparisons. Final Report to the World Intellectial Property Office (WIPO) (2008).

    38. 38.

      Quantifying the impact of weak, strong, and super ties in scientific careers. Proc Natl Acad Sci 112, E4671 (2015).

    39. 39.

      , , & A comparison of us and european university-industry relations in the life sciences. Management science 48, 24–43 (2002).

    Download references

    Data Citations

    1. 1.

      Morrison, G., Riccaboni, M. & Pammolli, P. Figshare (2017).


    We acknowledge support from the Crisis Lab Project (MIUR, Italy) and from the CERM Foundation (Switch Project). We are also grateful for the assistance of Anna Horodok on the manual disambiguation and for many useful conversations with Orion Penner.

    Author information


    1. Department of Physics, The University of Houston, Houston, Texas, USA

      • Greg Morrison
    2. IMT Institute for Advanced Studies, Lucca, Italy

      • Greg Morrison
      • , Massimo Riccaboni
      •  & Fabio Pammolli
    3. Department of Managerial Economics, Strategy and Innovation, K.U. Leuven, Leuven, Belgium

      • Massimo Riccaboni
    4. Politecnico di Milano, Milan, Italy

      • Fabio Pammolli


    1. Search for Greg Morrison in:

    2. Search for Massimo Riccaboni in:

    3. Search for Fabio Pammolli in:


    All authors contributed to the design of the method and reviewed the paper. G.M. performed the calculations, generated the figures, organized the data, and wrote the paper.

    Competing interests

    The authors declare no competing financial interests.

    Corresponding author

    Correspondence to Fabio Pammolli.

    Supplementary information

    Creative CommonsThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit The Creative Commons Public Domain Dedication waiver applies to the metadata files made available in this article.