skip to main content
research-article
Open access

Analyzing the Direct and Transitive Impact of Vulnerabilities onto Different Artifact Repositories

Published: 07 February 2022 Publication History

Abstract

In modern-day software development, a vast amount of public software libraries enable the reuse of existing implementations for reoccurring tasks and common problems. While this practice does yield significant benefits in productivity, it also puts an increasing amount of responsibility on library maintainers. If a security flaw is contained in a library release, then it may directly affect thousands of applications that are depending on it. Given the fact that libraries are often interconnected, meaning they are depending on other libraries for certain sub-tasks, the impact of a single vulnerability may be large, and is hard to quantify. Recent studies have shown that developers in fact struggle with upgrading vulnerable dependencies, despite ever-increasing support by automated tools, which are often publicly available. With our work, we aim to improve on this situation by providing an in-depth analysis on how developers handle vulnerability patches and dependency upgrades. To do so, we contribute a miner for artifact dependency graphs supporting different programming platforms, which annotates the graph with vulnerability information. We execute our application and generate a data set for the artifact repositories Maven Central, NuGet.org, and the NPM Registry, with the resulting graph being stored in a Neo4j graph database. Afterwards, we conduct an extensive analysis of our data, which is aimed at understanding the impact of vulnerabilities for the three different repositories. Finally, we summarize the resulting risks and derive possible mitigation strategies for library maintainers and software developers based on our findings. We found that NuGet.org, the smallest artifact repository in our sample, is subject to fewer security concerns than Maven Central or the NPM Registry. However, for all repositories, we found that vulnerabilities may influence libraries via long transitive dependency chains and that a vulnerability in a single library may affect thousands of other libraries transitively.

1 Introduction

Software reuse has proven to be an essential part of software development [5, 18]. A vast amount of public repositories exist, where developers can choose open-source software libraries for a variety of different tasks and programming environments. Various build systems like NPM [19], Nuget [25], or Maven [14] conveniently integrate the management of external components into development workflows, thus allowing developers to browse and add components by the click of a button.
As a result of the continuous improvement in tooling support, the amount of open-source library code has been shown to be more than \(45\%\) of the total project code for open-source Java projects, with only \(10\%\) of projects not reusing code at all [18]. Similarly, the 2021 Synposis Open Source Security Report finds that of 1,546 commercial codebases analyzed in 2020, \(98\%\) contained open-source library code, with the average application being comprised of \(75\%\) open source [42].
It has been established by W. C. Lim and others, that software reuse has several positive effects on a software product and its development process, including increased productivity and shorter time-to-market [22]. By May 2021, almost 7 million artifacts are available on Maven Central, which is only one of many repositories for open-source Java libraries [28]. A similar trend in the availability of reusable software artifacts can be observed in other repositories as well, with examples like the NPM Registry and NuGet.org hosting over 16 million artifacts combined (cf. Section 4.3).
However, using external open-source libraries from repositories like NuGet.org also introduces some distinct disadvantages. Most prominently, those libraries impose a new level of complexity on a project. Many of them have dependencies to other libraries of the same repository, thus creating a potentially large dependency graph (often called dependency tree), containing both the direct and transitive dependencies of a project. Decan et al. have shown in 2019 that these graphs can exhibit depths larger than six, with roughly 25% of all artifacts having a dependency graph of depth at least three. The authors also observe that the ratio of transitive to direct dependencies roughly equals 15 for both the NPM Registry and NuGet.org [11]. Overseeing the different responsibilities, release dates and available versions of all libraries in that graph is a complex task that is likely to add additional overhead to the development process [32].
This fact becomes especially relevant when considering vulnerabilities in software libraries. A single vulnerability in a certain library may break the security of all libraries and projects using it, be it directly or transitively. While vulnerabilities being present in a dependency does not necessarily mean vulnerable code is being used [6], Lawrence and Frohoff have shown that in Java vulnerabilities may be exploited simply because the corresponding code is on the classpath [21]. By May 2021, one single release of a popular library like jackson-databind is directly used by more than 18,000 other libraries on Maven Central alone [29], which underlines the impact a single vulnerability may have onto the whole repository.
To minimize the risk of developers not knowing about vulnerabilities in their library dependencies, those vulnerabilities are often disclosed using the CVE standard. The list of all known CVE vulnerabilities is publicly available [9] and frequently updated. The list has been started in 1999 and has, as of May 2021, more than 153,000 entries, averaging at more than 19 vulnerabilities per day. While those entries are not solely related to software libraries but also applications, these numbers still indicate that fixing vulnerabilities in software libraries is not a rare case, but something that must be anticipated during the development process.
In general, the process of safely dealing with a newly published vulnerability is threefold:
(1)
The maintainers of the affected library publish a new version that does not contain the vulnerability anymore. We call this new version a patch for the vulnerability.
(2)
The maintainers of all libraries directly using the affected library need to publish new versions that upgrade their dependencies to use the patch instead of the vulnerable version of the library. This process needs to be iterated recursively to account for transitive dependencies.
(3)
Similar to library maintainers, project developers need to upgrade their dependencies for any of the (transitively) affected libraries to not use vulnerable versions anymore.
While in theory, this process is adequate to deal with the problem at hand, in reality, there are several issues that negatively influence its applicability. Most importantly, it requires the developers of all libraries to be actively maintaining their code. Moreover, they have to constantly monitor all published vulnerabilities and patches, and correlate that data with the dependency graph of their library. While floating version references (i.e., dependencies with a range of valid version numbers) may in some cases enable maintainers to automatically benefit from vulnerability patches, they still need to verify that the vulnerability does not apply anymore.
As the act of monitoring vulnerability publications for every artifact of a dependency graph is both cumbersome and repetitive, to this day several tools have been developed in an effort to automate this task. A prominent example is Snyk, which can be added to software repositories as a CI check that informs developers of vulnerabilities in the project’s dependencies, as well as their severity and possible patches. Going one step further, Dependabot automatically upgrades vulnerable dependencies and creates a corresponding pull request.
Despite the availability of tools like Snyk and Dependabot addressing the problem of vulnerable dependencies, a 2017 study performed by Kula et al. found that \(69\%\) of developers claim to be unaware of being affected. As a result, more than \(81\%\) of the systems they analyzed contained vulnerable dependencies [20]. In addition to that, some library maintainers may never release a patched version, as they decided to stop maintaining their library. This information is not easy to obtain, as there is no standard way of publishing it. Pashchenko et al. [32] argue that such halted dependencies may be identified by using a heuristic based on the average interval between different releases.
In their study, Kula et al. conclude that one reason for this is the developer’s perception of security patches and dependency upgrades as added responsibility yielding little to no benefits for the user [20]. While the act of upgrading a dependency does not consume much time, it requires careful consideration to account for compatibility, performance, and security. Therefore, a good understanding of the respective repository and its patching behavior is key to making informed decisions for dependency upgrades, thus ultimately improving application security.
However, the question of how and when developers release patches and upgrade dependencies in different artifact repositories has yet to be answered. This is due to a lack of empirical data on the subject, especially regarding the differences between repositories. To this day, there is no vulnerability-enriched dependency graph available for any of the major artifact repositories, although that data is freely accessible online. As a result, no formal analysis has yet been conducted on patching behavior in the presence of multi-level dependency graphs.
The main goal of our work is to provide detailed insights into the problem of vulnerable software artifacts and their dependencies by conducting an in-depth analysis that investigates the patching behavior across different repositories. Based on that, we derive development strategies that increase application security. In summary, we contribute:
(1)
A distributed application for mining the dependency graphs of different software repositories and annotating it with vulnerability information. Our implementation is presented in Section 4 and published on Zenodo [12].
(2)
An analysis on the patching behavior of library developers in NuGet.org, the NPM Registry, and Maven Central. We highlight our analysis design, methodology, and results in Section 5.
(3)
A set of resulting threats to application security and possible mitigation strategies, which are discussed in Section 6.

2 Background

Build Systems like Maven or NPM resolve dependency specifications in so-called software artifact repositories. Popular examples for such repositories are Maven Central with almost 7 million artifacts by May 2021 [28], the NPM Registry (around 13.5 million artifacts, cf. Figure 4), and NuGet.org (around 2.8 million artifacts) [26]. These repositories generally expose a web-based interface granting full read-access to any client that wants to consume the contained artifacts. For Maven Central this interface is file-based,1 while for other examples like NuGet.org2 and the NPM Registry3 it is a JSON-based artifact index.
When developers reference an artifact in their project, that artifact may itself be depending on other artifacts of the same repository, thus introducing transitive dependencies to the project. Consequently, each project has a dedicated dependency graph, which may span multiple levels and is resolved by the build system. Developers do not necessarily know about the transitive dependencies in their project, as they are typically resolved transparently. However, commands like mvn dependency:tree4 may be used to explicitly list the full dependency graph.
There are two ways of referencing software artifacts in build systems for the context of this work. The first one, which is used predominantly by Maven, is to specify the unique library identifier and the version number of the target artifact. We call this a fixed artifact reference, as it references exactly one target artifact. While Maven also supports the use of version ranges [30], we find that these are rarely used in any of the artifacts online. However, floating artifact references, which are used in NPM and NuGet, consist of the unique library identifier and a range of valid version numbers. For this type of reference, the build system selects one of the valid target artifacts, which usually is the most recent release of the respective library.
Vulnerabilities in software artifacts can be discovered both unintentionally, e.g., by developers using the artifact, or intentionally, e.g., by users reviewing the source code of an artifact. Frei et al. illustrate the different options a discoverer has upon finding a vulnerability [15]. These include selling it on the black or white market, creating an exploit or a patch, or disclosing it to the public. For the context of this work, the last option is of special interest.
Launched by the MITRE Corporation in 1999, the CVE list [9] is an online database of such publicly disclosed vulnerabilities, and has since emerged as a de facto standard for identifying and referencing vulnerabilities [15]. For each vulnerability, it holds a unique identifier, a description, and a public reference. The NVD was created in 2005 and builds upon the CVE list by providing additional information on fixes, severity, and impact of each CVE entry [8].
Several tools incorporate those data sources to inform developers about vulnerabilities. Prominent examples include the Maven plugin Dependency-Check Maven [38] and Dependabot.5 The services provided by Snyk6 rely on a proprietary vulnerability database, which incorporates not only vulnerabilities published by the NVD but also findings from academic collaborations and proprietary research, as well as vulnerabilities reported by the community. According to Snyk, each finding is manually tested for accuracy and enriched with additional metadata, including an explicit link to the set of affected software artifacts [24]. Furthermore, for vulnerabilities not imported from public sources like the NVD, Snyk distinguishes the date of disclosure and date of publication. Upon disclosure, Snyk assess the severity and risks of vulnerabilities and, following an explicit Disclosure Policy, contacts the maintainers of the affected modules. After that, Snyk cooperates with the module maintainers and publishes the vulnerability information following a public disclosure timeline. Finally, as Snyk is officially registered as CVE Central Naming Authority, it assigns a CVE identifier for the vulnerability [23].

3 Approach

Our approach incorporates three different phases. First, we aggregate data on software artifact dependencies for different repositories and annotate vulnerability information. We then define research questions and answer them using the data produced in the previous phase. With our research questions, we investigate how and when library maintainers upgrade a directly or transitively vulnerable release, and how many artifacts are affected by vulnerabilities in the first place. Finally, we discuss the results, identify potential security threats and derive mitigation strategies.
For data aggregation, we design and implement a distributed application called Miner. This application builds the artifact dependency graph for different repositories, which is done by querying the respective web interfaces. In a second step, the Miner utilizes a customized dump of the Snyk database to annotate the dependency graph with vulnerability information. The graph is then stored in a database for further processing. We present the design of our Miner in Section 4.1, highlight implementation details in Section 4.2, and present insights into the resulting data set in Section 4.3.
In the second phase, we conduct our analysis, which is guided by four different research questions. We present those questions and define their scope in Section 5.1. To answer those questions, we implement different analysis applications that extract metrics from the annotated dependency graph, and also implement tools for manual data exploration. We provide the results for each analysis in Section 5.3, which includes diagrams and plots of the previously generated metrics, as well as the outcomes of our manual analyses.
Finally, we discuss the results of our analysis in Section 6. We systematically derive potential threats to application security from our findings and analyze which repositories are affected by them. We then propose possible mitigation strategies for those threats, which are meant to help software developers improve the security of their development process.

4 Data Aggregation

To aggregate the data set that we perform our analysis on, we developed a distributed application called Miner. It extracts the artifact dependency graph from different repositories, annotates it with vulnerability information and stores it in a persistent fashion for further processing. In this chapter, we present the design and implementation of said application.

4.1 Design

In this section, we present the core design decision upon which we based application development. We aimed to include all relevant properties that might increase analysis expressiveness in the final data model.

Selecting Repositories.

While we designed our application with extensibility in mind, we selected an initial set of artifact repositories to identify a superset of relevant artifact properties. We decided to include repositories for three programming platforms: the JVM, .NET, and JavaScript. To ensure the relevance of our work, we consider only the most popular repository for each platform, yielding Maven Central (JVM), NuGet.org (.NET), and the NPM Registry (JavaScript).

Defining a Suitable Data Model.

The creation of a suitable data model for the resulting data set involves defining properties for artifacts, vulnerabilities, and the relation of those two entities.
We define our data model for software artifacts by inspecting the data provided by the three selected repositories. Based on that, we identify common properties and evaluate their usefulness in the context of our work.
While named differently, all three repositories incorporate the concept of a unique library identifier, which is always accompanied by a corresponding version number. In Maven Central, the library identifier is composed of two attributes called groupId and artifactId, for NuGet.org and the NPM Registry it is a simple attribute called id and name, respectively. The concatenation of library identifier and version number yields a unique artifact identifier. Also, all three repositories expose the publication date of an artifact.
Naturally, a dependency graph cannot be constructed without information about an artifact’s dependencies. We represent each dependency as a tuple, containing both a library identifier and a domain-specific expression encoding a range of valid version numbers. As a result, the NPM dependency ”lodash”:”⌃4.17.4” references all artifacts of library lodash with a version number greater than or equal to 4.17.4 up until the next major version 5.0.0, excluding. For our data model, we introduce an entity that holds three properties: the target’s repository name and library identifier, as well as a repository-specific version range specifier. For the sake of traceability, we keep the original definition of the range specifier in our datamodel, and only normalize the data when resolving references in a later stage (cf. Section 4.2).
So in summary, the set of properties for software artifacts in our data model contains a unique library identifier, a version number, the release date, and a list of artifact dependencies.
Our source of information on software vulnerabilities is a customized data dump of the Snyk Vulnerability DB, which was provided to us by Snyk Ltd. in an effort to support our research, and is up-to-date as of 16 July 2020. Based on this data source, we define our data model for vulnerabilities, which includes the vulnerability’s internal Snyk identifier, its severity (Critical, High, Medium or Low) and a reference to an associated CVE identifier, if available. Furthermore, we incorporate the date of disclosure and date of publication. Last, vulnerabilities must be linked to the software artifacts that they affect. In Snyk, this is done by providing a unique library identifier and domain-specific version range specifier, for which they use a syntax that is different in all three repositories. Similar to artifact dependencies, we keep the domain-specific range specifier in our data model, and normalize the data when it is processed.
We summarize our data model in Figure 1. It features a single entity called ArtifactReference, which is used to model both artifact dependencies and vulnerability-to-artifact relations. It has an associated current target, which is the referenced artifact with the highest version number. Furthermore, we identified additional relations that are required to perform a meaningful analysis: A Patching Relation connects each vulnerability to zero or more artifacts that incorporate the respective patch. Furthermore, there are two distinct relations ordering releases for each unique library, one of them considering the time of release (next release), the other one the version number (next version).
Fig. 1.
Fig. 1. UML description of the overall data model.
Figure 2 illustrates the data model by showing a sample instantiation for a vulnerability that is associated to CVE-2020-110237 and affects the jQuery library in NuGet.org. For the sake of readability, only five entities of type SoftwareArtifact are shown here; in reality, there are 58 artifacts affected by this vulnerability.
Fig. 2.
Fig. 2. Example instance of the data model shown in Figure 1.

Defining Components.

To enable a modular and distributed implementation of our miner, we decompose the task of data aggregation into sub-tasks, each of which is addressed by a different component.
The Database component is responsible for storing the data set in a persistent fashion. It exposes an interface that allows structured querying of the data, which is required for the subsequent analysis phase. The Artifact Miners enumerate all software artifacts for their respective repository and store the results in the database using the data model defined in Figure 1. As the interfaces of all three repositories are different from each other, there is one dedicated miner implementation per repository.
The task of parsing the list of vulnerabilities and transforming them to the aforementioned data model is handled by the Vulnerability Processor. Finally, the Reference Resolver enumerates all ArtifactReference entities in the database, which implicitly reference a set of artifacts, and converts them to explicit entity-to-entity relations. This is done so that the subsequent analysis does not need to re-resolve those implicit relations. Using those explicit relations, the Patch Detector then enumerates all vulnerabilities in the data set and finds those artifacts that patch the vulnerability, i.e., the first artifacts of the targeted library that are not affected by the vulnerability anymore. The vulnerability and its patching artifact are connected via an entity-to-entity relation in the database.

4.2 Implementation

As previously mentioned, we design a distributed application for the task of aggregating our data set. We separately implement and package each component, and manage their deployment using Docker and Portainer. For our implementations, we use Python 3.8, and our Docker images are based on the python-alpine image in version 3.8.2-alpine3.11.
We publish our miners and the respective components for postprocessing as open-source Docker images via GitHub.8 Due to our data on vulnerabilities being the intellectual property of Snyk Ltd., we cannot publish any components handling this data. However, they can be obtained directly from Snyk Ltd.
In the following, we highlight the most important implementation details, the challenges we faced, and the tools we utilized.

Database.

As our data set is essentially a large dependency graph with annotated vulnerability nodes, we select the graph database Neo4j as our storage backend. We deploy it using the official Docker image9 in version 4.0.5.

Miners.

For each repository, we develop and implement a dedicated miner component. For Maven Central, we initially relied on an existing implementation by Benellalam et al. [4], which also stores a full dependency graph using Neo4j. However, due to multiple problems we faced when executing this implementation (cf. Section 4.3), we had to re-implement the Maven miner from scratch.
All our implementations follow an asynchronous Producer-Consumer-Architecture: a producer enumerates artifacts, which are grouped into batches and piped to a pool of consumer threads. Those consumers store all artifacts in the database. By decoupling those two processes, periods of low storage throughput do not slow down the artifact enumeration process.
For NuGet.org, artifacts are enumerated by consuming an append-only catalog that keeps track of all types of artifact-related events. For each publication event, the miner retrieves corresponding artifact properties by issuing a single HTTP request to the NuGet.org API. The catalog consists of multiple pages, for which the index is located at api.nuget.org/v3/catalog0/index.json.
In contrast to that, the NPM Registry provides a single index file at replicate.npmjs.com/_all_docs, which contains identifiers for all libraries ever created. Our NPM Miner utilizes this identifier to retrieve a list of all artifacts belonging to that library at the NPM Registry API.
Finally, for Maven Central, we leverage an Apache Lucene index of all artifact identifiers, which is located at https://repo.maven.apache.org/maven2/.index/ and updated weekly. For each identifier, the miner downloads the corresponding pom.xml file from Maven Central and extracts information about the artifact and its dependencies. However, in contrast to other repositories, Maven’s dependencies cannot always be extracted by looking at a single file. They may be specified in one of the artifact’s parent POM files, or use property values that need to be resolved in the POM file hierarchy. These problems can be addressed by using the Maven command-line interfaces to resolve dependencies, which does, however, take a very long time and uses a significant amount of disk space for POM file caching. Instead, we developed our own dependency resolver using the standard Java XML libraries, which significantly increases the performance of our application. To guarantee correctness, we use a secondary dependency resolver as a fallback: Whenever our XML resolver fails to resolve a property value (which may happen in cases of so-called import-scope dependencies), we re-resolve dependencies using the Eclipse Aether library.10 As this library is solely available for the JVM platform, we implement the Maven miner in Java and deploy it using the official maven Docker image in version 3.6.0-jdk-8-alpine.

Vulnerability Processor.

The vulnerability processor consumes our snapshot of the Snyk Vulnerability Database, which has been provided in the JSON file format and creates corresponding entities of type Vulnerability and ArtifactReference in the database.

Postprocessing.

After the Miners and Vulnerability Processor have aggregated all data necessary, two distinct components perform postprocessing steps to prepare the data for analysis.
The Reference Resolver converts a number of implicit relations between database entities into explicit relations, which enables us to greatly improve the performance of our subsequent analyses and simplifies transitive queries to the database. It first processes all artifacts for each unique library identifier and creates relations of type NEXT and NEXT_RELEASE. These represent the order of versions (according to the Semantic Versioning 2.0.0 specification11) and the order of release dates for all artifacts of the given library, as illustrated in our data model in Figure 1. Afterwards, the Reference Resolver connects all database entities of type ArtifactReference with their target artifacts by creating direct relations of type REFERENCES. This is done by parsing the specific version range qualifier and then filtering all artifacts of the target library for matching versions. The Reference Resolver also creates a special relation called CURRENT_TARGET between each ArtifactReference and its target artifact with the highest version number. The result of a successful run based on a reference to the Nuget.org library jQuery can be seen in Figure 2.
Finally, the Patch Detector builds on top of the transformations performed by the Reference Resolver and detects the set of patches for each vulnerability. As the range of version numbers affected by a vulnerability may be discontinuous, it is possible that there are multiple patches for a single vulnerability. The component incorporates two different approaches: First it checks whether we have explicit data on the patching software artifact, which is true for most of the vulnerabilities. If no such data exists, then the patching artifacts are calculated by finding the first non-affected library releases for which the predecessor is affected by the vulnerability. This process is illustrated in Figure 3, where version numbers highlighted red are affected by a vulnerability, and green version numbers indicate the detected patches. In the database, a vulnerability is connected to each of its patches with a relation of type PATCHED_BY.
Fig. 3.
Fig. 3. Detecting patching library releases.

4.3 Execution Results

We initially executed our miner throughout the course of 70 days. The first phase was dedicated to running the Miners for all three repositories, which took a total of 43 days. The second step of executing the Vulnerability Processor took less than one minute, whereas the postprocessing phase required another 27 days to complete. We identified two main factors with significant impact on the application performance:
Maven Miner. For our Maven Miner component, we initially reused an existing implementation published by Benelallam et al. [4] in 2019. During its execution, we observed a large variation regarding the artifact mining speed. Furthermore, some internal components of the application required constant manual restarts, which lead to additional performance penalties. Due to those problems, we aborted the program execution and instead developed our own Maven Miner implementation from scratch (cf. Section 4.2).
Reference Resolver. The Reference Resolver identifies potential targets of a reference, which involves a lookup via their unique library identifier. For all three repositories, these identifiers are generally case-insensitive. However, in NuGet.org references often use a style of capitalization that differs from the original definition. Therefore, the Reference Resolver is required to do a case-insensitive property lookup in the Neo4j database, which turned out to be significantly slower than an exact-match lookup. For future runs of the NuGet.org Miner, this problem has been averted by converting all package names to lowercase.
Running our own Maven Miner implementation took additional 14 days, including the second phase of postprocessing. Our final data for NuGet.org and the NPM Registry is up-to-date as of June 2020, while the data for Maven, which was calculated at a later stage, is valid as of April 2021. Furthermore, our data on vulnerabilities is up-to-date as of July 2020. To avoid inconsistencies in the data set, we disregard all information on artifacts and vulnerabilities released after June 2020, thus creating a snapshot of that point in time for all three repositories and their associated vulnerabilities.
Our final data set contains a total of 21.8 million software artifacts spread across 1.9 million libraries, and incorporates 7110 vulnerabilities, 27% of which do not have associated CVE identifiers. There is a total of 3736 distinct CVE identifiers referenced in the data set. Figure 4 illustrates how those values are distributed across the three different repositories. The data is stored in two separate Neo4j instances, which in their uncompressed state require a combined disk space of 51 GB.
Fig. 4.
Fig. 4. Number of artifacts and libraries per repository.

5 Analysis

Based on our data set presented as part of Section 4.3, we design and conduct an analysis that characterizes the patching behavior exhibited by maintainers of the different artifact repositories. We aim to identify how much influence software vulnerabilities have, especially considering transitive artifact dependencies. Furthermore, we investigate the time it takes library maintainers to release patches and to upgrade vulnerable dependencies.
In this section, we first present the design of our analysis (Section 5.1), which consists of four different research questions (RQs). After that, we present our methodology in Section 5.2, and finally, we outline our results in Section 5.3.

5.1 Design

With our analysis design, we aim to define the scope of the term patching behavior as precisely as possible. We do so by deriving a set of research questions, each of which involves metrics that contribute to our subjective understanding of the term. Thus, answering those questions yields a characterization of the patching behavior in each of the three repositories. This, in turn, enables us to compare the results across different repositories and to identify characteristic differences regarding the handling of software vulnerabilities.
In total, we answer four distinct research questions, which constitute our analysis. In the following, we present the goal and scope of each question in detail.
RQ1: How long does it take developers to release patches?. We identify the duration between the publication of a vulnerability and the release of an appropriate patch. This value is of critical interest, as it describes a period of time in which a vulnerability is publicly known without developers having an option to mitigate the risk. Also, the duration between vulnerability disclosure and patch release is of interest, as well as a comparison of the two values.
RQ2: What is the transitive impact of vulnerabilities?. We determine the number of artifacts that are affected by a given vulnerability in the first place. While this is rather trivial to calculate for artifacts directly affected by a vulnerability, it is interesting to also identify artifacts that are transitively affected. It is especially relevant to compare these values across different repositories, since the influence of vulnerabilities is expected to vary. The results are likely to help us understand the net impact a vulnerability may have, depending on the respective repository.
RQ3: How long does it take developers to upgrade vulnerable direct dependencies?. We analyze the artifacts directly depending on a vulnerable artifact. Our overall goal is to understand when vulnerable dependencies are fixed by library developers, as compared to the vulnerability’s publication. It is also interesting to see whether or not these upgrade durations depend on the respective library, and how they differ across multiple repositories. The results may help library developers to understand the consequences of vulnerable dependencies, and other developers to carefully select dependencies for their project.
RQ4: Do developers adhere to published security advisories?. We investigate whether or not developers use vulnerable dependencies, for which the corresponding security advisory has already been made public. The results indicate whether or not vulnerability publications are being monitored and considered during development.

5.2 Methodology

We extract different metrics from our data set to answer the RQs defined above. Besides those automated analyses, in some cases, we also require utilities for manual data exploration, which we implement using Python, Plotly,12 and Dash.13
We implement two different modes of operation regarding the detection of dependencies for all metrics that involve transitive relations. In the data set, a dependency is modeled using a node of type ArtifactReference, which is connected to all valid releases of the target library (as seen in Figure 2). Considering all of those targets as possible dependencies vastly increases the search space of the respective analysis, and does not correspond to reality, where a package manager picks exactly one of the valid target releases. Therefore, this mode of operation results in data that is an overapproximation ( \(\mathbb {M}_{\gt }\) ) of reality.
Another option for dependency detection is to select only one of the possible targets for any ArtifactReference. The most reasonable choice is to select the most recent release of the target library, as this is what any package manager does in the absence of version conflicts. Our analyses leverage the CURRENT_TARGET relation annotated by the Reference Resolver to do so. This mode of operation yields an underapproximation ( \(\mathbb {M}_{\lt }\) ) of reality.
All of our analyses that consider transitive dependencies can be executed in both the \(\mathbb {M}_{\gt }\) and \(\mathbb {M}_{\lt }\) mode. However, since package managers always select one target per dependency, and only deviate from the most recent release in the rare case of version conflicts, we consider \(\mathbb {M}_{\lt }\) to yield data that is intuitively closer to reality. Therefore, all results presented in the following Section 5.3 have been obtained using this mode of operation.

5.3 Results

In this section, we present our analysis results for the aforementioned research questions. For each of them, we briefly present the metrics involved and illustrate the results per repository using different types of diagrams.

RQ1: How long does it take developers to release patches?.

For this first RQ, we analyze the time from a vulnerability’s publication or disclosure to the release of an appropriate patch by the developers of the target library. A previous study on the NPM Registry from 2018 [10] suggests that patches are often published before the date of publication, and even before the date of disclosure. We, therefore, expect that most patching durations are negative.
For each vulnerability and patching artifact, our analysis extracts the duration from vulnerability publication to patch release ( \(\delta _{P}\) ) and from vulnerability disclosure to patch release ( \(\delta _{D}\) ) in seconds. This is done by issuing the Cypher query illustrated in Listing 1, and calculating the differences between the datetime values release and dis / pub for each patching artifact. As this process is limited to vulnerabilities that have patches associated to them, a separate database query first retrieves the number of unpatched vulnerabilities for each repository.
Cypher query to extract patching durations.
The bar chart in Figure 5 illustrates the number of unpatched vulnerabilities per repository. In addition to that, the percentage of vulnerabilities with wildcard version range specifiers is shown. These vulnerabilities by definition cannot be patched by any artifact, as they reference the entirety of their target library, for example with a version range qualifier of ”*” or ”[0,]”.
Fig. 5.
Fig. 5. Percentage of unpatched vulnerabilities and wildcardversion range specifiers per repository.
While on average 35.8% of all vulnerabilities are not patched, NPM has by far the highest amount with a percentage of almost 50%. Upon further investigation, we found that this is partially due to so-called Security Holders, which we found to be referenced by 14% of all NPM vulnerabilities. Security Holders are symbolic artifacts that act as a placeholder for library identifiers that have previously been in use and are not usable anymore for security reasons [43]. Figure 5 shows the values for NPM when excluding security holders as NPM WSH (without security holders).
In Figure 6, we illustrate the distribution of severities for those unpatched vulnerabilities. While we initially expected those vulnerabilities to be of minor severity, our data indicates that in fact more than 50% are either of severity high or critical across all three repositories. This is mainly influenced by the NPM repository, in which more than 70% of all unpatched vulnerabilities have a high or critical severity. For Nuget and Maven, we observed a behavior that is more in line with our expectations, as for both of them around 66% of the unpatched vulnerabilities are either of severity low or medium.
Fig. 6.
Fig. 6. Severity of unpatched vulnerabilities.
For the remainder of this section, we only consider vulnerabilities with existing patches for our analysis. In Table 1, we present the mean and median values for \(\delta _D\) and \(\delta _P\) , as well as the respective Interquartile Range(IQR) in days. Regarding \(\delta _P\) , we observe that vulnerabilities in NuGet.org seem to be patched the earliest, followed by Maven Central and then the NPM Registry. For \(\delta _D\) , we observe very similar median values for all three repositories, however, there are large differences in the IQR values, indicating a much more skewed distribution for Maven Central (184.0) as compared to the NPM Registry (25.4) or NuGet.org (3.8).
Table 1.
 MavenNPMNuget
Mean \(\delta _D\) 73.17.9 \(-3.7\)
Median \(\delta _D\) \(-0.6\) 0.80.7
IQR \(\delta _D\) 184.025.43.8
Mean \(\delta _P\) \(-188.2\) \(-159.8\) \(-209.8\)
Median \(\delta _P\) \(-28.2\) \(-7.9\) \(-30.3\)
IQR \(\delta _P\) 311.8143.5395.8
Table 1. Key Measures for \(\delta _{P}\) and \(\delta _{D}\) in Days
In general, in Table 1, we observe rather large IQR values, especially for \(\delta _P\) . To gain more meaningful insights, we plot a stacked histogram of both the \(\delta _{D}\) and \(\delta _{P}\) values in Figure 7. For each repository, the chart indicates the amount of patches released before vulnerability disclosure ( \(\delta _{D} \le 0\) ), between disclosure and publication ( \(\delta _{D} \gt 0 \gt \delta _{P}\) ) and after publication ( \(\delta _{P} \ge 0\) ).
Fig. 7.
Fig. 7. Distribution of \(\delta _{P}\) and \(\delta _{D}\) values per repository.
We observed that, in total, almost 50% of all vulnerabilities are patched before their date of disclosure. This suggests that library developers employ vulnerability detection processes, and manage to significantly reduce the attack surface of their library before any external actor informs them about a particular threat. Furthermore, around 73% of all vulnerabilities are patched before their publication, meaning that without any insider knowledge, an attacker may only be able to exploit the remaining 27% for an attack. In addition to that, our data suggests that an additional 7% of vulnerabilities are patched within the first month after publication.
While this average data does match our general expectation of the patch release date distribution, there are some notable differences between repositories. While for Maven Central over 50% of all vulnerabilities are patched before their disclosure, the corresponding value for NuGet.org is only 33%. However, with 86% being patched before publication, NuGet.org outperforms Maven Central, which is below 75% for this value. This suggests that maintainers of the NuGet.org repository rely on vulnerability disclosure to a larger degree than their colleagues at Maven Central, but are still able to patch a larger amount of vulnerabilities before they are made public.
Over 25% of all vulnerabilities are unpatched, with the majority being of high or critical severity. The main driver for both trends is the NPM Registry repository. Almost 75% of all patches are released before vulnerability publication, on average this happens 184 days prior. NuGet.org patches are released the earliest compared to vulnerability publication.

RQ2: What is the transitive impact of vulnerabilities?.

We inspect the net impact of vulnerabilities by measuring the number of artifacts that are affected, both directly and transitively. As part of this, we also analyze the length of transitive dependency chains that propagate such vulnerabilities.
To do so, we query our data set and for each vulnerability extract the number of unique affected libraries ( \(\mathbb {I}_{Lib,n}\) ) and unique affected artifacts ( \(\mathbb {I}_{Art,n}\) ) per level \(n\) , where the level captures the length of the transitive chain. \(\mathbb {I}_{Art,0}(v)\) therefore corresponds to the number of artifacts directly affected by vulnerability \(v\) , whereas \(\mathbb {I}_{Art,1}(v)\) represents the number of artifacts directly depending on a level zero artifact for the same vulnerability. As described in Section 5.2, the detection of dependencies is executed in an underapproximation mode ( \(\mathbb {M}_{\lt }\) ). Due to performance reasons, we cut off our analysis at a depth of six. Listing 2 presents an example query that is used to obtain values for a single vulnerability in the Maven Central dependency graph, with the cutoff being specified as the recursive multiplicity of the DEPENDS_ON relation in line 2.
Cypher query used to obtain values \(\mathbb {I}_{Art,n}\) and \(\mathbb {I}_{Lib,n}\) for a single vulnerability.
In Figure 8, we illustrate the transitive impact of vulnerabilities. We say a vulnerability \(v\) has impact on level \(n\) if it affects at least one library on that level, i.e., \(\mathbb {I}_{Lib,n}(v) \gt 0\) . We observe that vulnerabilities in the Maven Central Repository have the most impact on the first six levels. Here, more than 70% of all vulnerabilities have an impact on the first level, and roughly 50% have an impact on the sixth level. For the NPM Registry these values are significantly lower. The former one being not more than 60% and the latter just over 15%. Finally, vulnerabilities in NuGet.org exhibit the least impact of all three repositories, merely 3% have an impact on level six. As we cut off our automated analysis at level six, it does not reveal any information about the total length of impact chains in the data set. Therefore, we conducted a structured manual analysis on the most impactful vulnerabilities, which are likely to exhibit the longest impact chains. However, it must be noted that there may be longer chains for less impactful vulnerabilities, therefore our results may only be seen as an indication for the general trend in terms of impact chain length. Table 3 reports the maximum chain length \(\mathbb {L}_{MAX}\) that we observed during manual inspection for each repository. The longest chain is comprised of 38 artifacts and is part of the Maven Central repository. The impact chains for the NPM Registry and NuGet.org proved to be shorter, here we observed maximum lengths of 15 and 13, respectively.
Fig. 8.
Fig. 8. Amount of vulnerabilities with impact per level.
Apart from the length of impact chains, we also investigate the \(\mathbb {I}_{Art,n}\) values for each level \(n\) . To enable a comparison between repositories, we normalize those values per 100,000 total artifacts for each repository. Table 2 presents the mean ( \(\overline{X}\) ) and median ( \(\tilde{X}\) ) number of affected artifacts per level, as well as the corresponding IQR. Interestingly, we observe turning points for \(\overline{X}\) at levels two and four for the NPM Registry and NuGet.org, while the values for Maven Central exhibit continuous growth until level six. This matches both our observation of very long impact chains in Maven Central (cf. Table 3) and the large number of Maven vulnerabilities with impact on level six or higher (cf. Figure 8). However, the direct comparison between Maven Central and NuGet.org reveals a stark contrast in vulnerability impact: While in NuGet.org on average only 2 artifacts per 100,000 are affected by vulnerabilities on level 6, the corresponding values for Maven Central is almost 6,000 times larger.
Table 2.
 Maven CentralNPM RegistryNuGet.org
  \(\overline{X}\) \(\tilde{X}\) IQR \(\overline{X}\) \(\tilde{X}\) IQR \(\overline{X}\) \(\tilde{X}\) IQR
Level 00.50.20.50.30.10.30.80.70.9
Level 188.51.218.426.90.02.26.90.00.0
Level 2587.21.475.835.90.01.08.60.00.0
Level 31,878.01.0190.731.90.00.28.30.00.0
Level 44,200.00.5353.825.80.00.09.30.00.0
Level 57,649.30.2444.620.20.00.04.10.00.0
Level 611,645.90.0634.315.50.00.02.00.00.0
Table 2. Mean, Median and IQC Values for \(\mathbb {I}_{Art,n}\) Per 100,000 Total Artifacts for Each Level and Repository
Table 3.
RepositoryTarget LibraryCVE IdentifierSeverity \(\mathbb {L}_{MAX}\)
Maven Centrallog4j:log4jCVE-2019-17571Critical38
NPM RegistrylodashCVE-2019-1010266Medium15
NuGet.orglog4netCVE-2018-1285High13
Table 3. Vulnerability with Maximum Impact Chain Length Seen During Manual Analysis
Vulnerabilities in the Maven Central repository affect the most artifacts and do so via the longest transitive dependency chains. However, NuGet.org vulnerabilities have the shortest transitive impact chains, and affect less than 0.02% the number of artifacts compared to Maven.

RQ3: How long does it take developers to upgrade vulnerable direct dependencies?.

We analyze how long it takes library maintainers to upgrade a vulnerable direct dependency, as compared to the vulnerability’s publication. We are especially interested in finding out whether the publication of vulnerabilities triggers an immediate response from library maintainers. Since maintainers are generally not involved in the disclosure process, we do not compare the time of upgrades to vulnerability disclosure dates here. Furthermore, we present information on the relative number of affected libraries per repository and investigate the upgrade ratio for such libraries.
For each unique library identifier, our analysis collects the number of total vulnerable dependencies ( \(n_{VD}\) ), as well as the number of vulnerable dependencies that have not been upgraded for this library, which we call unpatched dependencies ( \(n_{UD}\) ). Based on that, we define the patching ratio \(r_{P}\) for each library as \(r_{P} = 1 - \frac{n_{UD}}{n_{VD}}\) . Furthermore, for every library that has at least one patched vulnerable dependency, we extract the mean duration between vulnerability publication and dependency upgrade, \(\delta _{PU}\) , in seconds. This methodology is illustrated in Algorithm 1.
Table 4 summarizes the main characteristics of our results for this RQ. It contains not only the total but also the relative amount of libraries that have ever been affected by direct vulnerable dependencies. Here, we observe significant differences between repositories: While in Maven Central more than one-fourth of all libraries are affected, in NuGet.org this value is just above 1%. In addition to that, Maven’s average \(n_{VD}\) value of 8.7 is notably larger than the values for the NPM Registry (3.4) and NuGet.org (2.8). Despite these differences, the average values for \(r_{P}\) are rather similar for all three repositories, with values ranging from \(32\%\) to \(40\%\) . On average, vulnerable dependencies are patched up to 503 days before vulnerability publication (NPM Registry). Upon investigating the distribution of patching rates, which is illustrated in Figure 9, we made an interesting observation: Library maintainers tend to either upgrade all vulnerable dependencies ( \(r_P=100\%\) ), or none of them ( \(r_P=0\%\) ). We observed partial upgrades for only 17% (Maven Central) to 1% (NuGet.org) of all libraries. The distribution of all \(\delta _{PU}\) values is plotted in Figure 10. As we expected, most values for the NPM Registry are below \(-400\) days, which matches the low average reported in Table 4. However, in Maven Central and NuGet.org between 10% and 13% of all values are larger than 400 days, which indicates that a substantial amount of vulnerable dependencies is upgraded more than a year after the corresponding security advisory was published. Interestingly, both Maven Central and NPM Registry exhibit a slight peak in the \([0; 100)\) interval, which may indicate that the publication of a vulnerability prompts some developers to immediately upgrade their dependencies. Another reason for this observation may be the use of automated tools for vulnerability detection, as mentioned in Section 2.
Vulnerable dependencies affect only 1% of all libraries in NuGet.org, but up to 29% in Maven Central. Upgrades of vulnerable dependencies usually happen more than 200 days prior to vulnerability publication. For some libraries, vulnerability publication seems to trigger an immediate dependency upgrade, suggesting the use of automation.
Fig. 9.
Fig. 9. Distribution of \(r_{P}\) values.
Fig. 10.
Fig. 10. Distribution of \(\delta _{PF}\) values for each repository in days.
Table 4.
Repository# LibrariesRelative # Libraries \(\mathbb {E}[n_{VD}]\) \(\mathbb {E}[r_{P}]\) \(\mathbb {E}[\delta _{PU}]\) in Days
Maven Central99,16228,97%8.738% \(-192\)
NuGet.org3,1441.16%2.840% \(-212\)
NPM Registry102,6667.96%3.432% \(-503\)
Table 4. Key Characteristics of the Analysis Results for RQ3 Per Repository

RQ4: Do developers adhere to published security advisories?.

Once published, a vulnerability and all artifacts affected by it are publicly accessible by any software developer. We inspect whether or not developers adhere to those public security advisories. To do so, we calculate the number of artifacts that use a vulnerable dependency but were released after the vulnerability’s publication.
For each such artifact, we compute the number of vulnerable dependencies, for which the respective vulnerability had already been published at the time of artifact release ( \(n_{V}\) ). Furthermore, for each artifact, we calculate the minimum duration for which those vulnerabilities have been publicly accessible at that time in seconds, which we call \(\delta _{PR}\) . We collect the raw data for this analysis by executing the Cypher query presented in Listing 3.
Cypher query used to obtain data for RQ4.
Similar to RQ3, Table 5 presents the total and the relative number of affected artifacts per repository. Furthermore, the table holds the average values for \(n_{V}\) and \(\delta _{PR}\) . Here, we observe significant differences between repositories. While in NuGet.org only 0.5% of all artifacts are affected, in the NPM Registry this value is more than 32 times higher. In addition to that, affected artifacts in NuGet.org on average have 1.3 vulnerable dependencies with published advisories, compared to 3.3 and 2.8 vulnerable dependencies per artifact in Maven Central and the NPM Registry, respectively. The average duration between vulnerability publication and artifact release ranges from 155 to 470 days, which implies that some maintainers use dependencies known to be vulnerable for well over a year. We further investigate the distribution of those \(\delta _{PR}\) values in Figure 11.
Fig. 11.
Fig. 11. Distribution of durations \(\delta _{PR}\) in months.
Table 5.
Repository# ArtifactsRelative # Artifacts \(\mathbb {E}[n_{V}]\) \(\mathbb {E}[\delta _{PR}]\) in Days
Maven Central807,20015,46%3.3470
NuGet.org15,1190.50%1.3155
NPM Registry2,203,63916.25%2.8431
Table 5. Results of RQ4 Per Repository
The distribution of values in the NuGet.org repository most closely matches our general expectations. Here, more than 90% of all artifacts have been released no longer than one year after vulnerability publication, and only 0.8% after more than two years. In contrast to that, over 20% of all \(\delta _{PR}\) values for both Maven Central and the NPM Registry are greater than two years. Compared to NuGet.org, the values for both repositories are distributed more evenly. This may indicate that in NuGet.org, compared to the other two repositories, security advisories have a more direct influence on dependency upgrades.
In the NPM Registry, 16.25% of all artifacts are released with dependencies for which public security advisories are available. This phenomenon also affects Maven Central (15.46%) but has a lesser impact on NuGet.org (0.5%). On average, affected artifacts have between 1.3 (NuGet.org) and 3.3 (Maven Central) such dependencies, which advisories being published for up to 470 days (Maven Central).

6 Discussion

In this section, we discuss our analysis results and their implications. We start by pointing out security risks that result from our findings and go on to derive possible mitigation strategies that may improve the current situation.

6.1 Security Risks

Based on our analysis results presented in Section 5.3, we identify a set of potential threats for software security. We present this set, which we name \(\mathbb {T}\) , in Table 6. In the following paragraphs, we highlight the core issues of each member of the set \(\mathbb {T}\) , and present empirical evidence found in the analysis results.
Table 6.
IdNameMaven CentralNPM RegistryNuGet.org
\(\mathbb {T}_{1}\) Patch Release after Vulnerability PublicationYesYesYes
\(\mathbb {T}_{2}\) Excessive Transitive Vulnerability ImpactYesNoNo
\(\mathbb {T}_{3}\) Libraries Keep Vulnerable DependenciesYesYesYes
\(\mathbb {T}_{4}\) Publication has no Impact on Upgrade DurationNoNoYes
\(\mathbb {T}_{5}\) New Releases have Vulnerable DependenciesYesYesNo
Table 6. Set of All Threats \(\mathbb {T}\) with Repositories They Apply To
As part of our analysis for RQ1, we found that while most patches are released before vulnerability publication, a non-negligible amount of over 25% is in fact released afterwards ( \(\mathbb {T}_{1}\) ). This implies that for a certain period of time, software libraries were publicly known to be vulnerable, making them a prime target for exploitation. This threat is most pronounced for Maven Central (29%) and the NPM Registry (23%) but also applies to NuGet.org (13%).
The propagation of vulnerabilities along transitive dependency chains leads to an exponentially growing number of artifacts potentially affected and makes it increasingly hard to decide whether vulnerabilities apply to a specific software artifact ( \(\mathbb {T}_{2}\) ). In our results for RQ2, we found that vulnerability impact chains in Maven Central are substantially longer than in the other two repositories. This is true both on average (cf. Figure 8) and at maximum (cf. Table 3). Furthermore, the amount of Maven artifacts transitively affected by a vulnerability is over 750 times higher than for the NPM Registry, and almost 6,000 times higher than for NuGet.org.
As part of our results for RQ3, we found that between 1.16% (NuGet.org) and 28.97% (Maven Central) of all repository libraries have at one point been affected by vulnerable direct dependencies. A further classification discovered that, out of all affected libraries, between 60% and 66% did not upgrade any of those vulnerable dependencies at all (cf. Figure 9). As a result, from 0.7% (NuGet.org) up to 17.4% (Maven Central) of all repository libraries have potentially multiple vulnerable dependencies that have not been upgraded. This fact poses a major security threat ( \(\mathbb {T}_{3}\) ), as consumers of those libraries may not be aware of the vulnerable dependencies, and do not expect the latest release to be vulnerable.
When analyzing the distribution of upgrade durations for RQ3, we found that the publication of a vulnerability seems to trigger an immediate dependency upgrade for some libraries in Maven Central and the NPM Registry (cf. Figure 10). This indicates that some library maintainers may be actively monitoring vulnerability publications, and respond if needed. However, we found no evidence for the same kind of behavior in the NuGet.org repository, meaning that vulnerable dependencies may affect artifacts longer than necessary ( \(\mathbb {T}_{4}\) ).
During the previous analysis for RQ4, we found that some software artifacts use vulnerable dependencies, for which the respective vulnerability has been published before the artifact release. This implies that developers did either not consult or deliberately ignore public security advisories like the CVE list. As a result, consumers of the corresponding repositories can not even rely on new releases of a library to be vulnerability-free, which may lead to a false sense of security when upgrading dependencies to the latest release ( \(\mathbb {T}_{5}\) ). According to Table 5, this issue affects artifacts in all three repositories, but is most pronounced for the NPM Registry (16.25% of all artifacts) and Maven Central (15.46%), while only affecting 0.5% of all artifacts in NuGet.org.

6.2 Mitigation Strategies

As a final contribution of our work, in this section, we present a set of mitigation strategies for maintainers and consumers of all three repositories, which aim to mitigate the threats presented in the previous section. Table 7 presents the set of strategies, which we call \(\mathbb {S}\) , alongside the corresponding threats and the target audience. Most of the strategies presented here are well-known in software engineering research. However, here we provide justification for their usefulness, as we link them to threats that we observed based on empirical findings in our dataset.
Table 7.
IdNameThreatsAudience
\(\mathbb {S}_{1}\) Closely Monitor Vulnerability Publications \(\mathbb {T}_{1}\) , \(\mathbb {T}_{4}\) Maintainers
\(\mathbb {S}_{2}\) Employ Bug Bounty Programs \(\mathbb {T}_{1}\) Maintainers
\(\mathbb {S}_{3}\) Use Tools for Dependency Monitoring \(\mathbb {T}_{2}\) , \(\mathbb {T}_{3}\) , \(\mathbb {T}_{5}\) Maintainers & Consumers
\(\mathbb {S}_{4}\) Check for Vulnerable Dependencies on Release \(\mathbb {T}_{5}\) Maintainers
\(\mathbb {S}_{5}\) Minimize the Number of Dependencies \(\mathbb {T}_{2}\) Maintainer & Consumers
Table 7. Set \(\mathbb {S}\) Containing Mitigation Strategies Derived from \(\mathbb {T}\)
Threat \(\mathbb {T}_{1}\) refers to a substantial amount of library maintainers releasing patches a long time after the corresponding vulnerability has been published. Furthermore, \(\mathbb {T}_{4}\) points out that maintainers of libraries with vulnerable dependencies do not act upon the vulnerability’s publication. As a result, there are artifacts that are subject to publicly known vulnerabilities without a patch being available in all three repositories. An easy way of mitigating this risk is to closely and actively monitor the publication of vulnerabilities ( \(\mathbb {S}_{1}\) ), so that maintainers can immediately upgrade the respective dependency or start implementing a patch, thus reducing the time for which an artifact is vulnerable. There are multiple ways to automatically monitor new vulnerability publications. The NVD provides an RSS feed on the latest vulnerabilities that have been analyzed,14 and the MITRE corporation does the same on their Twitter feed.15 Developers may also be informed about vulnerabilities in dependencies by the use of tools for dependency monitoring, which are regularly updated with the latest set of vulnerabilities. Examples include the services provided by Snyk, as well as the npm audit command.
While adhering to strategy \(\mathbb {S}_{1}\) reduces the time of exposure to published vulnerabilities significantly, it is still likely to take library maintainers a couple of days or weeks to develop a patch and release it. In an effort to further reduce this time, and therefore minimize the risk imposed by threat \(\mathbb {T}_{1}\) , maintainers may encourage members of the open-source community to find and report security vulnerabilities on their own, thus increasing the chances of discovering vulnerabilities before their official publication ( \(\mathbb {S}_{2}\) ). Such initiatives are often called Bug Bounty Programs and are already offered by a variety of corporations and institutions, including Microsoft [27] and the U.S. Pentagon [17]. While for large corporations these programs often include incentives like monetary rewards, smaller groups of library maintainers may still be able to offer some sort of reward, which may even be an honorable mention in the library description. It must be noted that both \(\mathbb {S}_1\) and \(\mathbb {S}_2\) can only be applied for libraries that are still being actively maintained, as opposed to libraries that have been discontinued.
Strategy \(\mathbb {S}_{1}\) already suggested the usage of automated tools for dependency monitoring, which helps library maintainers monitor vulnerability publications. In fact, this strategy also applies to regular repository consumers and has the potential to mitigate several risks identified in the previous section ( \(\mathbb {S}_{3}\) ). As those tools can often be integrated into a CI workflow, they inform library maintainers about vulnerable dependencies on every commit, making it effectively impossible to not be aware of their existence. Therefore, the risks imposed by \(\mathbb {T}_{2}\) , \(\mathbb {T}_{3}\) , and \(\mathbb {T}_{5}\) are reduced substantially. Furthermore, the usage of automated tools for detecting vulnerable dependencies may result in repository consumers preferring more secure libraries for their projects. The more consumers show an obvious interest in security, the more likely library developers are to shift their development focus on security-related aspects.
Due to the larger amount of artifacts being released despite their dependencies being affected by published vulnerabilities, as implied by \(\mathbb {T}_{5}\) , strategy \(\mathbb {S}_{4}\) encourages library maintainers to inspect all their library’s dependencies on every release. By establishing a defined process of doing so, the risk of unintentionally using vulnerable dependencies is effectively eliminated. Ideally, this process is automated by making use of similar tools to the ones discussed for \(\mathbb {S}_{3}\) . While maintainers may still intentionally decide to keep vulnerabilities for a given release, this guideline assures that at least one developer has evaluated the resulting risk and decided that it does not have a critical impact on library consumers.
The impact of threat \(\mathbb {T}_{2}\) is directly correlated to the density of the corresponding dependency graph, and therefore to the number of dependencies per artifact. Consequently, the impact may be reduced by reducing the average number of dependencies for each software library ( \(\mathbb {S}_{5}\) ). A similar argument can be made for repository consumers, where the number of potential security threats is directly linked to the number of project dependencies. While it is a rather simple task to remove redundant dependencies from an application or library, replacing those that are actually used can only be achieved by implementing the desired functionality from scratch. However, this would be in stark contrast to the concept of software reuse, which has been proven to be effective. Nevertheless, especially in the NPM Registry, some of the libraries are small enough to be replaced by a single function definition. Examples include libraries like array-first, which provides a single function for returning the first \(n\) elements of an array, or to-capital-case, which capitalizes a string value. These examples illustrate that in some cases a trade-off between software reuse and potential security risks must be considered. This has also been observed in a 2020 study on the use of trivial packages by Abdalkareem et al. [1], where the authors observe that trivial packages make up between 10.5% and 16% of PyPI and the Npm Registry, despite the fact that up to 72% of such packages to not incorporate any tests. Similarly, Soto-Valero et al. find that many dependencies in public software artifacts are actually unused, and present a tool called DepClean to tackle this issue [41].

7 Threats to Validity

In this section, we present possible threats to the validity of our work, and how we chose to mitigate them as much as possible.

Construct Validity.

In our analyses for RQ2, RQ3, and RQ4, we investigate metrics that involve vulnerabilities affecting (transitive) dependencies of a software artifact. It must be noted, that the mere presence of such a vulnerable dependency does not imply that the vulnerability applies to the artifact. In fact, researchers associated with the FASTEN Project propose to use method-level call graph analysis to decide whether or not vulnerabilities in dependencies actually apply to a project [31]. They argue that analyzing vulnerabilities on the package-level leads to false positives, and therefore wastes development resources, which is also supported by the findings of an empirical study conducted by Zapata et al. in 2018 [13]. However, the findings of this project are still preliminary and have yet to be implemented for entire software ecosystems [6]. However, sometimes vulnerabilities in dependencies can be exploited simply because the corresponding code is on the classpath, as shown by Lawrence and Frohoff [21]. As our analysis handles vulnerabilities on the package level, we note that our findings generally have to be considered an overapproximation, which may be refined by incorporating precise call graph information for each artifact in the future.

Internal Validity.

As mentioned in Section 5.2, our algorithm for dependency resolving supports two modes of execution, yielding either all valid targets (overapproximation) or the most recent valid target (underapproximation). As package managers only refrain from using the most recent target in case of version conflicts, our analysis results have been obtained using the latter. Therefore, there might be some dependency paths that have not been counted toward our results for RQ2, RQ3, and RQ4.
For the threats presented in Section 7, we often assume that not being aware of vulnerabilities in dependencies is a major reason for not performing an upgrade or not releasing a patch. However, there may in fact be other reasons involved in this decision, for example, libraries may not prioritize fixes, or maintainers may decide that vulnerabilities do not apply to their specific usage scenario. Further investigations need to be performed to identify those reasons and their relevance in this context.
We rely on the correctness of the Synk dataset we use throughout the analysis. As the dataset is manually verified in a structured process and includes official data from the NVD, we think we can safely assume the data to be correct.

External Validity.

We used vulnerability data that is the intellectual property of Snyk Ltd. and can, consequently, not be published. Therefore, others can not reproduce our analysis, as we are only able to make our tools for generating dependency graphs publicly available. We chose this approach as the data provided by Snyk both contains more vulnerabilities than sources like the NVD, and is manually verified to be correct. We strive to include open vulnerability sources and repeat our analysis in the future. In the meantime, interested researchs may of course obtain the vulnerability data through the same official channels we used.

Conclusion Validity.

As many of our analysis metrics imply, artifacts from the NuGet.org repository seem to be somehow less affected by vulnerabilities compared to the NPM Registry and Maven Central. While this may be an indication of better processes and awareness by the respective developers, it may also be a result of the size and popularity of the repository. Out of the three repositories observed in this work, NuGet.org has both the least amount of artifacts and the fewest downloads per week, which may lead to fewer vulnerabilities being introduced, and fewer vulnerabilities being discovered. While we normalized our measures to include the size of repositories where applicable, further analysis is necessary to identify the effects of repository size onto the behavior we observed.

8 Related Work

Dependencies between software artifacts for certain software systems have been the topic of a variety of publications. Furthermore, many researchers have analyzed the impact and lifecycle of software vulnerabilities, as well as possible mitigation strategies. In this section, we present the most prominent examples of related work.
In Reference [4], Benelallam et al. present a snapshot of the Maven Central dependency graph as of September 6, 2018. The authors implemented a tool called Maven Miner, which processes artifacts from the Maven Central repository, calculates their dependencies and creates a graph representation that is stored in a Neo4J graph database. For our work, we initially used the Maven Miner to gather an up-to-date version of the dependency graph, but failed to do so in time (cf. Section 4.3).
In 2018, Decan et al. performed an empirical study on the impact of software vulnerabilities in the NPM Registry [10]. Based on 700 vulnerabilities, which the authors manually gathered from Snyk.io, they computed a set of affected package releases using the online service at libraries.io. Among other things, the authors observed an ever-growing number of vulnerabilities and affected package releases over time, with more pronounced rates of growth for medium and high severity vulnerabilities. A similar observation is made for commercial applications by Mike Pittenger in Reference [35]. Decan et al. also analyzed the time it takes package maintainers to discover vulnerabilities, and report that the majority was found in packages older than 28 months. Similar to this work, the authors found that most vulnerabilities are patched before their publication, but there “is still a non-negligible proportion of vulnerabilities that take a long time to be fixed” [10].
Bavota et al. performed an in-depth study on 147 Java projects of the Apache ecosystem in 2015 [3], in which they analyze how and why the dependencies between projects are upgraded, left unchanged, or dropped. The authors evaluate the influence of the client project size, the number of LOC changed in the library project, the number of bugs fixed, and other factors on the upgrade behavior. They find that substantial changes to the library project, especially bug fixes, are adopted earliest, which matched the authors’ observations in developer communications (e.g., via mailing lists). Similarly, a 2021 empirical study conducted by Chinthanet et al. analyzes the adoption of vulnerability-related fixes in NPM projects on GitHub [7]. By inspection of 231 package-side fixing releases, the authors find that such releases consist of up to 86% code unrelated to the actual fix, and that quick availability of fixes does not ensure a fast adoption process by clients. Nevertheless, a 2019 study by Gkortzis et al. finds evidence for an inverse correlation between code reuse and the number of vulnerabilities, which seems to indicate that “[ . . . ] a high reuse ratio is associated with a lower vulnerability density” [16].
A 2017 study by Kula et al. analyzes the dependency upgrade behavior of software developers in detail [20]. In an empirical study, the authors collected 4,600 Java projects that are using Maven dependencies from GitHub. They found that while developers heavily rely on third-party libraries, they do not often upgrade their dependencies and tend to stick to popular old releases. An additional case study revealed that Awareness Mechanisms like Release Announcements and Security Advisories have a mixed influence on developers. The likelihood of upgrading a library dependency is observed to be decreasing with increasing migration effort. A subsequent interviewing process revealed that almost 70% of all developers were unaware of a vulnerable dependency in their project. The other 30% named mostly project-specific reasons for which a dependency upgrade was not performed or not prioritized, e.g., because the developers found that it does not actually apply to the project. This intuition is confirmed in a 2018 study conducted on 60 NPM projects by Zapata et al., which finds that as much as 73.3% of such vulnerable dependencies may not actually apply to a project, concluding that package-level vulnerability analysis leads to significant overapproximation [13].
In 2018, Jukka Ruohonen published an analysis on vulnerabilities in Python packages that target web development [39]. The base data is collected from the Python Package Index (PyPI) and the Safety DB,16 a curated list of vulnerabilities in Python packages. The author argues that, while often used in scientific work, the CVE list of vulnerabilities does often not include smaller libraries or lesser-known exploits. A subsequent analysis of packages related to web development showed that most vulnerabilities in the data set are of mild severity, with Input Validation and XSS being the most common attack vectors. A final conclusion states that a more meaningful analysis would have to take the dependencies between artifacts into account.
Ponta et al. [36] present a novel approach to detect whether or not a software artifact is affected by a vulnerability. Instead of consuming metadata files (e.g., pom.xml), their tool Vulas combines static and dynamic analyses to decide whether or not the actual vulnerable code construct (e.g., a method) is reachable from the analyzed artifact. This way, the number of false positives is reduced and the intuition of vulnerabilities not applying in certain contexts, as reported by Kula et al. in Reference [20], is formalized. In their latest publication [37], Ponta et al. report on the current state of Vulas, and perform a comparative study on 300 enterprise projects. Compared to OWASP Dependency Check(OWASP DC), their tool identified about 1,800 additional true positive findings, whereas almost 89% of 17,000 vulnerabilities only reported by OWASP DC turned out to be false positives.
In their 2016 publication, Alqahtani et al. use semantic web technologies to establish links between sources on software vulnerabilities and source code repositories [2]. They argue that traditionally, information sources on vulnerabilities and project metadata are heterogeneous, which hinders the estimation of vulnerability impact and artifact security.
Shahzad et al. published their study on the life cycles of software vulnerabilities in 2012 [40]. They obtained a sample of 46,310 vulnerabilities from different sources, including the NVD, and present an analysis regarding multiple data dimensions. Based on that, they extract association rules regarding the patching and exploitation behavior, which are then used for detecting patterns in the data set.
In their 2018 publication, Pashchenko et al. argue that existing tools for vulnerable dependency detection often produce false positives, which in turn lead to inefficient development processes, especially in commercial software development [32]. The authors state that dependency scopes are often not evaluated in existing approaches, as well as halted dependencies. The article presents an analysis approach that is able to detect vulnerable dependencies for Maven artifacts while improving on the deficiencies mentioned above. In particular, the authors use heuristics based on the average release time of a library to identify halted dependencies. Instead of automated vulnerability aggregation, the authors rely on manual identification of vulnerable code for each vulnerability. Finally, Pashchenko et al. conduct a study of their approach on 200 open source Maven libraries. They find that 14% of all dependencies in their sample are halted, with 1% of them being affected by known vulnerabilities. They conclude that employing their approach helped to reduce the time developers have to spend on false positives in vulnerable dependency detection.
Building on their previous work, in 2020 Pashchenko et al. present Vuln4Real, a methodology for overcoming the inaccuracies of traditional approaches for vulnerable dependency detection [33]. They implement the methodology for the Maven build system and perform an empirical study on 500 libraries. The authors find that 80% of all vulnerable dependencies can potentially be fixed by a direct dependency upgrade. Furthermore, Vuln4Real decreases the number of false positives by 27%, thus effectively saving developers the time they would otherwise have to spend on analyzing an inaccurate alert.
In 2020, Pashchenko et al. published a qualitative study on the security implications of developer decisions regarding dependency management [34]. The authors interviewed a total of 25 developers from different enterprises to investigate the trade-off between functional and security-related concerns. Their results imply that selecting an appropriate dependency is a complex task, and security concerns are often neglected in these scenarios. Furthermore, the authors observe that patches are more likely to be adopted by developers if they are not bundled with functional changes.

9 Conclusion

Our work presented three core contributions, that mitigate the resulting threats to application security and foster a general understanding of how vulnerability patches and dependency upgrades are performed in Maven Central, NuGet.org, and the NPM Registry.
At first, we designed and implemented a distributed application that aggregates the dependency graphs for artifacts from all three repositories. An additional post-processing step leverages a Snyk data dump to annotate information about CVE software vulnerabilities to the graph. By executing the application, we obtained a vulnerability-enriched dependency graph for Maven Central (as of September 2018), NuGet.org, and the NPM Registry (both up-to-date as of June 2020). It contains a total of 19 million software artifact nodes, which belong to 1.8 million libraries and have 4.9 million unique dependency specifications, with 5,378 vulnerabilities being annotated.
Our second contribution is a detailed analysis of the data that we previously generated. It was conducted based on four different research questions and comprised the creation of applications for data processing as well as interactive data visualizations for manual data exploration.
We found that a substantial amount of vulnerabilities is still unpatched (over 25%), but existing patches are often released long before vulnerability publication. We also observed that vulnerabilities may transitively affect an artifact with a transitive depth of up to 25, making it hard to manually discover a potential security risk. In addition to that, our analysis revealed that up to 20% of all libraries (Maven Central) in a repository have been affected by direct vulnerable dependencies, and 60% of those libraries did not upgrade any of them. Interestingly, while most dependency upgrades happen long before a vulnerability is published, the act of publishing does seem to trigger an immediate dependency upgrade for some maintainers in Maven Central and the NPM Registry, thus indicating that they may be actively responding to security advisories.
Our third and final contribution is a set of observed security threats and mitigation strategies, which we derived from the analysis results described above. These strategies are meant to mitigate the security threats that result from our findings, thus potentially increasing application security.
With each of our contributions, we aim to improve the current state-of-the-art regarding the development of patches and the upgrades of vulnerable dependencies, thus ultimately reducing the number of exploitable security flaws in software applications. Our findings provide the foundation for a common understanding of the topic and indicate how to tackle some of the problems that we pointed out. Our results may be extended by incorporating additional repositories or data sources.

10 Acknowledgments

Special thanks to Snyk Ltd. for sharing their data on vulnerabilities for the context of this research project.

Footnotes

1
repo1.maven.org/maven2.
2
api.nuget.org/v3/catalog0/index.json.
3
skimdb.npmjs.com/registry/_all_docs.
4
maven.apache.org/plugins/maven-dependency-plugin/examples/resolving-conflicts-using-the-dependency-tree.html.
5
dependabot.com/.
6
snyk.io.
8
The repository is located at https://github.com/sse-labs/dependency-graph-miner. An archived version is available via Zenodo at https://doi.org/10.5281/zenodo.5040439.
9
hub.docker.com/_/neo4j.
10
wiki.eclipse.org/Aether.
11
semver.org/.
12
plotly.com/python/.
13
plotly.com/dash/.
14
nvd.nist.gov/vuln/data-feeds#RSS.
15
twitter.com/CVEnew.
16
github.com/pyupio/safety-db.

REFERENCES

[1]
Rabe Abdalkareem, Vinicius Oda, Suhaib Mujahid, and Emad Shihab. 2020. On the impact of using trivial packages: An empirical case study on npm and PyPI. Empir. Softw. Eng. 25, 2 (Mar.2020), 1168–1204.
[2]
Sultan S. Alqahtani, Ellis E. Eghan, and Juergen Rilling. 2016. Tracing known security vulnerabilities in software repositories—A semantic web enabled modeling approach. Sci. Comput. Program. 121 (2016), 153–175. Special Issue on Knowledge-based Software Engineering.
[3]
Gabriele Bavota, Gerardo Canfora, Massimiliano Di Penta, Rocco Oliveto, and Sebastiano Panichella. 2015. How the apache community upgrades dependencies: An evolutionary study. Empir. Softw. Eng. 20, 5 (Oct.2015), 1275–1317.
[4]
Amine Benelallam, Nicolas Harrand, César Soto-Valero, Benoit Baudry, and Olivier Barais. 2019. The maven dependency graph: A temporal graph-based representation of maven central. In Proceedings of the 16th International Conference on Mining Software Repositories (MSR’19). IEEE Press, 344–348.
[5]
B. Boehm. 1999. Managing software productivity and reuse. Computer 32, 9 (1999), 111–113.
[6]
Paolo Boldi and Georgios Gousios. 2021. Fine-grained network analysis for modern software ecosystems. ACM Trans. Internet Technol. 21, 1 (Feb.2021), 1–14.
[7]
Bodin Chinthanet, Raula Gaikovina Kula, Shane McIntosh, Takashi Ishio, Akinori Ihara, and Kenichi Matsumoto. 2021. Lags in the release, adoption, and propagation of npm vulnerability fixes. Empir. Softw. Eng. 26, 3 (Mar2021).
[8]
MITRE Corporation. 2019. CVE and NVD Relationship. Retrieved from https://cve.mitre.org/about/cve_and_nvd_relationship.html.
[9]
MITRE Corporation. 2021. CVE List Home. Retrieved from https://cve.mitre.org/cve/.
[10]
Alexandre Decan, Tom Mens, and Eleni Constantinou. 2018. On the impact of security vulnerabilities in the npm package dependency network. In Proceedings of the 15th International Conference on Mining Software Repositories (MSR’18). Association for Computing Machinery, New York, NY, 181–191.
[11]
Alexandre Decan, Tom Mens, and Philippe Grosjean. 2019. An empirical comparison of dependency network evolution in seven software packaging ecosystems. Empir. Softw. Eng. 24 (Feb.2019), 381–416.
[12]
Johannes Düsing and Ben Hermann. 2021. sse-labs/dependency-graph-miner: Full release for final paper.
[13]
Rodrigo Elizalde Zapata, Raula Gaikovina Kula, Bodin Chinthanet, Takashi Ishio, Kenichi Matsumoto, and Akinori Ihara. 2018. Towards smoother library migrations: A look at vulnerable dependency migrations at function level for npm JavaScript packages. In Proceedings of the IEEE International Conference on Software Maintenance and Evolution (ICSME’18). 559–563.
[14]
Apache Software Foundation. 2020. Maven—Introduction. Retrieved from https://maven.apache.org/what-is-maven.html.
[15]
Stefan Frei, Dominik Schatzmann, Bernhard Plattner, and Brian Trammell. 2010. Modeling the Security Ecosystem—The Dynamics of (In)Security. 79–106.
[16]
Antonios Gkortzis, Daniel Feitosa, and Diomidis Spinellis. 2019. A double-edged sword? Software reuse and potential security vulnerabilities. In Reuse in the Big Data Era, Xin Peng, Apostolos Ampatzoglou, and Tanmay Bhowmik (Eds.). Springer International Publishing, Cham, 187–203.
[17]
Hackerone. 2020. Hack the Pentagon. Retrieved from https://www.hackerone.com/hack-the-pentagon.
[18]
Lars Heinemann, Florian Deissenboeck, Mario Gleirscher, Benjamin Hummel, and Maximilian Irlbeck. 2011. On the extent and nature of software reuse in open source Java projects. In Proceedings of the Top Productivity Through Software Reuse: 12th International Conference on Software Reuse (ICSR’11), vol. 6727, 207–222.
[19]
NPM Inc.2020. About NPM. Retrieved from https://docs.npmjs.com/about-npm.
[20]
Raula Gaikovina Kula, Daniel M. German, Ali Ouni, Takashi Ishio, and Katsuro Inoue. 2018. Do developers update their library dependencies? Empir. Softw. Eng. 23, 1 (Feb.2018), 384–417.
[21]
Gabriel Lawrence and Chris Frohoff. Marshalling Pickles—How Deserializing Objects Can Ruin Your Day. Retrieved from https://www.slideshare.net/frohoff1/appseccali-2015-marshalling-pickles.
[22]
W. C. Lim. 1994. Effects of reuse on quality, productivity, and economics. IEEE Softw. 11, 5 (1994), 23–30.
[23]
[24]
Snyk Limited. 2021. Snyk Intel Vulnerability Database. Retrieved from https://snyk.io/product/vulnerability-database/.
[25]
Microsoft. 2019. An Introduction to NuGet. Retrieved from https://docs.microsoft.com/en-us/nuget/what-is-nuget.
[26]
Microsoft. 2020. Create .NET Apps Faster with NuGet. Retrieved from https://www.nuget.org/.
[27]
Microsoft. 2020. Microsoft Online Services Bounty Program. Retrieved from https://www.microsoft.com/en-us/msrc/bounty-microsoft-cloud?rtc=1.
[28]
MvnRepository. 2021. Central Repository. Retrieved from https://mvnrepository.com/repos/central.
[30]
Oracle. 2015. Understanding Maven Version Numbers. Retrieved from https://docs.oracle.com/middleware/1212/core/MAVEN/maven_version.htm#MAVEN402.
[31]
OW2. 2021. FASTEN Project. Retrieved from https://www.fasten-project.eu/view/Main/.
[32]
Ivan Pashchenko, Henrik Plate, Serena Elisa Ponta, Antonino Sabetta, and Fabio Massacci. 2018. Vulnerable open source dependencies: Counting those that matter. In Proceedings of the 12th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM’18). Association for Computing Machinery, New York, NY, Article 42, 10 pages.
[33]
I. Pashchenko, H. Plate, S. E. Ponta, A. Sabetta, and F. Massacci. 2020. Vuln4Real: A methodology for counting actually vulnerable dependencies. IEEE Trans. Softw. Eng. 48, 5 (2020), 1.
[34]
Ivan Pashchenko, Duc-Ly Vu, and Fabio Massacci. 2020. A qualitative study of dependency management and its security implications. In Proceedings of the ACM SIGSAC Conference on Computer and Communications Security (CCS’20). Association for Computing Machinery, New York, NY, 1513–1531.
[35]
Mike Pittenger. 2016. The state of open source security in commercial applications. Black Duck Open Source Security Analysis. https://sq-software.com/wp-content/uploads/2016/12/2016-12-OS-Security-Analysis.pdf.
[36]
S. E. Ponta, H. Plate, and A. Sabetta. 2018. Beyond metadata: Code-centric and usage-based analysis of known vulnerabilities in open-source software. In Proceedings of the IEEE International Conference on Software Maintenance and Evolution (ICSME’18). 449–460.
[37]
Serena Elisa Ponta, Henrik Plate, and Antonino Sabetta. 2020. Detection, assessment and mitigation of vulnerabilities in open source dependencies. Empir. Softw. Eng. 25, 5 (Sep.2020), 3175–3215.
[38]
Open Web Application Security Project. 2020. Dependency-Check Maven. Retrieved from https://jeremylong.github.io/DependencyCheck/dependency-check-maven/index.html.
[39]
J. Ruohonen. 2018. An empirical analysis of vulnerabilities in python packages for web applications. In Proceedings of the 9th International Workshop on Empirical Software Engineering in Practice (IWESEP’18). 25–30.
[40]
M. Shahzad, M. Z. Shafiq, and A. X. Liu. 2012. A large scale exploratory analysis of software vulnerability life cycles. In Proceedings of the 34th International Conference on Software Engineering (ICSE’12). 771–781.
[41]
César Soto-Valero, Nicolas Harrand, Martin Monperrus, and Benoit Baudry. 2021. A comprehensive study of bloated dependencies in the Maven ecosystem. Empir. Softw. Eng. 26, 3 (Mar.2021).
[42]
Synopsys, Inc.2021. Open Source Security and Risk Analysis Report. Retrieved from https://www.synopsys.com/software-integrity/resources/analyst-reports/open-source-security-risk-analysis.html.
[43]
NPM Security Team. 2016. Security Holding Package. Retrieved from https://www.github.com/npm/security-holder#readme.

Cited By

View all
  • (2024)AROMA: Automatic Reproduction of Maven ArtifactsProceedings of the ACM on Software Engineering10.1145/36437641:FSE(836-858)Online publication date: 12-Jul-2024
  • (2024)VulNet: Towards improving vulnerability management in the Maven ecosystemEmpirical Software Engineering10.1007/s10664-024-10448-629:4Online publication date: 5-Jun-2024
  • (2023)Large Scale Study of Orphan Vulnerabilities in the Software Supply ChainProceedings of the 19th International Conference on Predictive Models and Data Analytics in Software Engineering10.1145/3617555.3617872(22-32)Online publication date: 8-Dec-2023
  • Show More Cited By

Index Terms

  1. Analyzing the Direct and Transitive Impact of Vulnerabilities onto Different Artifact Repositories

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image Digital Threats: Research and Practice
    Digital Threats: Research and Practice  Volume 3, Issue 4
    December 2022
    232 pages
    EISSN:2576-5337
    DOI:10.1145/3572830
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 07 February 2022
    Accepted: 24 June 2021
    Revised: 17 May 2021
    Received: 30 November 2020
    Published in DTRAP Volume 3, Issue 4

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Vulnerability
    2. impact analysis
    3. repository mining
    4. patch detection

    Qualifiers

    • Research-article
    • Refereed

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)962
    • Downloads (Last 6 weeks)100
    Reflects downloads up to 04 Sep 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)AROMA: Automatic Reproduction of Maven ArtifactsProceedings of the ACM on Software Engineering10.1145/36437641:FSE(836-858)Online publication date: 12-Jul-2024
    • (2024)VulNet: Towards improving vulnerability management in the Maven ecosystemEmpirical Software Engineering10.1007/s10664-024-10448-629:4Online publication date: 5-Jun-2024
    • (2023)Large Scale Study of Orphan Vulnerabilities in the Software Supply ChainProceedings of the 19th International Conference on Predictive Models and Data Analytics in Software Engineering10.1145/3617555.3617872(22-32)Online publication date: 8-Dec-2023
    • (2023)Vulnerability Propagation in Package Managers Used in iOS Development2023 IEEE/ACM 10th International Conference on Mobile Software Engineering and Systems (MOBILESoft)10.1109/MOBILSoft59058.2023.00015(60-69)Online publication date: May-2023
    • (2023)UpCy: Safely Updating Outdated DependenciesProceedings of the 45th International Conference on Software Engineering10.1109/ICSE48619.2023.00031(233-244)Online publication date: 14-May-2023
    • (2023)The GitHub Development Workflow Automation EcosystemsSoftware Ecosystems10.1007/978-3-031-36060-2_8(183-214)Online publication date: 26-May-2023

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Get Access

    Login options

    Full Access

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media