Chapter 3 lists cyber hard problems from what might be called the point of view of a “consumer” relying on cybersecurity in a product or service. This provides a tidy taxonomy but does not describe specific issues or problems that, if addressed, would represent significant progress toward solving them. This chapter describes these specific challenges, which can also be seen as the perspective of the “producer” who needs a well-characterized set of independent principles and procedures—technical, policy, and operational—that are prerequisites to addressing the cyber hard problems.
The consumer and producer lists do not map neatly to one another for the following reasons:
Some of the new cyber hard problems, such as the integration of hardware and software into a cyber-physical system (CPS) or artificial intelligence (AI), depend on essentially all of the producer cyber hard problems. However, they bring new important subproblems that are critical and unsolved.
For clarity and brevity, the producer cyber hard problems are described below in terms of concrete functional, operational, new technology, or policy problems. Some producer cyber hard problems may seem to duplicate consumer cyber hard problems. For example, “secure, resilient design” from a consumer’s point of view involves solving many subproblems because it is a characteristic of an entire cyber system and its operation. The “secure design” problem in this section addresses the technical problems involved in designing a more or less fully specified system (development tools, testing tools, design practices, needed workforce competencies, etc.). This also applies to secure composition. Accurate provisioning of data and information may be directly visible to a consumer, but it may be a characteristic of the training data used to produce an AI model, which may be of no direct interest to a consumer.
Some of the producer cyber hard problems can be solved in a rather satisfying manner by principled techniques that often go under the rubric “the science of security.” Examples include complete access and information flow models.1 This sort of solution is the “gold standard” for scientific progress but has only been applied to very carefully described and constrained subproblems.
Functional cyber hard problems deal with the design of secure, interoperable products and infrastructure.
___________________
1 F.B. Schneider, 2012, “Blueprint for a Science of Cybersecurity,” The Next Wave 19(2):47–57, https://www.cs.cornell.edu/fbs/publications/SoS.blueprint.pdf.
The emergence and popularity of cloud computing are due to convenience, efficiency, and (in some cases) cost-saving. However, while almost everyone uses or relies on cloud computing services today, there are few dominant providers, and their operations are unknown to outsiders. The monoculture and opacity result in key challenges to building and operating resilient cloud systems.
Many of the drivers for cloud computing adoption involve cost and convenience, but the trade-off most pertinent for security is control and understanding. While models of shared responsibility usually exist between the customer and provider, they are not always consistent or complete. Multi-tenancy in a provider’s environment can affect the visibility available to each customer when the provider cannot separate logging, backups, or forensic data for each tenant; sometimes the physical location of a provider’s data center is confidential. This is the “isolation” problem in shared resources. Shared hardware increases the risk of lateral movement by an attacker from one customer to another, and the network traffic needed by one customer may require that the provider cannot block some network traffic even if some tenants want them to. Finally, depending on the service, customers may not have visibility into traffic and interactions that happen within the provider’s environment, only the traffic that happens directly between the customer’s instance and the customer’s own location. Business email compromise (BEC) for the purpose of redirecting payments is a big problem and shows no sign of slowing.2 It is often accomplished by adding filtering rules redirecting emails pertinent to payments to a scammer acting as a “man in the middle” who instructs the payments to be redirected. Outsourcing email to “the cloud” where such critical rule changes for a single email user may not be quickly caught is a common problem.
Besides visibility issues, another hard problem for cloud computing is orchestration. It is difficult to find reliable statistics on multi-cloud use that do not come from a single cloud provider, and control policies between providers can be inconsistent. A look at the latest Cloud Controls Matrix3 from the Cloud Security Alliance tells the story: 197 control objectives in 17 domains. Organizations face security challenges regardless of the strategy they embrace. Using one provider can risk a single point of failure,4 and
___________________
2 P. Harr, 2024, “The Weaponization of AI: The New Breeding Ground for BEC Attacks,” Forbes Technology Council, June 14, https://www.forbes.com/councils/forbestechcouncil/2024/06/14/the-weaponization-of-ai-the-new-breeding-ground-for-bec-attacks.
3 K. Rundquist, 2024, “Cloud Security Alliance Announces Implementation Guidelines v2.0 for Cloud Controls Matrix (CCM) in Alignment with Shared Security Responsibility Model,” BusinessWire: Cloud Security Alliance (CSA), June 4, https://www.businesswire.com/news/home/20240604212963/en/Cloud-Security-Alliance-Announces-Implementation-Guidelines-v2.0-for-Cloud-Controls-Matrix-CCM-in-Alignment-with-Shared-Security-Responsibility-Model.
4 Intelligent Transportation Systems Joint Program Office, “ITS Deployment Evaluation,” Department of Transportation, https://www.itskrs.its.dot.gov/2019-l00856, accessed December 5, 2024.
using more than one provider incurs complexity and management costs (as well as the increased systemic risk of any one provider having an outage).
A particular risk with cloud computing is the durability of artifacts; although using only what you need, when you need it can help reduce costs, the practice requires more rigor in managing the life cycle of those instances. For example, the Colorado Department of Transportation fell victim to a ransomware attack in 20185 when a virtual server was not secured properly because it was intended to be temporary, and yet it was connected to the agency’s active directory domain, which allowed the attacker to gain additional privileges. Retention periods for backups and logs are other examples of critical artifact management properties.
The start of the COVID-19 pandemic in 2020 forced more organizations to embrace cloud computing with remote access, which, in turn, drove the development of architecture changes such as “edge computing.” Remote user traffic that had to pass through on-premises infrastructure to access cloud-based resources resulted in network bottlenecks and latency; the Secure Access Service Edge emerged in response, putting the users and resources closer to one another. All this reliance on third-party providers has opened new areas of attack as well as complicated security management;6 because any given end-to-end interaction now involves additional personnel, terms of service, and levels of visibility and control different from a simple governance model.
Assessing how well a provider secures its offerings is one challenge covered further in the section on supply chain security below. A related cyber hard problem is the process of incident response, for the reasons outlined above. Putting these together creates an overall cyber hard problem that deserves attention because one provider’s outage or compromise can affect literally thousands or even millions of customers. The notion that every organization is solely responsible for securing itself is outdated, as is acknowledged in the most recent White House strategy document cited earlier. The answer lies not just in another technology product that customers must layer on top of the already complex infrastructure (complexity to manage complexity), but includes aligned incentives and clearer, more consistent security standards and responsibilities for these “linchpin” cloud providers.
Cloud infrastructure is included in cyber hard problems 1, 2, 3, 4, 5, and 10.
The democratization of technology has increased the number of accounts each user relies on dramatically, but the demographics of those users have changed. Consumers as
___________________
5 Ibid.
6 Office of the National Cyber Director (ONCD), 2023, “The National Cybersecurity Strategy,” The White House, March 2, https://bidenwhitehouse.archives.gov/oncd/national-cybersecurity-strategy.
young as 3 and as old as 103 must (securely!) authenticate themselves for purposes as varied as online games, banking, education, government services, medical care, and employment. Attackers have taken advantage of the gaps in implementation and differing levels of user sophistication by capturing credentials or tricking the legitimate account holder into exposing authentication information. Identity, authentication, and access control have become a significant battlefield and a favored “soft target” even as viable technical solutions have been accepted.
In response, identity and access management (IAM) technology has evolved to meet this need.7 IAM trends include the following:
___________________
7 National Institute of Standards and Technology (NIST), 2023, Digital Identity Guidelines, SP 800-63, https://pages.nist.gov/800-63-3.
8 Federal Bureau of Investigation, 2022, “Criminals Increasing SIM Swap Schemes to Steal Millions of Dollars from US Public,” Public Service Announcement: Alert Number I-020822-PSA, February 8, https://www.ic3.gov/PSA/2022/PSA220208.
Around the early 2000s, the UK-based Jericho Forum proposed a stronger authentication model, called a “collaboration-oriented architecture.” With the principle of explicitly authenticating every access request, regardless of where it originated, “zero trust”9 resulted in many additional authentication factors being developed, such as GPS-based location, just-in-time analysis of the security state of the device being used for access, WiFi fingerprinting, biometrics, passkeys, and more. The additional options in this factor portfolio also made it more difficult to build and test consistent access policies.
Another complicating factor in IAM is time. Authentication has moved beyond a one-time event into continuous evaluation of the user’s factors, including location or security state changes during the session and activity alerts. Depending on the assessed risk and policy, the system might invoke a step-up authentication process with more factors to ensure that the access is legitimate. The continuous assessment may also take specific events into account, using data received either from the system owner’s own infrastructure (such as network telemetry, application changes, or a change in user access from the identity provider) or from collaborating entities. For example, a password change should invalidate all currently open sessions. The OpenID Foundation Shared Signals working group10 is tackling the challenge of standardizing and sharing access-related events, but as with many information-sharing initiatives in cybersecurity, misaligned incentives can hamper this goal.11 Finally, there is the concept of granting authorization dynamically, in a just-in-time fashion, rather than equipping an account with static permissions. The “zero standing privileges” approach is intended to harden existing user accounts in the face of attacks but carries its own set of associated management complications.
Identities themselves are now more widespread and context-specific than they were 20 years ago—they involve not just “who are you?” but “why should you have access to this particular resource at this point in time?” Establishing the right to access by verifying that the user is a citizen, a parent, a doctor, a partner, an employee, a customer, or a student can require collecting attributes from many different trusted parties.
___________________
9 S. Balaouras, J. Blankenship, D. Holmes, P. McKay, J. Burn, A. Tatro, and M. Belden, “The Business of Zero Trust Security,” Forrester, https://www.forrester.com/zero-trust, accessed February 6, 2025.
10 T. Cappalli, S. Miel, S. O’Dell, and A. Tulshibagwale, “Shared Signals Working Group—Overview,” OpenID, https://openid.net/wg/sharedsignals, accessed February 6, 2025.
11 Any time information sharing is voluntary, commercial drivers can get in the way of sharing useful and complete information. For example, security vendors with their own threat intelligence teams may avoid sharing unique data if it is seen as a competitive advantage or delay release past the point of timeliness in order to publish within marketing schedules.
The collection of personal data, often incentivized financially for marketing and resale, contributes to the attack surface for each individual, as threat actors can obtain a wider variety of demographic data and secrets needed to register, use, and recover access.12
Collecting this verified data, storing them securely, and only releasing them where necessary are all associated privacy challenges. Proposed solutions include a digital self-sovereign identity framework, such as the European eIDAS regulation13 that will require every European Union country to offer a digital identity wallet by 2026. U.S.-based government services such as Global Entry are now offering digital IDs for mobile devices. Large-scale public identity providers such as Apple, Facebook, and Google have been offering to ease usability for consumers by letting them use their account identities for logging in to other sites, making payment transactions, and so on. However, these varied offerings come with their own governance and privacy goals, which the general public may not be able to evaluate.
To make matters even more complicated, identity management and governance have moved beyond the realm of humans. Machine identities, workload identities, and the operational system accounts that underpin all types of infrastructure, from applications to network routers, all need to be addressed in a coherent way, particularly with the growth of the Internet of Things (IoT). Wherever access is not tied to an individual human, or wherever two entities communicate with one another without human initiation, the authentication, authorization, and identity issues still apply. The Workload Identity working group14 is addressing some of these issues, but the drudgery of tracking, auditing, and protecting dormant on-premises system accounts remains with the owners of the infrastructure. Not only do attackers regularly target default passwords on these systems, but the potential areas for attack now range from critical infrastructure (utilities, nuclear power plants, 911 systems, medical equipment) to security cameras, home thermostats, baby monitors, and indeed anything that is connected to the Internet under the guise of being “smart.” Non-human identities are equally important to combat cases where an attacker simulates a website or message to trick a user into supplying credentials; the machine needs to authenticate itself to the human.
One final point is that although IAM frameworks and technology have evolved, they are also extending the “long tail” of legacy systems that are too costly to retrofit. For every web-based application that now uses passkeys for authentication, there is also a decades-old banking mainframe or industrial storage tank that must still interoperate.
___________________
12 Department of Defense (DoD), 2023, “2020 DSB Summer Study on New Dimensions of Conflict: Executive Summary,” DoD Office of Prepublication and Security Review 23-S-2072, April, https://dsb.cto.mil/wp-content/uploads/reports/2020s/DSB-SS2020_NewDimensionsofConflict_Executive%20Summary_cleared.pdf.
13 European Commission, 2024, “eIDAS Regulation,” April 4, https://digital-strategy.ec.europa.eu/en/policies/eidas-regulation.
14 IETF Datatracker, “About the IETF Datatracker,” https://datatracker.ietf.org/release/about, accessed February 6, 2025.
The overarching technical hard problem within IAM is configuring and managing policy. Because of complicated mechanisms for performing authentication and authorization, expanded user demographics, and many, often ephemeral, systems that require IAM without governance, the need to protect these vulnerable attack points requires simplifying and possibly centralizing policy management. Today’s chief information security officer has no straightforward way to decide which factors to use, how to model the operational impact of a policy change, how to negotiate policy enforcement with external providers, or even how to get all the event data they need to make (or possibly automate) risk decisions involving policy across the entirety of the technologies and environments. One example of the current fractured state of IAM is universal access revocation, sometimes called “single logout”—the problem of identifying all accesses belonging to a departing user and revoking them, terminating any existing live sessions, and handling deeper layers of associated application and system access such as cryptographic tokens. This “holy grail” of IAM applies not only to access management in workplace scenarios but also to any active incident response involving a compromised user account. Achieving it comes with trade-offs—for example, tracking every place where a user is active can also have privacy implications.
As mentioned above, the nontechnical hard problem aspect of IAM involves governance. Identity management, authentication, and access control are driven by commercial entities and are fragmented accordingly, as governance falls to a population of private and public resource owners, not simply a central one in the role of an enterprise employer as in the past. In countries where each citizen has a single government-managed digital ID, the resulting centralization affords better technical solutions and enhances accessibility for underserved populations. Long-standing mistrust of centralized government in the United States stands in the way of creating a centralized digital identifier. Where cultural distrust of centralization is higher, it may be more practical to develop a broad federation model, allowing disparate resource owners and consumers to use a consistent and reliable framework for negotiating IAM features and processes.
Access controls affect cyber hard problems 1, 2, 3, 4, 6, 7, and 10, although it affects others to a more limited extent.
Cybersecurity is the property of technological artifacts, people, and processes to resist attacks by an adversary. At every stage of a system’s life cycle, including design, implementation, acquisition, testing, deployment, training, use, monitoring, maintenance, and retirement, there are a broad range of decisions to be made that will influence these properties. What authentication architecture should be chosen? What programming languages, tools, and processes should be used to minimize the introduction of
implementation bugs? When acquiring such a system, how should an organization compare the security it offers to competing systems? How should their information technology (IT) professionals configure the system to support its security assumptions? How should employees be trained to use it? When must a system be updated or retired due to security liabilities? Such questions, explicit and implicit, are being answered thousands of times a day. It is widely held that some of these choices are likely better than others. Indeed, many believe that there may be “best” choices for a given situation and that certain decisions, if taken, would seriously foreclose attacks.
While it is tempting to hope that these questions might be answered a priori—that with the proper levels of formal reasoning, systems might be designed and proven secure against reasonable threat assumptions—such results are rarely available. Real systems operate in a messy world, typically more complex than can be modeled, with countless deviations from idealized abstractions, with multiple humans in the loop, and adversaries who formulate their attacks based on the assumptions made by defenders.
This is strong motivation to place cybersecurity on a firm empirical footing—akin to evidence-based medicine—where careful data collection and analysis can differentiate and prioritize among the plethora of factors and approaches. However, a perennial challenge for the cybersecurity community has been to establish a rigorous evidentiary basis for evaluating such choices in a way that predicts outcomes. As a result, most of today’s established cybersecurity “best practices” are based on a combination of perceived common sense and received wisdom.
There are many reasons to question the quality of this status quo decision making.
___________________
15 USENIX, 2020, “29th USENIX Security Symposium,” August 12–14, https://www.usenix.org/conference/usenixsecurity20.
16 IEEE Symposium on Security and Privacy, 2017, “38th IEEE Symposium on Security and Privacy,” https://www.ieee-security.org/TC/SP2017.
___________________
17 A. Shostack, M. Smith, S. Weber, and M.E. Zurko, 2019, “Empirical Evaluation of Secure Development Processes,” Dagstuhl Reports 9(6)1–25, Schloss Dagstuhl – Leibniz-Zentrum für Informatik, https://doi.org/10.4230/DagRep.9.6.1.
18 R. McMillan, 2017, “The Man Who Wrote Those Password Rules Has a New Tip: N3v$r M1^d!” Wall Street Journal Pro-Cybersecurity, August 7, https://www.wsj.com/articles/the-man-who-wrote-those-password-rules-has-a-new-tip-n3v-r-m1-d-1502124118.
19 R. Morris and K. Thompson, 1979, “Password Security: A Case History,” Communications of the ACM 22(11):594–597.
20 A. Adams and M.A. Sasse, 1999, “Users Are Not the Enemy,” Communications of the ACM 42(12):40–46.
This is far from a new realization. “Metrics for Security” was identified as a key cyber hard problem in the 1995 version of the InfoSec Research Council’s Hard Problem List,24 the 2002 National Research Council consensus study report Cybersecurity Today and Tomorrow: Pay Now or Pay Later,25 the Computing Research Association’s 2003 “Four Grand Challenges in Trustworthy Computing,”26 the 2005 President’s Information Technology Advisory Committee report Cyber Security: A Crisis of Prioritization,27 and again in the InfoSec Research Council’s 2005 re-up of Hard Problem List28—relabeled as “Enterprise-level Security Metrics” (although not because the smaller scale problems had been solved, indeed that study states, “Most of the existing [security] metrics are of questionable utility, even with respect to individual software systems.”29,30). Indeed, almost 20 years later, the software-focused 2024 Office of the National Cyber Director (ONCD) report Back to the Building Blocks: A Path Toward Secure and Measurable Software opines that still “it is critical to develop empirical metrics that measure the cybersecurity quality of software.”31
___________________
21 D. Lain, T. Jost, S. Matetic, K. Kostiainen, and S. Capkun, 2024, “Content, Nudges and Incentives: A Study on the Effectiveness and Perception of Embedded Phishing Training,” arXiv:2409.01378.
22 L.F. DeKoven, A. Randall, A. Mirian, G. Akiwate, A. Blume, L.K. Saul, A. Schulman, G.M. Voelker, and S. Savage, 2022, “Measuring Safety Practices,” Communications of the ACM 65(9):93–102.
23 C. Thompson and D. Wagner, 2017, “A Large-Scale Study of Modern Code Review and Security in Open Source Projects,” PROMISE ‘17, November 8, https://people.eecs.berkeley.edu/~daw/papers/coderev-promise17.pdf.
24 The 1995 Infosec Research Council (IRC) Hard Problems is not easily found, but the problems themselves are available in Appendix A, “Retrospective on the Original Hard Problem List,” of the 2005 Hard Problem List report. See IRC, 2005, Hard Problem List, November, https://www.nitrd.gov/documents/cybersecurity/documents/IRC_Hard_Problem_List.pdf.
25 National Research Council, 2002, Cybersecurity Today and Tomorrow: Pay Now or Pay Later, National Academy Press, https://doi.org/10.17226/10274.
26 Computing Research Association, 2003, “Four Grand Challenges in Trustworthy Computing,” https://archive.cra.org/Activities/grand.challenges/security/grayslides.pdf.
27 President’s Information Technology Advisory Committee, 2005, Cyber Security: A Crisis of Prioritization, National Coordination Office for Information Technology Research and Development, February, https://www.nitrd.gov/pubs/pitac/pitac_report_cybersecurity_2005.pdf.
28 IRC, 2005, Hard Problem List.
29 IRC, 2005, Hard Problem List, p. 56.
30 D. Maughan, 2006, “Infosec Research Council Hard Problem Lists,” Department of Homeland Security, Science and Technology Directorate, January 26.
31 ONCD, 2024, Back to the Building Blocks: A Path Toward Secure and Measurable Software, The White House, February, https://bidenwhitehouse.archives.gov/wp-content/uploads/2024/02/Final-ONCD-Technical-Report.pdf, p. 11.
Given the clear need, why, then, has there been such limited progress in cybersecurity while other fields, such as medicine, have been able to incorporate empirical data to great success?
The traditional refrain, well described in Herley and van Ooorschot’s overview paper on the “science of security,”32 is a triumvirate of problems that describe the “unique challenge” in measuring cybersecurity.
While these challenges are real, none seem fundamentally at odds with the notion of empiricism or the scientific method. Indeed, a range of other disciplines faces one or more of these issues and still fruitfully make use of empiricism in practice. The common thread among all three of the problems is dynamism, which occurs in a variety of other disciplines as well. For example, while non-adversarial, insurers’ assumptions about various kinds of property damage risk have been repeatedly updated and revised in response to observations that prior measured likelihood distributions were no longer predictive (e.g., concerning hurricanes or wildfires). Finally, closest to cybersecurity, economists routinely drive policy decisions using empirical tools, despite addressing a system that is, at its core, both adaptive and quasi-adversarial (few would seriously argue
___________________
32 C. Herley and P.C. van Oorschot, 2017, “SoK: Science, Security and the Elusive Goal of Security as a Scientific Pursuit,” 2017 IEEE Symposium on Security and Privacy (SP) 99–120, https://oaklandsok.github.io/papers/herley2017.pdf.
33 D. Evans and S. Stolfo, 2011, “Guest Editors’ Introduction: The Science of Security,” IEEE Security & Privacy 2011(9):16–17, https://www.computer.org/csdl/magazine/sp/2011/03/msp2011030016/13rRUwh80sR.
that the Federal Reserve would make better decisions if only it ignored empirical data). While each of these settings differs from cybersecurity in key ways, all are also removed from the idealized notion in which measurements can be used to derive static generalizable laws that then offer perfect predictive power. That such empirical analyses may be imperfect or have limited lifetime does not eliminate their value—for the alternative is to act without the benefit of concrete evidence at all.
In the late 20th century, portions of the medical community popularized evidence-based medicine (EBM) to incorporate a range of empirical evidence to guide research and ultimately update practice guidelines with the singular goal of improving clinical outcomes. Firmly embedded in the scientific method, EBM fostered hypothesis generation from laboratory and qualitative studies, building into both prospective and retrospective case studies, then driving repeated randomized controlled trials of prospective treatments and filtered based on comprehensive meta-analysis. There is little debate that this effort has been transformative for the practice of medicine. While this precise formulation is unlikely to translate directly to the cybersecurity realm, a similar kind of focus and investment to pursue outcome-focused results is needed.
However, there is a range of obstacles that will need to be addressed to make progress, including the following:
___________________
34 CyberGreen, “Indirect Cost Policy,” https://cybergreen.net/technical-report-22-01/in, accessed February 6, 2025.
___________________
35 Verizon, 2024, “2024 Data Breach Investigations Report,” https://www.verizon.com/business/resources/reports/2024-dbir-data-breach-investigations-report.pdf.
Despite these roadblocks, there has been considerable innovation in the “evidence-based security” space since the last version of this report. Among the approaches that have borne fruit are the following:
___________________
36 USENIX, 2015, “24th USENIX Security Symposium,” August 12–14, https://www.usenix.org/conference/usenixsecurity15/technical-sessions/presentation/liu.
37 A. Searles, Y. Nakatsuka, E. Ozturk, A. Paverd, G. Tsudik, and A. Enkoji, 2023, “An Empirical Study & Evaluation of Modern CAPTCHAs,” pp. 3081–3097 in 32nd USENIX Security Symposium, https://www.usenix.org/conference/usenixsecurity23/presentation/searles.
38 K. Thomas, D. McCoy, C. Grier, A. Kolcz, and V. Paxson, 2013, “Trafficking Fraudulent Accounts: The Role of the Underground Market in Twitter Spam and Abuse,” 22nd USENIX Security Symposium, August 14–16, https://www.usenix.org/conference/usenixsecurity13/technical-sessions/paper/thomas.
39 Computer and Communications Security, 2014, CCS ‘14: Proceedings of the 2014 ACM SIGSAC Conference on Computer and Communications Security, Association for Computing Machinery.
40 A. Mirian, J. DeBlasio, S. Savage, G.M. Voelker, and K. Thomas, 2019, “Hack for Hire: Exploring the Emerging Market for Account Hijacking,” pp. 1279–1289 in WWW ‘19: The World Wide Web Conference, May 13, arianamirian.com/docs/www2019_hfh.pdf.
Finally, one highly-desired manifestation of evidence-based security is the establishment of security metrics—parsimonious measures allowing the evaluation and/or comparison of the security offered by a particular system. The kinds of data that will be needed for an empirically based security research agenda are clearly amenable to being shaped into metrics. However, what makes metrics attractive is that they abstract and simplify. One can easily use metrics as a decision criterion for whether an organization has improved or not, whether vendor A or vendor B is more secure, or as defense against liability. However, this same attractiveness creates strong incentives for standardization and can create institutional inertia that makes it difficult to change or react when the context in which the metrics were measured has changed. Even worse, failures in analysis or validation might elevate “bad” metrics, which ultimately incentivize less secure decisions. The same result can come from the incentive to “game” good metrics. Thus, it is critical to also consider how to deliver evidence-based security research, without enabling the most negative aspects of institutional desires for cheap decision making.
Many security metrics, from “speed to patch” to the collection of end user agent parameters, have been developed. However, there is no basis for believing any set of existing metrics provides an accurate prediction of safety or a root-cause analysis of previous losses.
While there is some research on software developers’ ability to produce secure software, this research seems to have had little impact on actual software development tools and processes. Secure development life-cycle practices and the tools that support them have largely come from industry and often have their own usability challenges.
___________________
41 Cybersecurity and Infrastructure Security Agency (CISA), 2016, “Alert: OpenSSL ‘Heartbleed’ Vulnerability (CVE-2014-0160),” October 5, https://www.cisa.gov/news-events/alerts/2014/04/08/openssl-heartbleed-vulnerability-cve-2014-0160.
42 Github, “AttackSufaceAnalzyer,” Microsoft, https://github.com/microsoft/AttackSurfaceAnalyzer/pulls, accessed February 6, 2025.
Some training to help software developers produce more security software can produce recommendations that seem on the surface impractical.
The inability to develop compact, predictive, measurable security metrics informs and affects essentially all of the consumer cyber hard problems.
Many of the traditional problems of “secure design and composition,” described in detail above, include careful specification, isolation, and partitioning of functionality—following the principle of “least privilege” by authenticating the principal on whose behalf actions are taking and verifying their “right” to take such an action (the basis for “zero trust”). There are also emerging technologies, such as “confidential computing,” which provides a strong, principled basis for authentication of programs (an important security principle) to establish a principled, distributed basis for partitioning, isolation, and trust management. Several “producer technologies” for achieving secure resilient design and composition are discussed below.
There is a wide diversity of engineering interventions, ranging from selecting safer programming languages such as Rust and TypeScript to making architectural choices that enhance the potential for resilient response to compromise. As noted above in this chapter, one of the most vexing cyber hard problems in cybersecurity is measurement. Measurement difficulties thwart progress, not just in implementing secure engineering practices but also in understanding trade-offs when there are choices to be made regarding which practices to adopt.
For example, how much benefit is to be obtained from using a safer programming language, and how does this compare, say, with using improved analysis tools? There may be trade-offs that involve traditional software engineering criteria—for example, how does architecting to reduce interdependency among components of a large system interact with choices to implement a design that limits trust assumptions among system components when they interact (in the same sense as zero trust, but at an internal implementation level)?
Further exacerbating this challenge is a set of enduring perceptions that engineering efforts that are directed at enhancing security and resilience have an uncertain return on investment.
When successful, however, secure engineering practice can have significant—and otherwise unattainable—benefits in reducing many aspects of cyber risk. Indeed, while it may be challenging to characterize the role of specific development or design practices, there is a range of proxies suggesting that “zero-day” exploits for commodity smart-phone and server platforms have become harder to procure over time, as evidenced by their increasing market price and owing to the increasing length and complexity of
exploit chains needed. Moreover, there are cases where secure engineering practice leads not only to improved risk posture but also improved productivity and, in some instances, enhanced system performance. It is reported, for example, that many users of memory-safe programming languages (e.g., Rust and TypeScript) make the adoption choice in the interests of productivity, with secondary consideration for the harder-to-measure security benefits. This means that, even if the security benefits cannot readily be measured, the adoption of improved practices based on concomitant benefits to productivity and performance can nonetheless be promoted.
There are many examples of guidance regarding secure engineering from firms, laboratories, and government agencies, including the National Security Agency, the Cybersecurity and Infrastructure Security Agency, and ONCD. Most recently, for example, ONCD issued advice regarding secure practices with a focus on memory safety and formal methods43 in the face of the paucity of cybersecurity metrics. Historically, the Microsoft Security Development Lifecycle (SDLC), which includes a mix of interventions and practices focused on process and product, has had broad adoption and, as perceived by engineering managers, meaningful benefits.44 Much of the guidance focuses on practices and processes that are associated with improved security outcomes. An example is “secure coding practice,” which involves making coding choices that reduce vulnerabilities, as demonstrated through techniques such as fuzz testing.45
Many security-related weaknesses and vulnerabilities go beyond simple coding practices and derive from the logical structure of software and firmware. These flaws can range from protocol and API misuse to erroneous business rules. These are logic flaws. The focus of logic flaw problems is not on the full scope of secure engineering practices, but rather on means to achieve verifiable assurances regarding the absence of certain categories of vulnerabilities that go beyond type safety and memory safety. This report focuses on logic flaws for two reasons. First, there is increasingly broad adoption of “traditional” secure engineering practices such as SDLC (and as assessed through instruments such as BSIMM).46 Second, logic flaws account for an increasing percentage of exploits,47,48 and these flaws are not readily detected using current techniques. The usual means to detect
___________________
43 ONCD, 2024, Back to the Building Blocks.
44 Microsoft, “Microsoft Security Development Lifecycle (SDL),” https://www.microsoft.com/en-us/securityengineering/sdl, accessed February 6, 2025.
45 OWASP Foundation, “Secure Coding Practices,” https://owasp.org/www-project-secure-coding-practices-quick-reference-guide/stable-en/02-checklist/05-checklist, accessed February 6, 2025.
46 BlackDuck, “What Is BSIMM?” https://www.blackduck.com/glossary/what-is-bsimm.html, accessed February 6, 2025.
47 OWASP, “Top Ten,” https://owasp.org/www-project-top-ten, accessed March 25, 2025.
48 S. McClure, 2024, “Safeguarding from Lurking Threats in Business Logic Flaws,” Fast Company, January, https://www.fastcompany.com/91013667/safeguarding-from-lurking-threats-in-business-logic-flaws.
and mitigate logic flaws is through inspection and testing. But it is well understood that these methods are imperfect.
There is an unhappily rich collection of security attributes, as revealed in the several taxonomies that are widely referenced (e.g., CIA,49 STRIDE,50 MITRE ATT&CK). Mathematical techniques can be used to address some of these, at different levels of complexity and scale. However, some attributes cannot currently be readily modeled mathematically. These include side-channel attacks. An example familiar to security researchers is cryptographic algorithms, which are mathematically correct and whose implementation is proved consistent with the algorithms, but whose implementations on actual physical processors creates vulnerabilities based on the physics of the operation of the processors, such as power fluctuations, RF emissions, and timing of executions.
Process compliance, in many cases, is seen as more affordable to achieve than actual measurable security. Although sometimes useful,51 compliance is expensive, subject to manipulation by “well-resourced” organizations, and it delays innovation and is not really very effective, generally, in “guaranteeing” security.
Formal methods (FM), in contrast, are direct techniques, focused on the operation rather than adherence to processes in the creation of that product. These techniques, including verification and program analysis for various functional and quality attributes, have a long history, going back at least to the 1960s with work by Robert Floyd and later Tony Hoare. For many years, the principal uses were in critical applications such as commercial flight controls, embedded medical devices, and national security applications. In the past 5 years, however, the scope of application has broadened significantly to include many commercial uses. Some evidence of this is cited in the Networking and Information Technology Research and Development publication regarding the FM@Scale workshops,52 where several at-scale commercial uses are highlighted. These use cases suggest the possibility that barriers of affordability, scale, usability, and integration can be overcome for a broader range of applications, with significant benefits not just to reducing important categories of vulnerabilities but also in providing evidence in support of cybersecurity risk assessment and certification. These successes depend on our ability to express models, including specifications, for quality attributes relevant to security. Improving the scope, expressiveness, and ease of use of modeling can significantly
___________________
49 CIA refers to confidentiality, integrity, availability.
50 STRIDE refers to spoofing, tampering, repudiation, information disclosure, denial of service, elevation of privilege.
51 The National Information Assurance Partnership (NIAP) Common Criteria, for example, can involve deep analysis of design artifacts and some sampling of code in an evaluated system. See NIAP, “Common Criteria: IT Security Evaluation,” https://www.nsa.gov/Portals/75/documents/resources/everyone/2023-02-NIAP_brochure_trifold_1.pdf, accessed February 6, 2025.
52 R.W. Floyd, 1967, “Assigning Meanings to Programs,” pp. 19–32 in Proceedings of Symposium on Applied Mathematics (19): Mathematical Aspects of Computer Science, J.T. Schwartz, ed., American Mathematical Society.
reduce the cycle of information loss and recovery that typically occurs, with great cost, as systems evolve over time. Information loss of this kind creates challenges for test and evaluation as well as for sustainment and evolution. It is one of the chief contributors to technical debt.53
An additional enabler, already in place, for maintaining continuity, enabling agility, and minimizing information loss across the full life cycle of software-based systems is the Software Acquisition Pathway. The benefits were reinforced in a recent memo from the Secretary of Defense.54,55
The following three examples illustrate the various ways that these barriers are now being overcome:
___________________
53 CISA, 2025, “Closing the Software Understanding Gap,” January 16, https://www.cisa.gov/resources-tools/resources/closing-software-understanding-gap.
54 DoD, 2025, “Directing Modern Software Acquisition to Maximize Lethality,” Memorandum for Senior Pentagon Leadership Commanders of Combatant Commands Defense Agency and DoD Field Directors, from the Secretary of Defense, March 6, https://media.defense.gov/2025/Mar/07/2003662943/-1/-1/1/DIRECTING-MODERN-SOFTWARE-ACQUISITION-TO-MAXIMIZE-LETHALITY.pdf.
55 DoD, 2020, “Operation of the Software Acquisition Pathway,” DoD Instruction 5000.87, October 2, https://www.esd.whs.mil/Portals/54/Documents/DD/issuances/dodi/500087p.pdf.
56 Defense Advanced Research Projects Agency (DARPA), “HACMS: High-Assurance Cyber Military Systems,” https://www.darpa.mil/program/high-assurance-cyber-military-systems, accessed February 6, 2025.
57 A. Miller, “HACMS,” GALOIS, https://galois.com/project/hacms-high-assurance-cyber-military-systems, accessed February 6, 2025.
It is evident that there is progress in FM and that barriers to its use can be overcome in specific cases, including usability, scale and composition, and integration into engineering workflows. These examples are illustrative, but there are significant barriers.62 The recent ONCD report highlights two areas—software and hardware memory safety, such as provided through Rust and similar languages, and FM to support affirmative claims—backed by evidence—regarding particular security properties.63
The solution of these problems and application and further development of these techniques acutely influences hard problems 1, 2, and 3 but also 8, 9, and 10.
A recent President’s Council of Advisors on Science and Technology (PCAST) report64 drew attention to critical CPS (which they call “operational technology systems”). These systems include the “integrated digital and physical resources that are crucial to
___________________
58 C. Thompson, 2023, “How Rust Went from a Side Project to the World’s Most-Loved Programming Language,” MIT Technology Review, February 14, https://www.technologyreview.com/2023/02/14/1067869/rust-worlds-fastest-growing-programming-language.
59 L. Lamport, 2024, “The TLA+ Home Page,” August 13, https://lamport.azurewebsites.net/tla/tla.html.
60 Amazon Web Services, “Provable Security Resources,” Cloud Security, https://aws.amazon.com/security/provable-security/resources, accessed February 6, 2025.
61 N. Rungta, 2022, “A Billion SMT Queries a Day,” Amazon Science, https://www.amazon.science/blog/a-billion-smt-queries-a-day.
62 P. Lincoln, W. Scherlis, and W. Martin, 2022, Formal Methods at Scale: 2019 Workshops Report, Computing-Enabled Networked Physical Systems Interagency Working Group, May, https://www.nitrd.gov/pubs/Formal-Methods-at-Scale-Workshops-Report.pdf.
63 ONCD, 2024, Back to the Building Blocks.
64 President’s Council of Advisors on Science and Technology (PCAST), 2024, Strategy for Cyber-Physical Resilience: Fortifying Our Critical Infrastructure for a Digital World, Executive Office of the President, February, https://bidenwhitehouse.archives.gov/wp-content/uploads/2024/02/PCAST_Cyber-Physical-Resilience-Report_Feb2024.pdf.
Americans’ daily lives, including the electrical grid, public water systems, internet and telecommunications, air traffic control and much more.” IoT and automated manufacturing systems that incorporate sensors or actuators, and indeed any CPS, can be included in this category as well. The PCAST report goes on to note:
Cyber-physical risk is high, while protections are disproportionately low. America’s infrastructure systems were created and operated long before they acquired cyber dependencies, with sensing, computing, and networking dependencies developing in different ways over time. There is no systemic, pervasive protection against cyber risk since our protections evolved over time.65
CPS (e.g., a car, a laboratory instrument, or a medical device in a doctor’s office) may have very long lifetimes. Future systems have to be shaped by cyber-informed engineering. Much of the technology that underpins cyber systems and CPS was engineered without appropriate consideration of security needs. Security and resilience elements are tacked on after systems are deployed, often imperfectly and at considerable expense.
As observed above, many CPS security practices lag far behind those of IT systems. These systems use old software versions and are often not upgraded, they are not managed under user policy, and there is seldom support for critical hardware features like “root of trust” nodes. Finally, they are especially vulnerable since they are exposed to physical attacks in addition to network-based attacks.
This is cyber hard problem 8, but it is affected by almost all the other “producer” cyber hard problems, including problem 10 (operational security).
Operational cyber hard problems address securely operating a scale infrastructure, including responding to attacks. This is cyber hard problem 10. Earlier discussion has already described the importance of continuous updating, resilient deployment and operations, monitoring, and situational awareness. These are the core elements of resilient secure operation, and large cloud providers have made great progress in this area. However, customer insight into the effectiveness of already introduced measures is modest, and small providers as well as “in house operations” often suffer in comparison to the effectiveness of the operational security of a large cloud provider. This includes the operational infrastructure of CPS.
___________________
65 PCAST, 2024, Strategy for Cyber-Physical Resilience, p. 12.
It is important to acknowledge that there are human attackers who make decisions and can be influenced to defenders’ advantage. These attackers have a profound influence on the secure operation of a cyber system.
For much of the history of cybersecurity, defenders have thought about attackers as an abstract and homogenous group. This meant that defenses were applied to technical attributes of attacks, such as techniques and tools, rather than through understanding the humans behind the attacks. Maturing threat models identifying threat groups can differentiate capabilities, objectives, and likely victims. A cyber hard problem, however, remains incorporating insights about the human attributes of attackers to create tailored defenses.
Attackers routinely exploit the human weaknesses of their victims; however, defenders lack sufficient insights to effectively incorporate human factors into defense. Historically, deception, such as honeypots, has been used to manipulate attackers to gain intelligence, but these techniques are neither widespread nor generally evaluated for effectiveness. In addition, it is difficult to persuade engineering organizations, which are mainly rewarded for adding new features to work on adversarial engineering projects. Recent research on adversarial human factors is starting to identify and analyze the human attributes of attackers. More research funding is needed. Industry partnership will likely be critical for developing commercial defensive capabilities that apply adversary human factors. Together, industry and the research community have to develop metrics for evaluating the effectiveness of the approach.
Understanding the human attributes of attackers is crucial for several reasons. Primarily, it allows for more sophisticated and targeted defense mechanisms that go beyond merely blocking attacks or deactivating users to anticipating and mitigating them. This knowledge can significantly improve the ability to prevent breaches and reduce the impact of successful attacks. Governments, businesses, and individuals all have a stake in this issue because the consequences of cyberattacks can range from financial loss and reputational damage to national security threats. By understanding the motivations, psychological traits, economic incentives, and behavioral patterns of attackers, defenders can craft strategies that are more likely to deter or disrupt malicious activities.
The complexity of this problem lies in the inherent variability and adaptability of human behavior. Attackers come from diverse backgrounds, possess different skill levels, and are driven by a range of motives, including financial gain, political activism, espionage, or personal vendettas. The dynamic nature of human attributes makes it challenging to create a one-size-fits-all defense. Moreover, ethical and privacy considerations must be balanced when researching and using human factors in cybersecurity.
Lawyers taking a position that even attackers have privacy rights limits defender access to their internal communications and communication with fraud victims. The incentives for attackers are high and constantly evolving, while defenders often face significant resource constraints and organizational inertia, making it difficult to implement and adapt new strategies to increase the cost seen by attackers.
Potential approaches to this problem include interdisciplinary research that combines insights from psychology, sociology, and behavioral economics with cybersecurity. Developing comprehensive attacker profiles and predictive models can help in understanding likely behaviors and vulnerabilities. Simulation and gaming techniques can be employed to study attacker behavior in controlled environments. Additionally, leveraging machine learning (ML) and AI to analyze patterns in large data sets can provide deeper insights into attacker traits and tactics. Collaboration between academia, industry, and government will be essential to foster innovation and translate research findings into practical applications. Note, however, that attackers will be increasingly autonomous (bots) and will also employ AI to probe systems and people.
Key players who can act include cybersecurity researchers, defense agencies, technology companies, and policy makers. Cybersecurity firms and tech companies can incorporate human factors insights into their security products and services. Governments can fund research initiatives and create frameworks that encourage information sharing and collaboration. Policy makers can help by enacting regulations that support ethical research, while researchers can make progress in blinding and masking personal information to make it less easy to reidentify, to protect individual privacy. Education and training programs for cybersecurity professionals need to also emphasize the importance of understanding attacker psychology and behavior.
Success in this endeavor can be measured through several indicators. A reduction in the frequency and severity of successful cyberattacks would be a primary indicator. Improved response times and more effective mitigation strategies in the face of attacks would also signify progress. An increase in the cost of credentials or breach data offered on dark markets would be a direct observable measure of success. Additionally, the development and widespread adoption of new defense technologies and methodologies that incorporate human attributes of attackers would demonstrate advancement in this field. Regular assessments and refinements based on real-world data and feedback will be necessary to ensure the ongoing effectiveness of these approaches.
This informs cyber hard problem 10.
There are few tools or techniques to predict how cyber systems may be used in the future, especially as relates to unforeseen use that can threaten an individual’s security,
the security of an organization, or the nation. Attacks that seem to have been largely unforeseen in the past include doxing, stalkers transitioning to the digital domain, false reporting to cause account deactivation, and disinformation and influence operations that aim to attack democratic processes.
The legacy of online anonymity, including the ease of creation of opaque personas in different forums, facilitates not just free speech but also an important category of adversarial information operations, the use of bots, artificial amplification of adverse memes, and impersonation of individuals and organizations. In certain online contexts, by contrast, identity has to be firmly established, such as for banking, health records access, and business process execution.
The challenge is to identify multifaceted approaches to identity that could thwart the adversarial behaviors that depend on lack of provenance while nonetheless supporting the many contexts where it is needed, including free speech and augmenting protections of the identity of victims and potential targets of abuse and attack (e.g., human rights workers).
Anonymity also facilitates socially destructive behaviors such as stalking, bullying, sexual harassment, and doxing (piercing the veil of the privacy of other people). What are technical approaches or potential remedies that do not require abandoning the free-speech benefits of anonymity?
Computer users are called on to make decisions and choices with cybersecurity consequences that they do not have sufficient knowledge and resources to make accurately or securely.
Decisions and choices may be explicit or implicit. Users may be using their own computer or a computer owned by their organization. They may be acting as an individual or as a member of an organization, such as an employee or even an administrator. They may be using special purpose tools as experts in a particular discipline (i.e., code development or threat hunting). The resources they lack to address the decisions thrust upon them, some of which may be irrevocable, are legion—information, knowledge, understanding, context, time, memory, attention, desire, and incentive.
The Verizon 2024 Data Breach Investigations Report66 calls out that approximately two-thirds of breaches involve a (non-malicious) human element. This reflects the tendency, as noted in the IAM discussion above, for all sorts of cyber products to “kick the can down the road” to required human decision making, when the humans involved do not have sufficient resources to do so. In the IAM case above, what is lacking is
___________________
66 Verizon, 2024, 2024 Data Breach Investigations Report.
transparency into the system and an understanding of the system’s complexity, which seems obviously like a too-heavy lift to require of a user.
Even in the most straightforward context, research on expert advice about what security practices users, as consumers, need to follow shows that experts do not agree on the most important.67 The amount of expert advice is overwhelming and not necessarily coherent.
The responsibility for users to make impossible security decisions creates vulnerabilities that attackers discover and exploit using social engineering. The security impact from such attacks may be to the users themselves, their computers, their organization, or to the Internet at large. The unmet expectations placed on users deteriorate presumed protection levels (which may go undiscovered until it is too late). This can contribute to a state of habituation, or worse, learned-helplessness, for users—Why follow the security advice if I am not able to do it all, or even know which matters?
Testing with appropriately representative users is both expensive and difficult. There are no tools, frameworks, principles, or automation to do the required testing without humans in an adequate and rigorous fashion. There are still gaps between the best-of-breed testing methodologies in research and what’s available to testers, who are often entry-level employees.
While the techniques for testing are largely known, choice of testing subject needs to consider the user’s skills, knowledge, and context. Since those vary, testing needs to include a range of such subjects. In addition, because some of the cybersecurity-related decisions may come about from an error in the system, or an attack, the suite of interactions needing testing can be difficult to identify, and difficult to replicate for testing.
When users interact with devices and systems in unmanaged environments (personal use), they acquire ad hoc habits that are ineffective or insecure, and these are difficult to unlearn. Because there is often no discernable difference between effective security advice and ineffective security habits, users become accustomed to ignoring security advice and experiencing no known harms.
Builders may “punt to the user” to make decisions in uncertain design situations, which compounds the issue at hand. This shows up as either warnings presenting choices or new configuration settings, which are often hidden. If used correctly (with omniscience), perhaps the security would be improved, however usability of said features goes untested.
Security and privacy mechanisms today reflect current knowledge on how to expose and control, which often differs from the user’s cognitive model of the system.
___________________
67 E.M. Redmiles, N. Warford, A. Jayanti, A. Koneru, S. Kross, M. Morales, R. Stevens, and M.L. Mazurek, 2020, “A Comprehensive Quality Evaluation of Security and Privacy Advice on the Web,” Proceedings of the 29th USENIX Security Symposium, August 12–14, https://www.usenix.org/system/files/sec20-redmiles.pdf.
Builders build what they know how to build. Users use what they are given. The resulting gap present a symmetry of ignorance that must be overcome.68
This is part of cyber hard problem 6, “human–system interactions,” and cyber hard problem 2, “secure development.”
Even if human-centered design were addressed, support for human security workers is still insufficient to ensure optimal security outcomes. The workforce is core to cybersecurity, in practice, but there is not necessary support in place to help security workers thrive and perform, with education, tools, and processes. This includes workers of all sorts, including developers, designers, architects, IT, and specialists such as security architects, chief security officers, blue teams, content moderators, and fact checkers.
There are two interconnected facets to this problem—inadequate training and the inability to provide cybersecurity without that support. For generalists who are also security workers, the support needed includes training, resources, and incentives baked into the job, not exogenous. Education available to learn to code would teach best practices for coding securely. Specialized security worker education would cover the myriad of topics needed to be a security subject-matter expert—from architecture, to design, to coding security functions, to specialized security testing, to security in deployment and use.69
The issue of inadequate tools and techniques for security workers to develop secure code is addressed in the cyber hard problems above.
The pool of trained security workers is limited, and training in one job can lead to the ability to move to a better paying job, creating an ongoing training need. Jobs that involve operational security vigilance are high pressure and can lead to burnout. Lack of certification means it is impossible to enforce education, training, and standards of professional conduct.
Since the 1996 New Security Paradigms Workshop,70 there has been substantial work, both in research and practice, on usable security and privacy, particularly for individuals and consumers. However, there is less work on the many other humans involved in creating, maintaining, operating, receiving, and even attacking, security and privacy, and how they can be supported, or repelled.
___________________
68 O. Pieczul, S. Foley, and M.E. Zurko, 2017, “Developer-Centered Security and the Symmetry of Ignorance,” Proceedings of the 2017 New Security Paradigms Workshop.
69 As examples, NIST and the SANS Institute have training materials available for all levels from end users to cybersecurity professionals. See NIST, “Free and Low Cost Online Cybersecurity Learning Content,” Applied Cybersecurity Division, https://www.nist.gov/itl/applied-cybersecurity/nice/resources/online-learning-content, accessed February 6, 2025.
70 M.E. Zurko and R.T. Simon, 1996, “User-Centered Security,” New Security Paradigms Workshop, https://www.nspw.org/papers/1996/nspw1996-zurko.pdf.
Research in practical secure deployment is even sparser and tends to focus on individuals, largely overlooking organizational issues. Some research exists on why, how, and when individuals will accept security-related updates. Studies and measurements exist showing the rate of patch deployment. Mistakes are made, and attacks take advantage of those gaps. Research on expert advice about what security practices to follow shows that experts do not agree on the most important.71
Resilience is more difficult to retrofit into existing systems as an afterthought. It requires thoughtful architectural design from the outset, considering factors such as graceful degradation and partitioning of mission-critical functions to minimize the impact of breaches. The goal is for systems to operate securely, albeit in degraded fashion, even when some components are compromised. This attribute is increasingly important as large-scale systems are interconnected into even larger-scale systems, with the larger goal of organizational resilience and the ability to operate essential business functions even when systems are impaired.72
The inherent uncertainty of cyberattacks often leads companies to delay investments despite the broader societal benefits of resilient systems. Therefore, integrating resilience into the initial design phase is far more effective than attempting to retrofit it, highlighting the importance of preparedness and proactive planning. In addition to the benefit of training, resilient design is still a matter of active research, especially regarding measuring the resilience properties of a composed system.
As cyber threats evolve, the need for resilience in maintaining operational continuity becomes increasingly urgent. Resilience is the key to ensuring that vital services, such as health care, finance, and critical infrastructure, can withstand and recover from attacks. This issue is paramount to the broader community, including businesses, consumers, and governmental bodies, as it directly impacts economic stability, public safety, and national security.
Potential approaches to enhancing resilience include both technical and organizational strategies. Again, as mentioned earlier, adopting architectural choices that support resilience in designs, including for distributed systems, as well as architecting for minimal trusted computing bases, can be beneficial. Formal or semi-formal verification of design and implementation can ensure that critical systems meet high standards of resilience.
On an organizational level, fostering a culture of preparedness and operational excellence is essential. This includes thorough risk assessments, continuous improvement
___________________
71 E.M. Redmiles, N. Warford, A. Jayanti, A. Koneru, S. Kross, M. Morales, R. Stevens, and M.L. Mazurek, 2020, “A Comprehensive Quality Evaluation of Security and Privacy Advice on the Web,” Proceedings of the 29th USENIX Security Symposium, August 12–14, https://www.usenix.org/system/files/sec20-redmiles.pdf.
72 CISA, “Secure by Design,” https://www.cisa.gov/securebydesign, accessed February 6, 2025.
of security configurations, and minimizing the “blast radius” of potential attacks by effectively partitioning critical and non-critical functions. Information-sharing models, akin to those used by the Federal Aviation Administration or anti-spam initiatives, can enhance collective resilience by disseminating threat intelligence and best practices across organizations. Finally, preparation requires deliberate practice to ensure that people and plans are effective, agile, and ready to respond to real-world incidents.
Advancing resilience is a collective effort that involves multiple stakeholders. Government agencies play a crucial role by setting standards and incentivizing resilient design, while industry groups can contribute by developing and promoting best practices. Companies, especially those operating critical infrastructure, have to make resilience a priority in their design and operational processes. It is important that resilience strategies be practical, scalable, and widely adopted.73
Various indicators can be used to “measure” success in designing for resilience. For example, these include the system’s ability to maintain functionality during attacks, local and network outages and natural disasters, the speed and effectiveness of recovery processes, and the overall reduction in the impact of cyber incidents. Regular testing, simulation of attack scenarios, and continuous improvement based on feedback and threat intelligence will help gauge progress. Ultimately, a resilient cybersecurity posture will not only mitigate the damage from attacks but also instill greater confidence in the security and reliability of all systems.
This is cyber hard problems 1, 2, and 3, as well as its effect on problem 10.
Growing software, system, and network size, complexity, and usage offer attackers increasing opportunities for both successful penetration (i.e., larger attack surface) and the ability to remain undetected and operate within the compromised environment (i.e., larger persistence volume). The former concern is generally addressed through secure system design and implementation, while the latter is addressed by intrusion detection and digital forensics. Although some progress has been made in software and system hardening (at least against certain classes of easier-to-exploit vulnerabilities), it appears that the dwell time of non-ransomware-focused sophisticated attackers remains high,74 despite significant investment in the collection, monitoring, and analysis of security-relevant events. Essentially, the duality of detection and evasion in cybersecurity
___________________
73 The Global Resilience Federation, a nonprofit, offers framework concepts for operational resilience for business. See Global Resilience Federation, “The Operational Resilience Framework,” https://www.grf.org/orf, accessed February 6, 2025.
74 The committee excludes ransomware because it inherently exhibits a very obvious and “noisy” behavior relatively soon after infection. Examples of more recent, long-dwell threat actors include TRIANGULATION and Volt Typhoon.
continues to trend in ways favorable to the sufficiently motivated and well-resourced attacker.
As a result, successful defense and remediation requires timely identification and disruption of malicious activities, whether proactively or reactively. The current state of practice relegates defenders to playing “whack-a-mole” while peering through a keyhole, with in-band network management tools that are potentially influenced by the attacker. The current cybersecurity tools and practices are not sufficiently precise to reliably identify the activities of slow-and-stealthy attackers over extended periods of time, nor efficient or fast enough to identify and stop rapid-moving attacks. Furthermore, a lot of emphasis has been placed on exfiltration detection, with significantly less on detecting other types of cyberattacks such as scheduled system-level denial of service.
Several technical factors contribute to confounding the ability of defenders to achieve a sufficient level of situational awareness. These include, but are not limited to, (1) improved threat actor tactics (low “signal”) that increasingly take advantage of native features and resources75 of the targeted environment; (2) high volume of benign system events (high “noise”) as a function of system size and complexity; (3) high volume of low-sophistication attack events (high “background radiation”) that lead to alert fatigue and misprioritization of response resources; (4) more complex and diverse system capabilities that, at least from an observability perspective, partially overlap with attacker capabilities, objectives, and behaviors (e.g., built-in screen recording, system-wide document search); (5) continued reliance on human-driven analysis (threat hunting), with the corresponding limitations on volume and pace of analysis; (6) the inability of analytics to keep up with ever-growing, security-driven telemetry data volumes; and (7) diminishing returns (low “gain”) but high, continuous, fixed cost for any additional type of telemetry collected and used, due to the rich set of pathways attackers can exploit to meet a given objective. Sociotechnical factors that also negatively contribute to the problem include the high cost for security-data storage and processing, the friction of information and data sharing and analysis across intra-organizational boundaries, lack of trained personnel, and (a sometimes real) conflict with other legal, privacy, or regulatory requirements.
No single solution appears sufficiently powerful to fully or substantively address the detection problem on its own. However, with appropriate investment for further scientific investigation, the following practices and technologies could play a positive role in addressing the problem:
___________________
75 This practice is often referred to as “living off the land.”
76 Such partitioning appears to offer several security benefits, at the potential cost of overall complexity.
Most of the above practices and technologies would be significantly aided by the availability of open, high-fidelity, experimental test beds. These would need to go beyond the typical goal of offering representative topologies, systems, and software to include realistic (ideally real) background data and activity sufficient to simulate actual environments and scenarios.
___________________
77 DARPA, 2014, “Transparent Computing,” https://www.darpa.mil/research/programs/transparent-computing.
78 MITRE, “ATT&CK,” https://attack.mitre.org, accessed February 6, 2025.
The fundamental and second-order metrics for success79 remain relevant and appropriate for evaluating individual technologies and, in some cases, combinations thereof. The primary challenges that need to be addressed are (1) providing higher-quality and unbiased evaluation of these technologies80 that goes beyond sparse, anecdotal empirical testing (e.g., human-driven red teaming) and (2) translating specific-technology in-lab effectiveness measurements to real-world impact. With respect to the latter, when better and consistent metrics of effectiveness become the norm, sharing of system and network security architecture patterns along with measurements would go a long way toward establishing a proper engineering discipline in this space.
Accurately determining the initial vector and subsequent impact of a cyberattack has always been a time-consuming and difficult task. A complete analysis would identify several actionable aspects of the attack, including the method of compromise; the software, systems, and users through which initial infection occurred; the full set of systems and data accessed, exfiltrated, or modified by the attacker; and any new software introduced, existing software modified, configuration changes made, or upstream and downstream services accessed during the attack—while at the same time filtering out the typically much larger volume of benign, legitimate activities that may be overlapping and interleaved with attacker activities in both time and space (i.e., in the same systems during the same time period). These are necessary for determining how an attacker was able to gain initial access (to prevent reinfection), what data were lost (e.g., to determine what intellectual property or customer personally identifiable information [PII] was stolen), what assets (e.g., critical infrastructure components) were tampered with, whether the attacker has been completely evicted, what latent access vectors an attacker may have introduced (again, to prevent reinfection), and what residual risk must be dealt with (and potentially through what methods). In many cases, the full extent of the damage incurred is often revealed only after significant time has elapsed since the initiation or even the discovery of the attack. Relying on reported extrinsic observables (e.g., reported financial fraud, identified damage on devices, or cyber-physical processes) negates much of the potential for timely intervention and prevention (or at least minimization) of said damage. For high-stakes events, teams of specialist forensic analysts must work manually over several weeks or months to produce an impact assessment. In the meantime, critical systems may remain exposed or even knowingly left compromised to avoid service disruption.
___________________
79 An incomplete list includes false-positive rate, false-negative rate, accuracy, precision, recall, mean-time-to-detection, and mean-time-to-remediation.
80 A fundamental limitation in the evaluation of almost all detection technologies remains the determination of false-negative rates.
A related problem is determining the set of actions necessary to restore the integrity and trustworthiness of a system or network (along with the relevant data) after a compromise. To the extent that recovery is driven by damage assessment, there is an obvious dependency. Although one could theoretically imagine a fully agnostic system and network reconstitution (e.g., a full data and system recovery from a combination of full backup and reinstallation), several factors make this impractical at scale. These include critical external dependencies (e.g., credentials for external services), system and business availability constraints, and friction related to legacy or embedded devices (e.g., out of support devices) and failure of parts under stress (e.g., network saturation due to recovery traffic). Perhaps the biggest issue is the uncertainty in how far back to recover from,81 especially as it pertains to data. As software systems become increasingly interdependent in both direct (e.g., cloud-enabled multi-device synchronization) and subtle ways (e.g., credential caching), the traditional fallback approach of reformatting and reinstalling becomes both untenable and insufficient.
The size and continuous piecemeal evolution of software, systems, and networks inhibits a sufficiently detailed understanding of their composition and functionality (even under attack-free conditions), which is a necessary step to identifying the aspects of system operation that were (or could be) accessed or tampered by an attacker. Combined with the inherent stealthiness of attacker activities, a timely reconstruction of a reasonably complete timeline of said activities and relevant system assets is currently infeasible except in limited situations. The telemetry or logging necessary to achieve the necessary degree of visibility, strongly correlated with but potentially more detailed than that needed for attack detection, can be cost- and performance-prohibitive to collect, store, and analyze in a timely fashion, even putting aside concerns about the integrity and reliability of the telemetry data in the presence of a sophisticated attacker. Furthermore, the need to restore or maintain system operations practically limits the time and resources that can be committed to the assessment analysis. In the (typically informal) risk analysis that drives the relevant parameters for the system recovery (i.e., which systems and data, how far back), this biases toward a focus on minimization of attack footprint, allowing for undiscovered latent access and other leave-behind artifacts introduced by the attacker. At the very least, it is important to capture in the final after-action report any specific assumptions made relative to the conclusions. For example, if the system was restored from a checkpoint created on a certain date, the inherent assumption is that the compromise occurred after that date.
In terms of potential solutions in the space of damage assessment, forensic reconstruction of attacker activities would benefit from the same type of solution as is needed for attack detection (see the section above on situational awareness), albeit with the
___________________
81 Using an older backup is less likely to contain attacker artifacts, at the cost of lost data.
need for higher fidelity. Other relevant knowledge and capability gaps that need to be addressed include the following:
Several metrics can be used to gauge progress in this space. System-wide, the goal is to reduce mean time to second compromise and mean time to system restoration, and to achieve higher completeness in attack path reconstruction relative to ground truth (potentially in the context of red team–based evaluations, where ground truth can be made available).
This is a core aspect of cyber hard problem 10.
New technology is bringing new cyber hard problems. A prime example is AI, another is CPS. The challenge in securing AI applications and CPS is a core contributor to cyber hard problem 9 but also affects 1, 2, 3, 4, and 10. It also profoundly affects cyber hard problem 7.
It is difficult to assure the security of AI applications, particularly for design patterns that leverage generative AI. While the following discussion highlights AI-specific security challenges, it is important to note that AI systems are themselves software systems and thus susceptible to the full range of traditional cyberattacks.
The adoption of AI as an integral component in modern applications has been among the most disruptive innovations in computing this century. Many of the largest software companies have transitioned to using generative AI, as has become evident in public statements by Microsoft, Google, Meta, Salesforce, and others. Although traditional application security principles—when appropriately adopted—can safely accommodate the inclusion of AI components in software systems, there are unique attributes of AI that make securing forthcoming AI applications a hard problem.82
At the component level, both predictive and generative ML models are “non-smooth” systems that may produce very different outputs for similar inputs. Generative AI models are stochastic systems that can produce different inputs for the same input. Their non-smooth and sometimes stochastic nature may present a reliability challenge when using AI as a component in a repeatable system since their function cannot be formally guaranteed, nor behavior be fully characterized. Since the models themselves are not readily interpretable, this makes their safety and security difficult to assure. Remediation in AI components is difficult since the weaknesses which arise from training cannot be patched directly in code as it might in a traditional software component.
Applications using generative large language models (LLMs) typify several AI challenges. In a basic AI chatbot application, the user interacts with an LLM that iteratively predicts the next token (word chunk) from a growing input consisting of the original system instructions, user input, and previously predicted tokens. The initial and subsequent set of predictions is heavily influenced by the system instructions, which are designed to guide toward—but cannot robustly guarantee—predictions conforming to a preferred style or topic. Because LMs are instruction-following machines, attackers may attempt either indirectly or directly to lead the application away from the intended use. This can be especially problematic in agentic systems, in which the LLM output is connected to services that act on behalf of an (untrusted) user or respond to context fetched from external (untrusted) sources by the agent components.
While AI systems are fundamentally software systems, their characteristics—supply chains that include data sets and training code and runtime nondeterminism and non-smoothness—necessitate new approaches to risk assessment and vendor trust evaluation.
___________________
82 A. Vassilev, A. Oprea, A. Fordyce, and H. Anderson, 2024, “Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations,” NIST Computer Security Resource Center, January, https://csrc.nist.gov/pubs/ai/100/2/e2023/final.
To that end, many emerging regulations have begun to specifically call for “AI red teaming” requirements. Regulations in the European Union83 and proposed regulations in the United Kingdom84 legislate requirements for model assessment, with an emphasis on safety and societal harms. However, these requirements still lack acceptable standards across the industry in what should be assessed, what are acceptable assessment outcomes, and how and to whom to disseminate the results of an assessment.
Although there are fundamentally still software systems that include software and third-party services in applications, the supply chain of AI applications also includes data and third-party pre-trained or fine-tuned models. In addition to the possibility that attackers may develop model or data deserialization-based file formats (e.g., pytorch, pickle, and numpy) to execute arbitrary code,85 the possibility exists that models may contain backdoor functionality encoded in the model’s architecture or model weights. Technology and processes to measure and mitigate risk in these supply-chain components are nascent. Specific challenges in supply chain include (see more at NIST Adversarial ML Taxonomy86) the following:
A key challenge for auditability in the AI supply chain is that there is not yet a standard for reporting the equivalent of an SBOM, although efforts to address this have emerged.87 The addition of model and data components can be accommodated by SBOM to include traditional static elements of component identification, dependency information, licensing, and versions. But for AI models, the behavioral reports should also be included that report on potentially risky runtime behaviors that have been
___________________
83 European Commission, 2024, “AI Act,” https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai.
84 Department for Science, Innovation and Technology and the Office for Artificial Intelligence, 2023, “AI Regulation: A Pro-Innovation Approach,” March 29, https://www.gov.uk/government/publications/ai-regulation-a-pro-innovation-approach.
85 Common Weakness Enumeration, “CWE-52: Deserialization of Untrusted Data,” https://cwe.mitre.org/data/definitions/502.html, accessed February 6, 2025.
86 A. Vassilev, A. Oprea, A. Fordyce, and H. Anderson, 2024, “Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations,” NIST AI 100-2 E2023, January, https://csrc.nist.gov/pubs/ai/100/2/e2023/final.
87 J. Bressers, 2023, “SBOM Everywhere and the Security Tooling Working Group: Providing the Best Security Tools for Open Source Developers,” Open Source Security Foundation (blog), June 30, https://openssf.org/blog/2023/06/30/sbom-everywhere-and-the-security-tooling-working-group-providing-the-best-security-tools-for-open-source-developers.
observed. Unfortunately, unlike a binary set of attributes or functions, the set of risky model behaviors may be incomplete, imprecise, and less actionable than in traditional software. Thus, it is still important to employ third-party audits and third-party guardrails to discover and control runtime behavior.
Defenders are unprepared for a dramatic increase in scale and complexity of cyber operations from offensive AI tools—when attackers leverage AI for traditional cybersecurity operations. The risks presented by using AI for offensive purposes are offset at least to some degree by the potential for defenders to leverage AI to implement compensatory security controls and mitigations, but these are not addressed here.
As highlighted in the National Security Commission on Artificial Intelligence Final Report,88 digital infrastructure may be increasingly indefensible against escalating, offensive, AI-enabled cyber capabilities without offsetting defensive controls. Threat actors are beginning to leverage AI for various malicious use cases, including offensive copilots, scaling social engineering attacks, and enhancing offensive operations.
Offensive AI is still nascent, but researchers are developing AI for various offensive purposes that will challenge defensive systems and processes. AI-driven offensive capabilities can increase the potency and speed of cyber campaigns and present significant threats to both digital infrastructure and human targets.
AI can expedite traditional cyber campaigns against digital infrastructure in several ways. For example, using LLMs, attackers can expedite the discovery, development, and delivery of exploits through automated code reversing, vulnerability discovery, and instrumentation of exploits for vulnerabilities. AI systems that reduce the time required for threat actors to execute attacks by automating labor-intensive tasks represent a sort of “offensive copilot” that can decrease the time to impact in cyber operations.
AI-powered tools can also assist attackers in more rapid maneuvering during hands-on parts of offensive campaigns to scale offensive operations. By integrating generative AI agentic frameworks with existing tools, attackers can orchestrate complex operations that cover large portions of an attack life cycle in a way that was not previously possible.
The impact on human targets using AI presents a formidable challenge. Disinformation campaigns that leverage deepfakes have already become part of public awareness due to several incidents involving elections89 and digital warfare that now requires
___________________
88 E. Schmidt, R. Work, S. Catz, E. Horvitz, S. Chien, A. Jassy, M. Clyburn, et al., 2021, Final Report, National Security Commission on Artificial Intelligence, released March 1, https://reports.nscai.gov/final-report.
89 E. Sayegh, 2024, “The Battle for Truth in Election Seasons: AI-Generated Deepfakes,” Forbes, May 14, https://www.forbes.com/sites/emilsayegh/2024/05/14/the-battle-for-truth-in-election-seasons-ai-generated-deepfakes.
news consumers to question the validity of reports.90 Since bad news tends to travel faster than good news, correcting disinformation is an asymmetric challenge.
These tools can also be used for fraud. Highly realistic and interactive social engineering attacks for fraud are now possible with generative AI. In this setting, attackers can create convincing impersonations or scenarios to manipulate individuals in a way that feels customized.91 AI’s potential to scale such attacks is a developing threat vector, where generative AI can create deepfakes and other convincing forms of fake identities for automated and interactive phishing or scamming operations.
The human challenges that this presents have been called out in other cyber hard problems. The key ingredient that AI brings is the sophistication and potential for scale. While fundamental security practices can ward off many of these attacks, the increased scale and sophistication allows attackers and fraudsters to affect a much broader set of victims. In a setting of fixed resource constraints of defenders, remediation and response can become intractable.
Many cyber and cyber-enabled systems include a data component, either creating new data, processing existing data, or transmitting data to achieve a particular purpose. The designers of the system or application build the service with specific security and privacy properties to mitigate the occurrence and impact of adverse events—that is, uses of the data that go beyond the intended purpose. The user of the system, and relevant regulatory or law enforcement entities, desire the ability to hold the data steward accountable for upholding the properties as promised while also ensuring that unexpected uses of the data are not possible (i.e., “the software does what it says with the data; no more, no less”).
In the absence of a solution to this problem, there is little choice but to trust that data are collected, used, and stored appropriately without much assurance. Prior to deployment or adoption, there is a requirement to convince the user and relevant authorities that the promised properties are sufficient and correctly implemented. Post-deployment, the data steward may need to modify the data use or protection terms and need to re-consent the data subject or owner, updating the presentation of the use and proposed protections and accurately recording the update.
___________________
90 D. Klepper, 2023, “Fake Babies, Real Horror: Deepfakes from the Gaza War Increase Fear About AI’s Power to Mislead,” Associated Press, November 28, https://apnews.com/article/artificial-intelligence-hamas-israel-misinformation-ai-gaza-a1bb303b637ffbbb9cbc3aa1e000db47.
91 H. Chen and K. Magramo, 2024, “Finance Worker Pays Out $25 Million After Video Call with Deepfake ‘Chief Financial Officer,’” CNN World, February 4, https://www.cnn.com/2024/02/04/asia/deepfake-cfo-scam-hong-kong-intl-hnk/index.html.
Retrospectively, there is a requirement to be able to determine as much as possible about what went wrong (e.g., whether the security properties were inadequate, if the security properties were incorrectly implemented, or if the system were modified in some way that impacted the security properties).
A sizable portion of technical innovation is rooted in advancing the state of the art of what can be done with data, yet the technical mechanisms for setting and enforcing policies throughout the data life cycle (e.g., data at rest, in transit, and in use) have not kept pace. It is extremely time-consuming and difficult to identify, mitigate, and prevent the misuse of data without policy-aware data systems. In the past, we accepted possession and access to data as a proxy for permission to use the data. The proliferation of devices that collect data; the inherent complexity of the software, hardware, and network ecosystem; and also the ease with which data can be transmitted to another party make fine-grained control over the use of data untenable for the future. The misuse of data and our inability to make verifiable claims about how data will be used degrades trust in IT systems and hampers future innovations.
Making progress on this hard problem will support better outcomes on avoiding adverse outcomes for end users related to the misuse of data, increasing trustworthiness of personal devices, and perhaps decreasing disinformation. Large-scale change is needed to evolve IT systems to be policy-aware when processing data. At a minimum, such change requires the following:
One open question is whether the desired policies are actually expressible (i.e., what kinds of policies are expressible and enforceable, and are these what people care about?). Although this has been tackled in traditional IAM systems (see above), it has not been done for shared data.
The emergence of cyber-mediated, human-targeted attacks of various sorts has a history of being considered cybersecurity or privacy problems. Examples include inducing the receiver of a malicious email to download and run an attachment, click on a link (to deliver malware), clinking on a link and type things in (to steal identity), and the recognition of stalkerware as a category and problem. Targeting individuals through spear phishing and catfishing are recognized cybersecurity attacks.
Although propaganda, disinformation, and military deception have a long history, technology-enabled creation and dissemination of disinformation is a newer and growing problem. Everything on the computer is mediated by technology. The Internet and World Wide Web, along with social media, expands the reach of disinformation. Automation and AI expand the scale and precision of disinformation, to bots, deepfakes, and written and spoken text that can increasingly mimic anyone trustworthy. In parallel, “broadcast” sources of journalism (television, radio, newspapers) are being replaced with peer-to-peer communications with poor or missing authentication. What people see is determined in whole or in part by algorithms that (typically) optimize for engagement.
Per the Verizon 2024 Data Breach Investigations Report,92 deepfake-like technology has already been used in many reported cases of fraud and misinformation. As generative AI only increases in abilities, scope, and scale, AI-generated fakes as weapons of disinformation will move beyond “deepfake” pictures of humans and audio fakes, to more complex scenes compellingly attesting to events that never occurred, compelling quotes, speeches, and “fake news” articles, and full videos. Generative AI that undermines artists of all professions today can become tomorrow’s tools of disinformation.
Technology-enabled disinformation can be used to undermine individual reputations and emotional well being (e.g., deepfake revenge porn), create conspiracy theories
___________________
92 Verizon, 2024, 2024 Data Breach Investigations Report.
targeted at public figures, and attack core democratic processes such as elections. Much of the U.S. economy relies on the reputations of the strength of businesses and financial infrastructure, making this a potential weapon against U.S. economic stability.
What makes this a hard problem is the following:
Content on the Internet (and later curated as part of another data set or AI function) often does not come with any trustworthy indication of the source or provenance of that information.
Much of the information people receive comes through someone else, either directly or through a communication artifact (e.g., books, newspapers). All the information received through computer interactions is the latter. Even a video call is intermediated by sophisticated software that can change backgrounds and faces. Information comes to consumers from or through a source, and their reaction to that information is potentially colored by knowledge of that source, from news to education to books, from religion to civics to politics. The reaction may be to the identity of the source, such as an individual (Walter Cronkite) or an organization (Fox News), or the reaction may be to the process and assumptions around the source’s communication type (autobiography, medical advice).
The pseudonymity promised by “no one know[ing] you’re a dog” is rapidly extending to all contexts on the Internet. “Fake news” is shared by people you know, and anyone can stand up a website claiming to be a news or publishing source. Identity, identity attributes, and source creation context are all at risk of being inaccurately relayed or assumed. Immersive virtual environments make alternate realities the norm.
Civil and societal institutions rely on some shared understanding of the authoritativeness of various kinds of information. Examples include news about communities, states, and nations, or results of the electoral process. The stability and safety of people’s economic supports and investments rely on reliable information about them.
Building blocks for enabling a more trustworthy information ecosystem might include digital signatures (including source devices that apply signatures at the point of capture), imperceptible signals in media streams (watermarking), and widespread and reliable conveyance of provenance information through social media channels.93 An alternate approach is centralized or decentralized fact checking and “community notes.”
What makes this a hard problem is the following:
___________________
93 Coalition for Content Provenance and Authenticity, “Overview,” https://c2pa.org, accessed February 6, 2025.
Many of the overarching cyber challenges described earlier are expressly amplified by missing or misaligned policies. It is difficult to design effective policy for complex systems that does not increase cost disproportionately to its benefit. Resilient system design and operations, even when properly guided by policy, can slow progress. Competing interests often retard policy solutions even when candidate solutions exist. Jurisdictional questions, including globalization and failures to provide federal preemption, further complicate policy solutions and effective remedies that would apply to providers and users in a single legal jurisdiction.94
Regulatory policy and economic incentives can be confounded by competing policy goals. For example, the desire for rapid problem identification can often be achieved by comprehensive authentication; however, this can often interfere with users’ privacy. Furthermore, policy that encourages disclosure in support of principled risk assessment can threaten providers’ intellectual property if done carelessly.
There has been policy progress that has helped ensure vulnerability and breach disclosure, but this is a fairly crude measure of resilience and safety.
The lack of effective policy (economic and regulatory) is one of the most dogged and influential of hard problems.
Policy hard problems need to be addressed by laws, policies, regulations. As described in the committee’s overarching problem framework, they affect almost all cyber hard problems.
Misaligned incentives in cybersecurity are a significant challenge, manifesting in the varied and often conflicting priorities of stakeholders such as vendors, consumers, insurers, and regulators.
___________________
94 Harmonizing policies across jurisdictions (state, federal, and international) is a super-hard problem.
Vendors prioritize speed to market and cost efficiency over security and say additional security measures would slow the pace of innovation, while consumers often choose products based on price rather than security features. Insurers, who have the potential to influence better security practices through underwriting conditions, struggle with accurately assessing risks and enforcing effective mitigations. Competitors, despite facing similar threats, are often unwilling to share valuable threat intelligence, undermining collective defense efforts. This misalignment results in suboptimal decisions that increase overall vulnerability and delay the benefits of addressing other cyber hard problems. Without progress on incentives, benefits from solving the other cyber hard problems will be disadvantaged or delayed.
Solving the issue of misaligned incentives is crucial for enhancing the overall security posture of the digital ecosystem. It matters to a wide array of stakeholders, including businesses that suffer financial losses from breaches, consumers whose personal information is compromised, and national and homeland security agencies tasked with protecting critical infrastructure. The economic impact of cyber incidents is substantial, with costs extending beyond immediate financial losses to include reputational damage, loss of consumer trust, and long-term recovery expenses. However, to date this has been inadequate to spur changes needed. Therefore, realigning incentives to promote better security practices is essential for reducing these risks and enhancing resilience against cyber threats.
The difficulty in addressing misaligned incentives stems from several factors. Economic and competitive pressures often discourage businesses from investing adequately in cybersecurity, as the benefits are not always immediately observable.95 The tendency to prioritize short-term gains over long-term security investments is pervasive, and the lack of standardized metrics for measuring cybersecurity return on investment or cyber-coverage quality complicates decision making. Organizations may sometimes feel that the most cost-effective method for limiting damage for faulty products is public relations, especially for categories of weaknesses that are not readily assessed or repaired. Regulatory and policy efforts to realign incentives have been slow and fragmented, with various proposals such as grants, tax incentives, and liability considerations failing to achieve widespread implementation. The complexity of the cyber threat landscape and the rapid evolution of attack techniques further exacerbate these challenges.
Potential approaches to realigning incentives include policy reforms and innovative economic models. Governments can play a pivotal role by introducing and enforcing regulations that mandate minimum security standards, introducing “safe havens”
___________________
95 This is related to myopic loss aversion. See R.H. Thaler, A. Tversky, D. Kahneman, and A. Schwartz, 1997, “The Effect of Myopia and Loss Aversion on Risk Taking: An Experimental Test,” The Quarterly Journal of Economics 112(2):647–661.
for good faith efforts accompanied by product design transparency, and by offering tax incentives or subsidies for businesses that invest in robust cybersecurity measures. Public–private partnerships can facilitate better information sharing and collective defense initiatives. An example of a public–private partnership that does such a thing successfully is the National Cyber-Forensics and Training Alliance, which brings together the business sector and law enforcement to disrupt cybercrime. Additionally, developing standardized metrics for assessing cybersecurity investments and outcomes can help businesses make more informed decisions. The implementation of mechanisms like the U.S. Cyber Trust Mark, which provides a recognizable standard of cybersecurity for consumers of wireless IoT devices, is a step in the right direction.
Those who can take action to realign incentives span across sectors. Policy makers and regulators can introduce and enforce laws that require graded security standards, depending on the kind of device and its use environment, and incentivize compliance. For example, autos that can be easily stolen because their keyfobs use weak cryptography and pervasive back doors in network-connected devices seem like areas that need to be addressed. Absent this, industry leaders and business executives will not prioritize cybersecurity as a critical component of their operational strategy and allocate appropriate resources. Insurers can refine their risk assessment models and offer premium reductions for policyholders that adopt best practices. Consumers can influence the market by demanding more secure products and services. Additionally, cybersecurity researchers and advocacy groups can continue to highlight the importance of aligned incentives and drive awareness.
Success in realigning incentives can be measured through several indicators. A notable decrease in the frequency and severity of cyber incidents would suggest that stakeholders are making more security-conscious decisions. Increased investment in cybersecurity by businesses, moving closer to the recommended 10 percent of budgets, would also be a positive sign. Ultimately, success will be reflected in a more resilient and secure digital ecosystem where the costs and benefits of cybersecurity investments are better aligned across all stakeholders.
An often-overlooked consequence of technology’s spread is the difficulty that organizations and individuals have in securing it. Originally described in 2011,96 the “security poverty line” is a concept that delineates the “haves” from the “have nots”: whether it is economically or technically feasible to implement what is generally assumed to be effective security, given real-world conditions. Just as with economic poverty, cybersecurity
___________________
96 W. Nather, 2011, “T1R Insight: Living Below the Security Poverty Line,” 451 Research, May 26, https://web.archive.org/web/20140203193523/https:/451research.com/t1r-insight-living-below-the-security-poverty-line.
poverty results from many complex dynamics and factors. This problem exacerbates the effects of cyber hard problems 1, 2, and 4.
Because there is no simple prescriptive blueprint for building secure systems, some have tried to measure effective mitigation of carefully scoped attack scenarios, such as MITRE’s Engenuity evaluations97; others have tried to calculate the projected cost of security technology according to a given compliance framework98 or simply following security professionals’ recommendations.99 Although peer benchmarking and trends reports describe how much some organizations spend on cybersecurity, the reports do not address whether the spending is effective or appropriate. Spending formulas, such as the percentage of IT budget, do not necessarily scale up or down, nor do they have any link to positive or negative outcomes. Some increasingly critical controls (e.g., logging) are not included in the minimum baseline edition of products but are premium priced.
Another confounding factor for organizations is expertise. Cybersecurity expertise is not simply education or training; it also includes the experience of securing new technology and diagnosing and responding to new vulnerabilities and attacks. Organizations find themselves competing for this expertise against the deeper pockets of security providers (according to Glassdoor, the total salary in 2024 for a senior cybersecurity analyst is $156,000–$234,000 per year).
Constraints within the environment also affect an organization’s capability to secure itself. For example, conventional best practice in cybersecurity calls for a system to be designed to fail safe rather than open; this is not an option in a safety-focused sector such as health care, where medical staff may never be barred access to equipment or data needed to treat patients in an emergency. Software that integrates with hundreds of different systems under a variety of countries’ regulatory environments can take months or years to update. Onsite upgrades for thousands of point-of-sale systems mean that retailers must choose carefully when to incur that downtime and expense, and certainly not during the heaviest shopping times of the year. Every cybersecurity risk framework or practice may need to be adapted substantially to work around these obstacles.
Finally, in an era where cybersecurity controls are spread among third-party providers (see the section “Supply Chain” below), organizations have to rely on the cooperation of other entities with whom they may have little to no legal or commercial influence. With a sufficiently large amount of money at stake or the possibility of negative public relations, a provider may be incentivized to meet the security requirements of a customer, but smaller organizations lacking this kind of influence cannot necessarily receive
___________________
97 MITRE, 2024, “Our ATT&CK Evaluations Methodology,” https://attackevals.mitre-engenuity.org.
98 Center for Internet Security (CIS), 2023, “The Cost of Cyber Defense,” CIS Controls Implementation Group 1, August, https://www.cisecurity.org/insights/white-papers/the-cost-of-cyber-defense-cis-controls-ig1.
99 A. Shimel, 2013, “What Is the Real Cost of Security?” NetworkWorld, April 4, https://www.networkworld.com/article/744780/opensource-subnet-what-is-the-real-cost-of-security.html.
the emergency services they need during an incident, force the timely remediation of an identified vulnerability, or reject provider conditions that may result in increased risk (such as allowing overly broad network access). Regulations will not work unless the regulated parties have access to tools that are affordable (sustainable) and actually reduce risk. Without uniform cybersecurity regulations or other incentives, most small- and medium-sized businesses, nonprofits, and local public-sector entities (including law enforcement) have to make do with the equivalent of security scraps, with support only available piecemeal from managed service providers, from a provider specified by their cyber insurer, or volunteer efforts such as the University of California, Berkeley–led Cybersecurity Clinics.100
Fully organic development shops such as Google and Apple have the advantage of full (internal) transparency in their software code bases. This facilitates direct analysis at scales ranging from lines of code in small components to design choices for APIs and architectural features. This also facilitates comprehensive assured refactoring; for example, updating a service API with potentially hundreds of clients, all incompatibly updated in an atomic action. Additionally, it facilitates a fully explicit linking of design models, implementation artifacts, test cases, analysis tooling, and any supporting elements.101,102
In other words, full transparency facilitates ongoing acceptance evaluation, rapid adaptation, and repairs without creating technical debt (i.e., expedient decisions that would later need to be revised in order to permit continued evolution of a system).
By contrast, large enterprise and mission systems are generally integrated from diversely sourced components and services (“system elements”), some of which are kept opaque to their clients in order to retain competitive advantage, protect sensitive data and algorithms, and enable update and enhancement without unwanted dependencies on (hidden) implementation choices. This means that even when one layer is revealed in a complex system, there can be multiple opaque layers beneath, analogous to “turtles all the way down.”
The integrated systems model poses challenges, however. One set of challenges relates to acceptance evaluation, due to opacity of system elements and uncertainty regarding compatibility of elements. Another set of challenges relates to update and evolution, deriving from compatibility issues as individual elements are on uncorrelated update cycles. (Services, for example, can be updated several times per day, while
___________________
100 Consortium of Cybersecurity Clinics, “Cybersecurity for the Public Good,” Center for Long-Term Cybersecurity, https://cltc.berkeley.edu/program/consortium-of-cybersecurity-clinics, accessed February 6, 2025.
101 “Why Google Stores Billions of Lines of Code in a Single Repository,” posted September 14, 2025, by @scale, YouTube, https://www.youtube.com/watch?v=W71BTkUbdqE.
102 H. Wright, 2019, “Lessons Learned from Large-Scale Refactoring,” 2019 IEEE International Conference on Software Maintenance and Evolution, December 5, https://ieeexplore.ieee.org/document/8919159.
open-source components may be updated every few weeks.) Evolution is also impaired by opacity, and many systems sustainment teams must engage in active reverse engineering to assess and document for repair security vulnerabilities, for example. There are also challenges related to overall systems architecting and design. Architectural decisions focused on reducing coupling, localizing variabilities, enhancing key quality attributes (particularly resiliency), and the like may need to be compromised to support compatibility of APIs, data representations, and service interfaces among elements that are meant to interoperate.
An extreme example is the incorporation of vendor components as original equipment manufacturer elements into integrated systems, such as commercial desktop systems into medical devices such as imaging systems. The end user, and possibly the IT support team, might not have sufficient visibility to be aware of the incorporated desktop as other than part of an appliance, and so that desktop may not, over a period of years, receive necessary updates and security patches. The resulting vulnerabilities have been exploited in ransomware attacks.103,104
An additional consequence of supply chain opacity is hidden dependencies, where a deeply embedded vulnerable system element can trigger disruptions in the event of compromise or, in the case of open source, loss of configuration control. Attacks on embedded supply chain elements can have broad consequences, and so these elements are a favored target by attackers. Examples include Blackbaud, a service provider to financial services and other organizations including critical nonprofits.105 A research report by the Cyentia Institute (a subsidiary of Mastercard) and RiskRecon106 identified ripple effects impacting between 800 and 1,000 downstream organizations. The network security vendor SolarWinds unintentionally delivered a malware payload embedded in a signed system update that was automatically distributed.107 A more recent extended global outage, caused by an automatically deployed update to CrowdStrike security software on Windows systems, affected millions of systems from banks to commercial aviation, health care, and critical infrastructure.
This can be an issue even when the embedded system element is a tiny fragment of code. One example from 2016 is leftpad, which is an 11-line module of code in the
___________________
103 L. Hautala, 2020, “Hospital Devices Exposed to Hacking with Unsupported Operating Systems,” CNET, March 10, https://www.cnet.com/health/medical/hospital-devices-exposed-to-hacking-with-unsupported-operating-systems.
104 C. Van Alstin, 2023, “RSNA 2023: Hospital Imaging Systems May Be Gateways for Ransomware, Expert Warns,” HealthImaging, November 30, https://healthimaging.com/topics/professional-associations/radiology-associations/radiological-society-north-america-rsna/rsna-2023-ransomware-medical-devices.
105 L. Fair, 2024, “FTC Says Blackbaud’s Lax Security Allowed Hacker to Steal Sensitive Data—and That’s Just the Beginning,” Federal Trade Commission (blog), February 1, https://www.ftc.gov/business-guidance/blog/2024/01/ftc-says-blackbauds-lax-security-allowed-hacker-steal-sensitive-data-thats-just-beginning-story.
106 Riskrecon, “New Report: Ripples Across the Risk Surface,” Riskrecon by Mastercard, https://www.riskrecon.com/ripples-across-the-risk-surface, accessed February 6, 2025.
107 L. Fair, 2024, “FTC Says Blackbaud’s Lax Security Allowed Hacker to Steal Sensitive Data.”
million-element open source NPM ecosystem widely used for web applications. The developer of this small component chose to delete it and other elements from the library due to a dispute over names for a software package. The deletion lasted only 2 hours but caused widespread disruption because of its pervasive use deep in the supply chain supplying web applications.108
There are also supply-chain issues in IoT, CPS, and networking infrastructure generally. One example was an attack on small office and home office (SOHO) routers, identified by Black Lotus Labs at Lumen Technology.109 In this case, more than 600,000 routers belonging to a single internet service provider were completely disabled, forcing the entire customer base to have their equipment physically replaced. In other cases, these SOHO routers are not regularly updated or are no longer supported by the vendor so that no security updates are available; the accumulating residue of vulnerabilities makes a perfect platform for attackers to take over infrastructure and use it for botnets or proxying services.
On the one hand, the approach to security vulnerabilities has been to encourage organizations to patch early and often, and preferably automatically. But as these examples show, automatic updates gone wrong can also cause catastrophic events. Victims of attacks and outages are caught in the middle between conflicting imperatives, and the cybersecurity industry owes them a better answer than to say “just patch.”
The recent mandates regarding use of an SBOM can be seen as transforming what is sometimes a full opacity into a kind of “translucency” where some information is provided downstream (i.e., to client users, integrators, and end customers) in order to overcome some of these challenges and, additionally, create some incentives within the supply chain to address security attributes more aggressively. An SBOM, representing something akin to a food ingredient list, can empower organizations to make better procurement decisions, but only when they have feasible alternatives.
This is, of course, a supply-chain cyber hard problem also affecting hard problems 1, 2, and 3.
Liability for faulty code or hardware represents a critical and complex issue in cybersecurity. Under “contracts of adhesion” vendors often sell software “as is,” disclaiming responsibility for defects that may lead even to significant breaches or failures. Cyber systems cannot be evaluated based on a quick inspection (like a vacuum cleaner) or even a diligent inspection. Choices for equivalent functionality in other products, are often
___________________
108 Ibid.
109 Black Lotus Labs, 2024, “The Pumpkin Eclipse,” Lumen, May 30, https://blog.lumen.com/the-pumpkin-eclipse.
very limited and, as discussed, there is a market failure that does not practically enable consumers to “select the model with the security they want.” Except for copyright and patent infringement and designated systems like medical devices and automotive, or software used in other regulated industries, where limitation of liability is circumscribed by law, most software is marketed with the “understanding” that some level of imperfection is acceptable. Thus “legal” remedies are largely ineffective even with expensive litigation. This becomes particularly problematic in cybersecurity where software faces sophisticated, evolving attackers. The challenge is compounded by the displacement of loss onto consumers rather than the companies responsible for the vulnerabilities.
The consequences of faulty code extend far beyond mere inconvenience. Consumers, businesses, and governments all suffer from the fallout of software failures. Establishing liability would incentivize vendors to prioritize security and quality, potentially reducing the frequency and severity of breaches. However, current practices and economic realities pose significant hurdles. Large companies, even after significant breaches, rarely face existential threats, and the costs are often borne by consumers and smaller entities.
Addressing this issue requires overcoming several barriers. First, there is a need for a cultural and operational shift within the software industry. The mantra that “we did the best we could” (even if true) must give way to more rigorous standards and accountability. Introducing liability necessitates robust metrics and frameworks to evaluate software safety and security, akin to those in place for other regulated industries. However, creating these standards is not straightforward. The dynamic nature of software development, coupled with the continuous evolution of threats, makes it difficult to establish a static baseline for safety. Furthermore, the global nature of the software supply chain complicates the assignment of liability, as many stakeholders—from developers to suppliers—are involved in the creation and maintenance of software products.
Potential approaches to this problem include both voluntary and mandatory assessment mechanisms. These standards would need to be continuously reviewed and updated to remain relevant. Another approach is to create a “safe harbor” for vendors who follow best practices, thus incentivizing compliance while recognizing the inherent challenges of achieving absolute security.110
Government agencies can establish and enforce regulatory standards, while industry groups can develop and promote good practices, with incremental adoption. Companies, particularly large enterprises with significant market influence, can lead by example, incorporating security into their development processes and advocating for broader industry changes. Collaboration between public and private sectors is essential
___________________
110 Lawfare (https://www.lawfaremedia.org/topics/cybersecurity-tech) has a number of relevant notes on this topic.
to ensure that standards are practical and effective. Although other domains such as medicine and civil and mechanical engineering have successfully employed professionalization standards, the software field generally does not possess the sort of widely accepted, comprehensive principles as the other fields.111
Success in this endeavor would be indicated by a measurable reduction in the frequency and impact of software-related breaches. Metrics could include the number and severity of vulnerabilities discovered and exploited, time to patch after exploit announcements, the financial and operational damage from breaches, and the rates of compliance under established standards. Ultimately, creating a more secure software ecosystem will require sustained effort and cooperation across the industry, but the potential benefits for all stakeholders make it a goal worth pursuing.
Most cyberattacks rely on communication over the Internet. The Internet, globally and even within countries, is not owned, operated, or controlled by a single legal or technical entity or by a closed set of governments, major corporations, or technology institutions; nor is it subject to a single set of policies. Rather, the Internet is an aggregate whose owners, operators, participants, and technologies function independently of one another but interoperate. As a result, long before they touch their target networks and endpoints, cyber operations traverse and in some cases leverage infrastructure, technology, and services owned, operated, and offered by different Internet infrastructure providers (IIPs). These companies include (without counting computing hardware companies) operating system developers, cybersecurity firms, Internet service providers, mobile telecommunications companies, cloud and virtual private server providers, content delivery networks, Domain Name System service providers, hosting providers, domain registrars, and a variety of Internet technology platforms, such as browser, e-mail, and search platforms.
These companies see themselves as neutral providers of global Internet services but not as critical infrastructure assets with a key role to play in systemic security and resilience. Although most of these companies invest substantially in cybersecurity measures, and occasionally cooperate operationally to degrade specific threats, they lack regulatory, financial, or other incentives to systematically address malicious actors’ use of their technology and services in cyber operations that do not directly and immediately impact them. Their cybersecurity efforts and cross-company cooperation are further hamstrung by a complex legal, privacy, regulatory, antitrust, government, and business environment.
___________________
111 Circumscribed certifications, like network administration, can be useful but they are scarcely comprehensive.
Internet infrastructure companies have unique (and in many cases the only legal) vantage points to observe attacks and detect and frustrate malicious cyber activity on a systematic, nationwide basis (and, for some of the large IIPs, even globally). If these entities were to take on the responsibility of discovering and limiting malicious cyber activities and implement successful regimes to do so, malicious actors’ ability to conduct cyber operations against U.S. targets would be substantially attenuated. Coordination across providers is required because attackers straddle their infrastructure across multiple providers for survivability against takedowns.
There have been several attempts at private-sector coordination (with or without U.S. government participation), primarily at the threat information sharing (TIS) level, over many years. Perhaps the biggest inhibitors to such coordination and threat information sharing revolve around liability concerns and the lack of a business case. While setting industry-wide standards and the exchange of best practices is non-controversial, the (at best) federated nature of the infrastructure landscape means that operational coordination that leads to systemic action (e.g., coordinated takedowns of malicious infrastructure) is the exception rather than the norm, and only occurs as a knee-jerk reaction to high-visibility events. Therefore, in practice, most coordination has taken the form of TIS. This need not be a handicap, if the information exchanged is accurate, timely, and conveys sufficient context to provide the necessary confidence so each participant can take action. Unfortunately, that is typically not the case, and critical information is often missing due to lack of collaboration across key industry verticals.
One of the potential liability concerns expressed is that TIS represents or can lead to collusion between companies (creating the perception of collusion to the public and to regulators), leading to privacy and antitrust concerns (and associated litigation). Making the government part of any threat information sharing arrangement focused solely on technical factors is one possible approach, but for the global firms, there will also be concerns on other countries’ reactions to such direct U.S. government participation. The most common liability concern comes from actions following high-profile attacks. All such incidents have been accompanied with various types of litigation, including litigation by different agencies of the U.S. government. For companies, the most damaging form of litigation alleges “willful neglect,” meaning the commercial entity knew or should have known about the specific problem but failed to act. This creates a perverse incentive for a deliberate lack of knowledge in some area as a cost-effective form of liability protection. There will be no long-term continuous incentive for TIS or any other meaningful collaboration without addressing these concerns.
There are at least four different concerns over liability: customer and third-party impacts; notification shortfalls; regulatory fines; and civil and class-action lawsuits. In many cases, this involves different agencies such as the Department of Justice, the
Securities and Exchange Commission, the Federal Trade Commission, the Department of Homeland Security Cybersecurity and Infrastructure Security Agency (CISA); and, in certain cases, the Department of Defense. While CISA has made progress with information sharing, the fear of liability still exists because it is not possible for CISA to fully protect companies reporting a breach from litigious actions taken by other government agencies due to different authorities and regulations. In addition, CISA cannot protect companies from civil lawsuits from other companies, customers, and third parties.
Accordingly, and starting from a recognition that its greatest strength is as a convener for collaboration on a technical level, the U.S. government could provide a safe, non-litigious forum for technical collaboration and data sharing without fear of liability from both the U.S. government and private industry, perhaps after the model of the Information Sharing and Analysis Centers.112 Such a forum would create potential for end-to-end visibility across the domestic Internet infrastructure. Such a forum could be coordinated with the U.S. Cyber Command, and potentially be tipped by the Intelligence Community—but, critically, it should not be used as a source for data by the latter. To the extent that technologies for private TIS exist or can be developed, they will play a significant role in countering narratives of collusion and negative public perception.
One option is to require all companies over a certain size to have cybersecurity insurance—and allow companies to work together to bundle end-to-end coverage to further sharing of information. This also creates the incentive for companies to follow standards and adopt technologies for lower insurance rates. Working through the insurance providers may be a more tractable proposition. However, cyber insurance coverage is very limited and for all the reasons mentioned earlier, insurance companies themselves are in no position to judge the security of the systems they insure. Careful mechanism and incentive design are needed to avoid simply transferring the intractable problem from the company to an insurance firm.
Individuals and organizations have a legal right to pursue those who violate their service agreements with civil courts. Creating a special cyber court and providing it with the necessary technical resources to fully pursue cyber criminals and threat actors (even nation-state sponsored ones, to the extent that it is not desirable to treat these as an act of war) may be appropriate given the technical understanding required by the U.S. government and all parties.
Another option is to create a “cyber fire department,” with broad authority to act on third-party, including private cyberinfrastructure. This could be operated by the U.S. government, by contracted private entities, or in a federated or localized manner to reflect constraints of specific sectors, geographical areas, or other considerations. It is
___________________
112 National Council of ISACs, “Information Sharing and Analysis Centers (ISACs),” https://www.nationalisacs.org, accessed February 6, 2025.
almost certain that significant new authorities conferred by legislation (including limited liability waiver) would have to be granted to such an entity.
Ultimately, if the United States wishes for the Internet infrastructure providers to play better defense, it will have to either create the right financial incentives for markets to value security more than they currently do, or directly pay for such better practices. Tax credits, bounties, fees on connectivity bills, security investment programs through the Small Business Administration, and subsidized cyber insurance (combined with heightened terms and conditions for such policies) are only some of the ways such financial support can be extended.