Previous Chapter: 3 Cyber Hard Problems
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

4
The Producer Perspective

Chapter 3 lists cyber hard problems from what might be called the point of view of a “consumer” relying on cybersecurity in a product or service. This provides a tidy taxonomy but does not describe specific issues or problems that, if addressed, would represent significant progress toward solving them. This chapter describes these specific challenges, which can also be seen as the perspective of the “producer” who needs a well-characterized set of independent principles and procedures—technical, policy, and operational—that are prerequisites to addressing the cyber hard problems.

The consumer and producer lists do not map neatly to one another for the following reasons:

  • Solving a consumer cyber hard problem may require solving many different (e.g., technical) producer problems, and the technical problems may affect many different consumer-level problems.
  • The consumer cyber hard problems themselves are interdependent. For example, risk assessment depends on secure design and composition as well as supply-chain integrity, the availability of metrics and the reliability of data and information and is profoundly affected by policy related to “remedies and incentives” that help making such an assessment and being able to reasonably rely on it.
  • Policy and economic incentives affect nearly all the producer cyber hard problems since incentives determine what producers do (or should do), what prices they can charge to consumers, and what redress consumers have if they fail to do so.
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • Security metrics (if they exist) and the related ability to determine the cybersecurity of a system at a reasonable cost, based on easily obtainable measurements or information, affect producers because it would allow them to determine whether they have succeeded in “secure, resilient design and composition” as well as determining the security characteristics of their own supply chain and risk of liability. Finally, the consumer’s ability to determine whether a system was securely designed or implemented and whether its operation is secure “in practice” would best be based on metrics instead of largely unsupported assertions by producers.

Some of the new cyber hard problems, such as the integration of hardware and software into a cyber-physical system (CPS) or artificial intelligence (AI), depend on essentially all of the producer cyber hard problems. However, they bring new important subproblems that are critical and unsolved.

For clarity and brevity, the producer cyber hard problems are described below in terms of concrete functional, operational, new technology, or policy problems. Some producer cyber hard problems may seem to duplicate consumer cyber hard problems. For example, “secure, resilient design” from a consumer’s point of view involves solving many subproblems because it is a characteristic of an entire cyber system and its operation. The “secure design” problem in this section addresses the technical problems involved in designing a more or less fully specified system (development tools, testing tools, design practices, needed workforce competencies, etc.). This also applies to secure composition. Accurate provisioning of data and information may be directly visible to a consumer, but it may be a characteristic of the training data used to produce an AI model, which may be of no direct interest to a consumer.

Some of the producer cyber hard problems can be solved in a rather satisfying manner by principled techniques that often go under the rubric “the science of security.” Examples include complete access and information flow models.1 This sort of solution is the “gold standard” for scientific progress but has only been applied to very carefully described and constrained subproblems.

FUNCTIONAL CYBER HARD PROBLEMS

Functional cyber hard problems deal with the design of secure, interoperable products and infrastructure.

___________________

1 F.B. Schneider, 2012, “Blueprint for a Science of Cybersecurity,” The Next Wave 19(2):47–57, https://www.cs.cornell.edu/fbs/publications/SoS.blueprint.pdf.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Cloud Visibility and Centralized Risk

The emergence and popularity of cloud computing are due to convenience, efficiency, and (in some cases) cost-saving. However, while almost everyone uses or relies on cloud computing services today, there are few dominant providers, and their operations are unknown to outsiders. The monoculture and opacity result in key challenges to building and operating resilient cloud systems.

Many of the drivers for cloud computing adoption involve cost and convenience, but the trade-off most pertinent for security is control and understanding. While models of shared responsibility usually exist between the customer and provider, they are not always consistent or complete. Multi-tenancy in a provider’s environment can affect the visibility available to each customer when the provider cannot separate logging, backups, or forensic data for each tenant; sometimes the physical location of a provider’s data center is confidential. This is the “isolation” problem in shared resources. Shared hardware increases the risk of lateral movement by an attacker from one customer to another, and the network traffic needed by one customer may require that the provider cannot block some network traffic even if some tenants want them to. Finally, depending on the service, customers may not have visibility into traffic and interactions that happen within the provider’s environment, only the traffic that happens directly between the customer’s instance and the customer’s own location. Business email compromise (BEC) for the purpose of redirecting payments is a big problem and shows no sign of slowing.2 It is often accomplished by adding filtering rules redirecting emails pertinent to payments to a scammer acting as a “man in the middle” who instructs the payments to be redirected. Outsourcing email to “the cloud” where such critical rule changes for a single email user may not be quickly caught is a common problem.

Besides visibility issues, another hard problem for cloud computing is orchestration. It is difficult to find reliable statistics on multi-cloud use that do not come from a single cloud provider, and control policies between providers can be inconsistent. A look at the latest Cloud Controls Matrix3 from the Cloud Security Alliance tells the story: 197 control objectives in 17 domains. Organizations face security challenges regardless of the strategy they embrace. Using one provider can risk a single point of failure,4 and

___________________

2 P. Harr, 2024, “The Weaponization of AI: The New Breeding Ground for BEC Attacks,” Forbes Technology Council, June 14, https://www.forbes.com/councils/forbestechcouncil/2024/06/14/the-weaponization-of-ai-the-new-breeding-ground-for-bec-attacks.

3 K. Rundquist, 2024, “Cloud Security Alliance Announces Implementation Guidelines v2.0 for Cloud Controls Matrix (CCM) in Alignment with Shared Security Responsibility Model,” BusinessWire: Cloud Security Alliance (CSA), June 4, https://www.businesswire.com/news/home/20240604212963/en/Cloud-Security-Alliance-Announces-Implementation-Guidelines-v2.0-for-Cloud-Controls-Matrix-CCM-in-Alignment-with-Shared-Security-Responsibility-Model.

4 Intelligent Transportation Systems Joint Program Office, “ITS Deployment Evaluation,” Department of Transportation, https://www.itskrs.its.dot.gov/2019-l00856, accessed December 5, 2024.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

using more than one provider incurs complexity and management costs (as well as the increased systemic risk of any one provider having an outage).

A particular risk with cloud computing is the durability of artifacts; although using only what you need, when you need it can help reduce costs, the practice requires more rigor in managing the life cycle of those instances. For example, the Colorado Department of Transportation fell victim to a ransomware attack in 20185 when a virtual server was not secured properly because it was intended to be temporary, and yet it was connected to the agency’s active directory domain, which allowed the attacker to gain additional privileges. Retention periods for backups and logs are other examples of critical artifact management properties.

The start of the COVID-19 pandemic in 2020 forced more organizations to embrace cloud computing with remote access, which, in turn, drove the development of architecture changes such as “edge computing.” Remote user traffic that had to pass through on-premises infrastructure to access cloud-based resources resulted in network bottlenecks and latency; the Secure Access Service Edge emerged in response, putting the users and resources closer to one another. All this reliance on third-party providers has opened new areas of attack as well as complicated security management;6 because any given end-to-end interaction now involves additional personnel, terms of service, and levels of visibility and control different from a simple governance model.

Assessing how well a provider secures its offerings is one challenge covered further in the section on supply chain security below. A related cyber hard problem is the process of incident response, for the reasons outlined above. Putting these together creates an overall cyber hard problem that deserves attention because one provider’s outage or compromise can affect literally thousands or even millions of customers. The notion that every organization is solely responsible for securing itself is outdated, as is acknowledged in the most recent White House strategy document cited earlier. The answer lies not just in another technology product that customers must layer on top of the already complex infrastructure (complexity to manage complexity), but includes aligned incentives and clearer, more consistent security standards and responsibilities for these “linchpin” cloud providers.

Cloud infrastructure is included in cyber hard problems 1, 2, 3, 4, 5, and 10.

Identity, Authentication, and Access Control in the Context of the Global Commons

The democratization of technology has increased the number of accounts each user relies on dramatically, but the demographics of those users have changed. Consumers as

___________________

5 Ibid.

6 Office of the National Cyber Director (ONCD), 2023, “The National Cybersecurity Strategy,” The White House, March 2, https://bidenwhitehouse.archives.gov/oncd/national-cybersecurity-strategy.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

young as 3 and as old as 103 must (securely!) authenticate themselves for purposes as varied as online games, banking, education, government services, medical care, and employment. Attackers have taken advantage of the gaps in implementation and differing levels of user sophistication by capturing credentials or tricking the legitimate account holder into exposing authentication information. Identity, authentication, and access control have become a significant battlefield and a favored “soft target” even as viable technical solutions have been accepted.

In response, identity and access management (IAM) technology has evolved to meet this need.7 IAM trends include the following:

  • Increasing the availability of two-factor and multi-factor authentication (2FA and MFA, respectively). Although 2FA was available long before the 2005 cyber hard problems list, more tools emerged with different ways of enforcing the “something you have” that contributes to more secure authentication. At the same time, the use of SMS as one of these authentication factors, although it is the easiest and cheapest to deploy at scale, particularly where feature phones (precursor to the “smart” phone era) are more widely available than “smart” phones, has come under widespread attack through both social engineering and the practice of “SIM swapping.”8
  • The growth of single sign-on (SSO) to evaluate and pass authentication data under time-limited conditions so that the user does not have to repeatedly authenticate to access multiple systems, applications, and data in the scope of a single task or enterprise. This generally takes place within one governance area (such as a corporate infrastructure), where a consistent access policy can be enforced; sometimes SSO can be federated among disparate cooperating entities.
  • The implementation of passkeys (cryptographically generated key pairs) that are securely stored within the user’s device or other hardware, taking the place of memorized (and usually overused and oversimplified) passwords.
  • The development of password managers, which store and in some cases autofill passwords at authentication time, making it easier for users to choose complex or unique passwords without having to remember them or type them in.

___________________

7 National Institute of Standards and Technology (NIST), 2023, Digital Identity Guidelines, SP 800-63, https://pages.nist.gov/800-63-3.

8 Federal Bureau of Investigation, 2022, “Criminals Increasing SIM Swap Schemes to Steal Millions of Dollars from US Public,” Public Service Announcement: Alert Number I-020822-PSA, February 8, https://www.ic3.gov/PSA/2022/PSA220208.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • Heavier use of biometrics (fingerprint, facial recognition, gait analysis, typing analysis, and other techniques) for the initial authentication in place of passwords, to unlock other stored credentials such as passkeys, or coercion detection.

Around the early 2000s, the UK-based Jericho Forum proposed a stronger authentication model, called a “collaboration-oriented architecture.” With the principle of explicitly authenticating every access request, regardless of where it originated, “zero trust”9 resulted in many additional authentication factors being developed, such as GPS-based location, just-in-time analysis of the security state of the device being used for access, WiFi fingerprinting, biometrics, passkeys, and more. The additional options in this factor portfolio also made it more difficult to build and test consistent access policies.

Another complicating factor in IAM is time. Authentication has moved beyond a one-time event into continuous evaluation of the user’s factors, including location or security state changes during the session and activity alerts. Depending on the assessed risk and policy, the system might invoke a step-up authentication process with more factors to ensure that the access is legitimate. The continuous assessment may also take specific events into account, using data received either from the system owner’s own infrastructure (such as network telemetry, application changes, or a change in user access from the identity provider) or from collaborating entities. For example, a password change should invalidate all currently open sessions. The OpenID Foundation Shared Signals working group10 is tackling the challenge of standardizing and sharing access-related events, but as with many information-sharing initiatives in cybersecurity, misaligned incentives can hamper this goal.11 Finally, there is the concept of granting authorization dynamically, in a just-in-time fashion, rather than equipping an account with static permissions. The “zero standing privileges” approach is intended to harden existing user accounts in the face of attacks but carries its own set of associated management complications.

Identities themselves are now more widespread and context-specific than they were 20 years ago—they involve not just “who are you?” but “why should you have access to this particular resource at this point in time?” Establishing the right to access by verifying that the user is a citizen, a parent, a doctor, a partner, an employee, a customer, or a student can require collecting attributes from many different trusted parties.

___________________

9 S. Balaouras, J. Blankenship, D. Holmes, P. McKay, J. Burn, A. Tatro, and M. Belden, “The Business of Zero Trust Security,” Forrester, https://www.forrester.com/zero-trust, accessed February 6, 2025.

10 T. Cappalli, S. Miel, S. O’Dell, and A. Tulshibagwale, “Shared Signals Working Group—Overview,” OpenID, https://openid.net/wg/sharedsignals, accessed February 6, 2025.

11 Any time information sharing is voluntary, commercial drivers can get in the way of sharing useful and complete information. For example, security vendors with their own threat intelligence teams may avoid sharing unique data if it is seen as a competitive advantage or delay release past the point of timeliness in order to publish within marketing schedules.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

The collection of personal data, often incentivized financially for marketing and resale, contributes to the attack surface for each individual, as threat actors can obtain a wider variety of demographic data and secrets needed to register, use, and recover access.12

Collecting this verified data, storing them securely, and only releasing them where necessary are all associated privacy challenges. Proposed solutions include a digital self-sovereign identity framework, such as the European eIDAS regulation13 that will require every European Union country to offer a digital identity wallet by 2026. U.S.-based government services such as Global Entry are now offering digital IDs for mobile devices. Large-scale public identity providers such as Apple, Facebook, and Google have been offering to ease usability for consumers by letting them use their account identities for logging in to other sites, making payment transactions, and so on. However, these varied offerings come with their own governance and privacy goals, which the general public may not be able to evaluate.

To make matters even more complicated, identity management and governance have moved beyond the realm of humans. Machine identities, workload identities, and the operational system accounts that underpin all types of infrastructure, from applications to network routers, all need to be addressed in a coherent way, particularly with the growth of the Internet of Things (IoT). Wherever access is not tied to an individual human, or wherever two entities communicate with one another without human initiation, the authentication, authorization, and identity issues still apply. The Workload Identity working group14 is addressing some of these issues, but the drudgery of tracking, auditing, and protecting dormant on-premises system accounts remains with the owners of the infrastructure. Not only do attackers regularly target default passwords on these systems, but the potential areas for attack now range from critical infrastructure (utilities, nuclear power plants, 911 systems, medical equipment) to security cameras, home thermostats, baby monitors, and indeed anything that is connected to the Internet under the guise of being “smart.” Non-human identities are equally important to combat cases where an attacker simulates a website or message to trick a user into supplying credentials; the machine needs to authenticate itself to the human.

One final point is that although IAM frameworks and technology have evolved, they are also extending the “long tail” of legacy systems that are too costly to retrofit. For every web-based application that now uses passkeys for authentication, there is also a decades-old banking mainframe or industrial storage tank that must still interoperate.

___________________

12 Department of Defense (DoD), 2023, “2020 DSB Summer Study on New Dimensions of Conflict: Executive Summary,” DoD Office of Prepublication and Security Review 23-S-2072, April, https://dsb.cto.mil/wp-content/uploads/reports/2020s/DSB-SS2020_NewDimensionsofConflict_Executive%20Summary_cleared.pdf.

13 European Commission, 2024, “eIDAS Regulation,” April 4, https://digital-strategy.ec.europa.eu/en/policies/eidas-regulation.

14 IETF Datatracker, “About the IETF Datatracker,” https://datatracker.ietf.org/release/about, accessed February 6, 2025.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

The overarching technical hard problem within IAM is configuring and managing policy. Because of complicated mechanisms for performing authentication and authorization, expanded user demographics, and many, often ephemeral, systems that require IAM without governance, the need to protect these vulnerable attack points requires simplifying and possibly centralizing policy management. Today’s chief information security officer has no straightforward way to decide which factors to use, how to model the operational impact of a policy change, how to negotiate policy enforcement with external providers, or even how to get all the event data they need to make (or possibly automate) risk decisions involving policy across the entirety of the technologies and environments. One example of the current fractured state of IAM is universal access revocation, sometimes called “single logout”—the problem of identifying all accesses belonging to a departing user and revoking them, terminating any existing live sessions, and handling deeper layers of associated application and system access such as cryptographic tokens. This “holy grail” of IAM applies not only to access management in workplace scenarios but also to any active incident response involving a compromised user account. Achieving it comes with trade-offs—for example, tracking every place where a user is active can also have privacy implications.

As mentioned above, the nontechnical hard problem aspect of IAM involves governance. Identity management, authentication, and access control are driven by commercial entities and are fragmented accordingly, as governance falls to a population of private and public resource owners, not simply a central one in the role of an enterprise employer as in the past. In countries where each citizen has a single government-managed digital ID, the resulting centralization affords better technical solutions and enhances accessibility for underserved populations. Long-standing mistrust of centralized government in the United States stands in the way of creating a centralized digital identifier. Where cultural distrust of centralization is higher, it may be more practical to develop a broad federation model, allowing disparate resource owners and consumers to use a consistent and reliable framework for negotiating IAM features and processes.

Access controls affect cyber hard problems 1, 2, 3, 4, 6, 7, and 10, although it affects others to a more limited extent.

Developing an Empirical Basis for Security Decisions

Cybersecurity is the property of technological artifacts, people, and processes to resist attacks by an adversary. At every stage of a system’s life cycle, including design, implementation, acquisition, testing, deployment, training, use, monitoring, maintenance, and retirement, there are a broad range of decisions to be made that will influence these properties. What authentication architecture should be chosen? What programming languages, tools, and processes should be used to minimize the introduction of

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

implementation bugs? When acquiring such a system, how should an organization compare the security it offers to competing systems? How should their information technology (IT) professionals configure the system to support its security assumptions? How should employees be trained to use it? When must a system be updated or retired due to security liabilities? Such questions, explicit and implicit, are being answered thousands of times a day. It is widely held that some of these choices are likely better than others. Indeed, many believe that there may be “best” choices for a given situation and that certain decisions, if taken, would seriously foreclose attacks.

While it is tempting to hope that these questions might be answered a priori—that with the proper levels of formal reasoning, systems might be designed and proven secure against reasonable threat assumptions—such results are rarely available. Real systems operate in a messy world, typically more complex than can be modeled, with countless deviations from idealized abstractions, with multiple humans in the loop, and adversaries who formulate their attacks based on the assumptions made by defenders.

This is strong motivation to place cybersecurity on a firm empirical footing—akin to evidence-based medicine—where careful data collection and analysis can differentiate and prioritize among the plethora of factors and approaches. However, a perennial challenge for the cybersecurity community has been to establish a rigorous evidentiary basis for evaluating such choices in a way that predicts outcomes. As a result, most of today’s established cybersecurity “best practices” are based on a combination of perceived common sense and received wisdom.

There are many reasons to question the quality of this status quo decision making.

  • Disagreement about best practices. Even among experts, there is frequent disagreement about the set of best practices, and which are the most important. For example, Redmiles and colleagues’ recent large-scale analysis of online security and privacy advice identified 374 distinct pieces of guidance, of which a set of experts surveyed identified 188 as being among the “top 5” practices that should be followed.15 Using a similar methodology, Reeder and colleagues surveyed 200 security experts who identified 152 distinct “top 3” practices.16 While consensus is no substitute for evidence, a lack of consensus suggests an information-poor environment in general.
  • Poor evidence even where there is consensus. In those cases where there has been agreement about best practice, it is frequently based on simplified abstractions of how systems are built, used, and attacked. Empirical data are

___________________

15 USENIX, 2020, “29th USENIX Security Symposium,” August 12–14, https://www.usenix.org/conference/usenixsecurity20.

16 IEEE Symposium on Security and Privacy, 2017, “38th IEEE Symposium on Security and Privacy,” https://www.ieee-security.org/TC/SP2017.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • ultimately needed to evaluate the impact of these best practices. For example, many major software developers follow some sort of secure development process (e.g., incorporating threat-modeling, code review, fuzz testing, such as Microsoft’s Security Development Lifecycle), but even after 20 years of experience, little is known about how effective these processes are for improving security. A 2019 Dagstuhl Seminar summarized the situation as follows: “There is little empirical data to quantify the effects that these principles, architectures and methodologies have on the resulting systems.”17 Similarly, there is broad acceptance of the security benefits of outsourcing key services to cloud providers with high-quality security teams, but there is no practical way to reason about how to compare these benefits with the independent risk of correlated losses when one of these cloud platforms is itself compromised.
  • The evidence that is collected repeatedly contradicts prior assumptions. In the handful of cases where there have been extensive empirical analyses of these assumptions, best practice has repeatedly been found wanting. For example, for decades the received wisdom concerning password content requirements was that requiring more character classes and longer passwords was superior, on the assumption that this would increase entropy and thus increase the work factor for an attacker to guess. Password reset policies (e.g., resetting passwords more frequently) was deemed superior, on the assumption that it reduces the window of vulnerability for using a password. While these rules are frequently ascribed to Bill Burr’s 2003 authorship of National Institute of Standards and Technology (NIST) SP 800-63 and its appendix that codifies these arguments,18 the history is significantly older and more diffuse. For example, these issues are discussed in Morris and Thompson’s “Password Security: A Case History”19 and 20 years later in Adams’s and Sasse’s “Users are Not the Enemy.”20 However, regardless of the origin, the key point is that these arguments have been widely accepted and implemented based on their implicit assumptions concerning attacker and user behavior, without any significant empirical scrutiny. Yet after a decade of research, it became clear that these approaches were in fact not superior in practice, had been based on incomplete assumptions about how users and attackers work, and frequently

___________________

17 A. Shostack, M. Smith, S. Weber, and M.E. Zurko, 2019, “Empirical Evaluation of Secure Development Processes,” Dagstuhl Reports 9(6)1–25, Schloss Dagstuhl – Leibniz-Zentrum für Informatik, https://doi.org/10.4230/DagRep.9.6.1.

18 R. McMillan, 2017, “The Man Who Wrote Those Password Rules Has a New Tip: N3v$r M1^d!” Wall Street Journal Pro-Cybersecurity, August 7, https://www.wsj.com/articles/the-man-who-wrote-those-password-rules-has-a-new-tip-n3v-r-m1-d-1502124118.

19 R. Morris and K. Thompson, 1979, “Password Security: A Case History,” Communications of the ACM 22(11):594–597.

20 A. Adams and M.A. Sasse, 1999, “Users Are Not the Enemy,” Communications of the ACM 42(12):40–46.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • interact with usability in ways that actually reduce security. A variety of recent studies have been unable to find clear and convincing empirical evidence that well-accepted practices actually improve security, including embedded phishing training,21 indicators of compromise-based threat intelligence sharing, and prompt browser updating.22 Some studies support actual benefits, but with only minor effect (e.g., Thompson and Wagner’s study of code review impact on security vulnerabilities23). For the vast majority of security decisions made, there are no studies at all. We simply act without evidence.

This is far from a new realization. “Metrics for Security” was identified as a key cyber hard problem in the 1995 version of the InfoSec Research Council’s Hard Problem List,24 the 2002 National Research Council consensus study report Cybersecurity Today and Tomorrow: Pay Now or Pay Later,25 the Computing Research Association’s 2003 “Four Grand Challenges in Trustworthy Computing,”26 the 2005 President’s Information Technology Advisory Committee report Cyber Security: A Crisis of Prioritization,27 and again in the InfoSec Research Council’s 2005 re-up of Hard Problem List28—relabeled as “Enterprise-level Security Metrics” (although not because the smaller scale problems had been solved, indeed that study states, “Most of the existing [security] metrics are of questionable utility, even with respect to individual software systems.”29,30). Indeed, almost 20 years later, the software-focused 2024 Office of the National Cyber Director (ONCD) report Back to the Building Blocks: A Path Toward Secure and Measurable Software opines that still “it is critical to develop empirical metrics that measure the cybersecurity quality of software.”31

___________________

21 D. Lain, T. Jost, S. Matetic, K. Kostiainen, and S. Capkun, 2024, “Content, Nudges and Incentives: A Study on the Effectiveness and Perception of Embedded Phishing Training,” arXiv:2409.01378.

22 L.F. DeKoven, A. Randall, A. Mirian, G. Akiwate, A. Blume, L.K. Saul, A. Schulman, G.M. Voelker, and S. Savage, 2022, “Measuring Safety Practices,” Communications of the ACM 65(9):93–102.

23 C. Thompson and D. Wagner, 2017, “A Large-Scale Study of Modern Code Review and Security in Open Source Projects,” PROMISE ‘17, November 8, https://people.eecs.berkeley.edu/~daw/papers/coderev-promise17.pdf.

24 The 1995 Infosec Research Council (IRC) Hard Problems is not easily found, but the problems themselves are available in Appendix A, “Retrospective on the Original Hard Problem List,” of the 2005 Hard Problem List report. See IRC, 2005, Hard Problem List, November, https://www.nitrd.gov/documents/cybersecurity/documents/IRC_Hard_Problem_List.pdf.

25 National Research Council, 2002, Cybersecurity Today and Tomorrow: Pay Now or Pay Later, National Academy Press, https://doi.org/10.17226/10274.

26 Computing Research Association, 2003, “Four Grand Challenges in Trustworthy Computing,” https://archive.cra.org/Activities/grand.challenges/security/grayslides.pdf.

27 President’s Information Technology Advisory Committee, 2005, Cyber Security: A Crisis of Prioritization, National Coordination Office for Information Technology Research and Development, February, https://www.nitrd.gov/pubs/pitac/pitac_report_cybersecurity_2005.pdf.

28 IRC, 2005, Hard Problem List.

29 IRC, 2005, Hard Problem List, p. 56.

30 D. Maughan, 2006, “Infosec Research Council Hard Problem Lists,” Department of Homeland Security, Science and Technology Directorate, January 26.

31 ONCD, 2024, Back to the Building Blocks: A Path Toward Secure and Measurable Software, The White House, February, https://bidenwhitehouse.archives.gov/wp-content/uploads/2024/02/Final-ONCD-Technical-Report.pdf, p. 11.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Given the clear need, why, then, has there been such limited progress in cybersecurity while other fields, such as medicine, have been able to incorporate empirical data to great success?

The traditional refrain, well described in Herley and van Ooorschot’s overview paper on the “science of security,”32 is a triumvirate of problems that describe the “unique challenge” in measuring cybersecurity.

  • Adaptive intelligent adversaries. This concern is, by far, the most common—the fact that the scope of adversarial behavior cannot be empirically accounted for in measurements of systems. An adaptive adversary, by definition, can change their method of attack—either in response to changes in the system itself, or changes in attacker motivation, knowledge, or investment. Thus, evidence concerning the cybersecurity of a system at a point in time may fail to predict changed outcomes driven by an attacker’s new behavior or abilities.
  • Lack of fundamental laws. Unlike the invariant natural laws of physics, cybersecurity is a moving target—“too entwined with human behavior and engineered systems to have universal laws.”33 Thus, it is believed that few aspects of system behavior extracted from empirical observation can be effectively generalized into firm rules that may be counted on for future decision making.
  • Dynamic artifacts. The modern computing environment is constantly changing, as are the behaviors and modes of use of its users. Thus, at best, one can obtain evidence for cybersecurity questions at a point in time, but the relevance of those answers will be of unknown length.

While these challenges are real, none seem fundamentally at odds with the notion of empiricism or the scientific method. Indeed, a range of other disciplines faces one or more of these issues and still fruitfully make use of empiricism in practice. The common thread among all three of the problems is dynamism, which occurs in a variety of other disciplines as well. For example, while non-adversarial, insurers’ assumptions about various kinds of property damage risk have been repeatedly updated and revised in response to observations that prior measured likelihood distributions were no longer predictive (e.g., concerning hurricanes or wildfires). Finally, closest to cybersecurity, economists routinely drive policy decisions using empirical tools, despite addressing a system that is, at its core, both adaptive and quasi-adversarial (few would seriously argue

___________________

32 C. Herley and P.C. van Oorschot, 2017, “SoK: Science, Security and the Elusive Goal of Security as a Scientific Pursuit,” 2017 IEEE Symposium on Security and Privacy (SP) 99–120, https://oaklandsok.github.io/papers/herley2017.pdf.

33 D. Evans and S. Stolfo, 2011, “Guest Editors’ Introduction: The Science of Security,” IEEE Security & Privacy 2011(9):16–17, https://www.computer.org/csdl/magazine/sp/2011/03/msp2011030016/13rRUwh80sR.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

that the Federal Reserve would make better decisions if only it ignored empirical data). While each of these settings differs from cybersecurity in key ways, all are also removed from the idealized notion in which measurements can be used to derive static generalizable laws that then offer perfect predictive power. That such empirical analyses may be imperfect or have limited lifetime does not eliminate their value—for the alternative is to act without the benefit of concrete evidence at all.

In the late 20th century, portions of the medical community popularized evidence-based medicine (EBM) to incorporate a range of empirical evidence to guide research and ultimately update practice guidelines with the singular goal of improving clinical outcomes. Firmly embedded in the scientific method, EBM fostered hypothesis generation from laboratory and qualitative studies, building into both prospective and retrospective case studies, then driving repeated randomized controlled trials of prospective treatments and filtered based on comprehensive meta-analysis. There is little debate that this effort has been transformative for the practice of medicine. While this precise formulation is unlikely to translate directly to the cybersecurity realm, a similar kind of focus and investment to pursue outcome-focused results is needed.

However, there is a range of obstacles that will need to be addressed to make progress, including the following:

  • Data collection. Evidence-based methods live or die based on the availability of evidence. Some kinds of analyses can be done in synthetic controlled settings, but much work requires in situ longitudinal data collection. Unfortunately, few organizations are organized around security-relevant data collection, and incentives—both due to privacy and liability—support holding such data closely. While it is straightforward to estimate how many Americans died of heart disease last year and their associated demographics, it is much harder to measure how many servers were compromised in each market sector, how many passwords were stolen from accounts in a position of authority, or the kind and number of new software vulnerabilities that were found and fixed and whether they were implicated in critical infrastructure (viz. Shostack’s Cyber Public Health34). There are legal, operational, ethical, and economic aspects to this problem. How should such data collection be incentivized? What protections should exist for collectors, and how should the privacy interests of individuals implicated in the data be ensured? How might the use of such data be limited to improving cybersecurity? These are not simple challenges, and they implicate roles for the government, the private sector, and

___________________

34 CyberGreen, “Indirect Cost Policy,” https://cybergreen.net/technical-report-22-01/in, accessed February 6, 2025.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • civil society. Improved cybersecurity data collection is not a problem that will be fixed in any single place, but it is “table stakes” for any significant progress.
  • Broadening analysis. There is a tendency of cybersecurity discussions to focus on concrete artifacts—products and services. These are clearly important in any assessment, but it cannot be done at the expense of understanding what happens after such artifacts are fielded, and the processes and behaviors around how these artifacts are deployed and used. It is well understood, for example, that system misconfiguration—how a system is specialized for the environment in which it is operating—is as much of a security problem in practice as vulnerabilities in the underlying software systems. Similarly, most data breaches implicate some amount of social engineering (e.g., via phishing).35 Systems are routinely developed to enforce a range of potential security policies, but there is only the most cursory understanding of what policies end up defined in practice, how and why they fail to capture the true security risks in their systems, and what solutions might better bridge gaps in IT staff knowledge and organizational behavior. As security is a holistic property, these inconvenient and hard to analyze realities cannot be ignored—from system administration to user training—and these are particularly places where empirical assessment will be critical.
  • Cultural change. The computing community has been slow in embracing empirical and experimental methods as an approach for reasoning about cybersecurity. The core of computer science has traditionally been mathematical and mechanistic, and there can be a cultural tendency to strongly prefer solutions of that flavor. Indeed, when appropriate, such formal approaches are highly desirable as they can offer qualitatively comprehensive answers to scoped problems. Writing code in a language with a memory-safe type system (particularly one whose implementation and underlying runtime interface have been subject to adversarial empirical scrutiny) will generally be preferable to empirical studies about the past probability of memory errors in an unsafe language being detectable. However, outside of crisp properties such as memory safety, complex systems swiftly enter realms that are very much human-driven and with large numbers of hidden variables. This will require both the training of technologists in such methods and a cultural shift to recognize that such statistical deconstructive approaches (describing reality) are valuable and necessary complements to traditional strengths in more formal constructive ways of reasoning (defining reality).

___________________

35 Verizon, 2024, “2024 Data Breach Investigations Report,” https://www.verizon.com/business/resources/reports/2024-dbir-data-breach-investigations-report.pdf.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Despite these roadblocks, there has been considerable innovation in the “evidence-based security” space since the last version of this report. Among the approaches that have borne fruit are the following:

  • Opportunistic use of available data. A community of researchers has been slowly building empirical analysis based on what data can be publicly gleaned. For example, open-source software revision control data, combined with bug databases and public repositories of vulnerabilities and exploits, have been used to identify a number of correlates of software vulnerability introduction, the efficacy of vulnerability-finding techniques such as fuzz testing, and the varied impact of “bug bounty” programs. Similarly, widespread active network scanning—enabled by tools such as zmap—and public breach reporting requirements have allowed correlating publicly visible security features with data breach outcomes.36 The combination of “hard” (i.e., measurements) and “soft” (i.e., surveys) data is starting to emerge helping unravel thorny questions about the role of human behavior in security outcomes.
  • Outcome proxies. Absent the ability to measure the impact of a security intervention directly, some research has explored the use of outcome proxies that capture some notion of adversarial hardness. For example, researchers have used the “underground price” of various criminal commodities and services to gauge the value of security interventions,37 bulk shutdown of compromised accounts,38 phone verification for account creation,39 and account compromise.40 The theory is that defenses that increase costs for attackers who offer their services at retail will naturally translate to an increase in their asking prices. In the same vein, others have offered bug bounty prices (both first party and from third parties such as Zerodium) as a metric to infer how hard it is to exploit particular popular platforms.
  • Second-order measurements. Absent the ability to concretely measure security outcomes, a parallel line of work has focused on empirical assessments

___________________

36 USENIX, 2015, “24th USENIX Security Symposium,” August 12–14, https://www.usenix.org/conference/usenixsecurity15/technical-sessions/presentation/liu.

37 A. Searles, Y. Nakatsuka, E. Ozturk, A. Paverd, G. Tsudik, and A. Enkoji, 2023, “An Empirical Study & Evaluation of Modern CAPTCHAs,” pp. 3081–3097 in 32nd USENIX Security Symposium, https://www.usenix.org/conference/usenixsecurity23/presentation/searles.

38 K. Thomas, D. McCoy, C. Grier, A. Kolcz, and V. Paxson, 2013, “Trafficking Fraudulent Accounts: The Role of the Underground Market in Twitter Spam and Abuse,” 22nd USENIX Security Symposium, August 14–16, https://www.usenix.org/conference/usenixsecurity13/technical-sessions/paper/thomas.

39 Computer and Communications Security, 2014, CCS ‘14: Proceedings of the 2014 ACM SIGSAC Conference on Computer and Communications Security, Association for Computing Machinery.

40 A. Mirian, J. DeBlasio, S. Savage, G.M. Voelker, and K. Thomas, 2019, “Hack for Hire: Exploring the Emerging Market for Account Hijacking,” pp. 1279–1289 in WWW ‘19: The World Wide Web Conference, May 13, arianamirian.com/docs/www2019_hfh.pdf.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • of qualities that hopefully correlate with actual security outcomes. These include approaches such as process measurement,41 relative measures of attack surface reduction,42 or comparative analysis of bug-finding tools (e.g., fuzzers) on software corpora, as well as exploring the efficacy of transparency approaches such as a software bill of materials (SBOM).
  • Attackers and victims. Multiple past and ongoing efforts have empirically explored concrete threat actors in great depth—for example, identifying attackers driven by social and financial reasons, or sponsored by nation states. Similarly, studies of victims and how the effectiveness of training, guidance, and nudges to reduce the likelihood of negative outcomes have led to significant improvements in major services.

Finally, one highly-desired manifestation of evidence-based security is the establishment of security metrics—parsimonious measures allowing the evaluation and/or comparison of the security offered by a particular system. The kinds of data that will be needed for an empirically based security research agenda are clearly amenable to being shaped into metrics. However, what makes metrics attractive is that they abstract and simplify. One can easily use metrics as a decision criterion for whether an organization has improved or not, whether vendor A or vendor B is more secure, or as defense against liability. However, this same attractiveness creates strong incentives for standardization and can create institutional inertia that makes it difficult to change or react when the context in which the metrics were measured has changed. Even worse, failures in analysis or validation might elevate “bad” metrics, which ultimately incentivize less secure decisions. The same result can come from the incentive to “game” good metrics. Thus, it is critical to also consider how to deliver evidence-based security research, without enabling the most negative aspects of institutional desires for cheap decision making.

Many security metrics, from “speed to patch” to the collection of end user agent parameters, have been developed. However, there is no basis for believing any set of existing metrics provides an accurate prediction of safety or a root-cause analysis of previous losses.

While there is some research on software developers’ ability to produce secure software, this research seems to have had little impact on actual software development tools and processes. Secure development life-cycle practices and the tools that support them have largely come from industry and often have their own usability challenges.

___________________

41 Cybersecurity and Infrastructure Security Agency (CISA), 2016, “Alert: OpenSSL ‘Heartbleed’ Vulnerability (CVE-2014-0160),” October 5, https://www.cisa.gov/news-events/alerts/2014/04/08/openssl-heartbleed-vulnerability-cve-2014-0160.

42 Github, “AttackSufaceAnalzyer,” Microsoft, https://github.com/microsoft/AttackSurfaceAnalyzer/pulls, accessed February 6, 2025.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Some training to help software developers produce more security software can produce recommendations that seem on the surface impractical.

The inability to develop compact, predictive, measurable security metrics informs and affects essentially all of the consumer cyber hard problems.

Secure Resilient Engineering and Formal Methods for High Assurance

Many of the traditional problems of “secure design and composition,” described in detail above, include careful specification, isolation, and partitioning of functionality—following the principle of “least privilege” by authenticating the principal on whose behalf actions are taking and verifying their “right” to take such an action (the basis for “zero trust”). There are also emerging technologies, such as “confidential computing,” which provides a strong, principled basis for authentication of programs (an important security principle) to establish a principled, distributed basis for partitioning, isolation, and trust management. Several “producer technologies” for achieving secure resilient design and composition are discussed below.

There is a wide diversity of engineering interventions, ranging from selecting safer programming languages such as Rust and TypeScript to making architectural choices that enhance the potential for resilient response to compromise. As noted above in this chapter, one of the most vexing cyber hard problems in cybersecurity is measurement. Measurement difficulties thwart progress, not just in implementing secure engineering practices but also in understanding trade-offs when there are choices to be made regarding which practices to adopt.

For example, how much benefit is to be obtained from using a safer programming language, and how does this compare, say, with using improved analysis tools? There may be trade-offs that involve traditional software engineering criteria—for example, how does architecting to reduce interdependency among components of a large system interact with choices to implement a design that limits trust assumptions among system components when they interact (in the same sense as zero trust, but at an internal implementation level)?

Further exacerbating this challenge is a set of enduring perceptions that engineering efforts that are directed at enhancing security and resilience have an uncertain return on investment.

When successful, however, secure engineering practice can have significant—and otherwise unattainable—benefits in reducing many aspects of cyber risk. Indeed, while it may be challenging to characterize the role of specific development or design practices, there is a range of proxies suggesting that “zero-day” exploits for commodity smart-phone and server platforms have become harder to procure over time, as evidenced by their increasing market price and owing to the increasing length and complexity of

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

exploit chains needed. Moreover, there are cases where secure engineering practice leads not only to improved risk posture but also improved productivity and, in some instances, enhanced system performance. It is reported, for example, that many users of memory-safe programming languages (e.g., Rust and TypeScript) make the adoption choice in the interests of productivity, with secondary consideration for the harder-to-measure security benefits. This means that, even if the security benefits cannot readily be measured, the adoption of improved practices based on concomitant benefits to productivity and performance can nonetheless be promoted.

There are many examples of guidance regarding secure engineering from firms, laboratories, and government agencies, including the National Security Agency, the Cybersecurity and Infrastructure Security Agency, and ONCD. Most recently, for example, ONCD issued advice regarding secure practices with a focus on memory safety and formal methods43 in the face of the paucity of cybersecurity metrics. Historically, the Microsoft Security Development Lifecycle (SDLC), which includes a mix of interventions and practices focused on process and product, has had broad adoption and, as perceived by engineering managers, meaningful benefits.44 Much of the guidance focuses on practices and processes that are associated with improved security outcomes. An example is “secure coding practice,” which involves making coding choices that reduce vulnerabilities, as demonstrated through techniques such as fuzz testing.45

Logic Flaws and the Need for Mathematical Techniques

Many security-related weaknesses and vulnerabilities go beyond simple coding practices and derive from the logical structure of software and firmware. These flaws can range from protocol and API misuse to erroneous business rules. These are logic flaws. The focus of logic flaw problems is not on the full scope of secure engineering practices, but rather on means to achieve verifiable assurances regarding the absence of certain categories of vulnerabilities that go beyond type safety and memory safety. This report focuses on logic flaws for two reasons. First, there is increasingly broad adoption of “traditional” secure engineering practices such as SDLC (and as assessed through instruments such as BSIMM).46 Second, logic flaws account for an increasing percentage of exploits,47,48 and these flaws are not readily detected using current techniques. The usual means to detect

___________________

43 ONCD, 2024, Back to the Building Blocks.

44 Microsoft, “Microsoft Security Development Lifecycle (SDL),” https://www.microsoft.com/en-us/securityengineering/sdl, accessed February 6, 2025.

45 OWASP Foundation, “Secure Coding Practices,” https://owasp.org/www-project-secure-coding-practices-quick-reference-guide/stable-en/02-checklist/05-checklist, accessed February 6, 2025.

46 BlackDuck, “What Is BSIMM?” https://www.blackduck.com/glossary/what-is-bsimm.html, accessed February 6, 2025.

47 OWASP, “Top Ten,” https://owasp.org/www-project-top-ten, accessed March 25, 2025.

48 S. McClure, 2024, “Safeguarding from Lurking Threats in Business Logic Flaws,” Fast Company, January, https://www.fastcompany.com/91013667/safeguarding-from-lurking-threats-in-business-logic-flaws.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

and mitigate logic flaws is through inspection and testing. But it is well understood that these methods are imperfect.

There is an unhappily rich collection of security attributes, as revealed in the several taxonomies that are widely referenced (e.g., CIA,49 STRIDE,50 MITRE ATT&CK). Mathematical techniques can be used to address some of these, at different levels of complexity and scale. However, some attributes cannot currently be readily modeled mathematically. These include side-channel attacks. An example familiar to security researchers is cryptographic algorithms, which are mathematically correct and whose implementation is proved consistent with the algorithms, but whose implementations on actual physical processors creates vulnerabilities based on the physics of the operation of the processors, such as power fluctuations, RF emissions, and timing of executions.

Process compliance, in many cases, is seen as more affordable to achieve than actual measurable security. Although sometimes useful,51 compliance is expensive, subject to manipulation by “well-resourced” organizations, and it delays innovation and is not really very effective, generally, in “guaranteeing” security.

Formal methods (FM), in contrast, are direct techniques, focused on the operation rather than adherence to processes in the creation of that product. These techniques, including verification and program analysis for various functional and quality attributes, have a long history, going back at least to the 1960s with work by Robert Floyd and later Tony Hoare. For many years, the principal uses were in critical applications such as commercial flight controls, embedded medical devices, and national security applications. In the past 5 years, however, the scope of application has broadened significantly to include many commercial uses. Some evidence of this is cited in the Networking and Information Technology Research and Development publication regarding the FM@Scale workshops,52 where several at-scale commercial uses are highlighted. These use cases suggest the possibility that barriers of affordability, scale, usability, and integration can be overcome for a broader range of applications, with significant benefits not just to reducing important categories of vulnerabilities but also in providing evidence in support of cybersecurity risk assessment and certification. These successes depend on our ability to express models, including specifications, for quality attributes relevant to security. Improving the scope, expressiveness, and ease of use of modeling can significantly

___________________

49 CIA refers to confidentiality, integrity, availability.

50 STRIDE refers to spoofing, tampering, repudiation, information disclosure, denial of service, elevation of privilege.

51 The National Information Assurance Partnership (NIAP) Common Criteria, for example, can involve deep analysis of design artifacts and some sampling of code in an evaluated system. See NIAP, “Common Criteria: IT Security Evaluation,” https://www.nsa.gov/Portals/75/documents/resources/everyone/2023-02-NIAP_brochure_trifold_1.pdf, accessed February 6, 2025.

52 R.W. Floyd, 1967, “Assigning Meanings to Programs,” pp. 19–32 in Proceedings of Symposium on Applied Mathematics (19): Mathematical Aspects of Computer Science, J.T. Schwartz, ed., American Mathematical Society.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

reduce the cycle of information loss and recovery that typically occurs, with great cost, as systems evolve over time. Information loss of this kind creates challenges for test and evaluation as well as for sustainment and evolution. It is one of the chief contributors to technical debt.53

An additional enabler, already in place, for maintaining continuity, enabling agility, and minimizing information loss across the full life cycle of software-based systems is the Software Acquisition Pathway. The benefits were reinforced in a recent memo from the Secretary of Defense.54,55

The following three examples illustrate the various ways that these barriers are now being overcome:

  • Defense Advanced Research Projects Agency’s (DARPA’s) High Assurance Cyber Military Systems program, circa 2012–2017 (and recently awarded a DARPA Game Changer Award), undertook case study exercises to assess and demonstrate the potential applications of FM techniques to mid-sized complex systems, including unmanned aerial vehicles. The approach to scaling is significant: Judicious choices regarding systems architecture and software infrastructure enabled a narrowing of focus in use of FM to a relatively small portion of the system, including the underlying operating system, seL4, and adherence to the key coupling constraints imposed by the systems architecture.56,57
  • One of the most significant FM interventions is the use of types—and more recently memory-safety features—in programming languages. Type declarations form a specification, and type checking verifies that the software is consistent with the declarations. The declarations are just another element of the language specification, and the type-checking algorithms, which are mathematically complex, are integrated into the compiler. Adding to the ease of use is type inference, where types can be determined algorithmically, enabling specifications to be largely omitted. The mathematical analysis, nearly fully invisible to users, can eliminate entire classes of vulnerabilities related to

___________________

53 CISA, 2025, “Closing the Software Understanding Gap,” January 16, https://www.cisa.gov/resources-tools/resources/closing-software-understanding-gap.

54 DoD, 2025, “Directing Modern Software Acquisition to Maximize Lethality,” Memorandum for Senior Pentagon Leadership Commanders of Combatant Commands Defense Agency and DoD Field Directors, from the Secretary of Defense, March 6, https://media.defense.gov/2025/Mar/07/2003662943/-1/-1/1/DIRECTING-MODERN-SOFTWARE-ACQUISITION-TO-MAXIMIZE-LETHALITY.pdf.

55 DoD, 2020, “Operation of the Software Acquisition Pathway,” DoD Instruction 5000.87, October 2, https://www.esd.whs.mil/Portals/54/Documents/DD/issuances/dodi/500087p.pdf.

56 Defense Advanced Research Projects Agency (DARPA), “HACMS: High-Assurance Cyber Military Systems,” https://www.darpa.mil/program/high-assurance-cyber-military-systems, accessed February 6, 2025.

57 A. Miller, “HACMS,” GALOIS, https://galois.com/project/hacms-high-assurance-cyber-military-systems, accessed February 6, 2025.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • consistency of interpretation of low-level representations and, with languages such as Rust, related to the safe mutation of shared data.58
  • The Amazon Web Services (AWS) cloud service is a highly complex distributed system that could manifest diverse errors related to concurrency. One of the challenges of concurrency is that errors are intermittent and at a potentially low rate. This means that testing is unlikely to find the contributing faults in the code, but that at-scale execution in a distributed cloud could nonetheless manifest errors at high frequency. Using TLA+, AWS engineers have been able to find subtle design bugs and then, on repair, prove absence of similar design bugs.59 There are many other applications. One example is the use of FM to support verification and querying of AWS security access policy implementations using satisfiability modulo theories (SMT) provers; the provers run billions of times each day to support this verification.60,61

It is evident that there is progress in FM and that barriers to its use can be overcome in specific cases, including usability, scale and composition, and integration into engineering workflows. These examples are illustrative, but there are significant barriers.62 The recent ONCD report highlights two areas—software and hardware memory safety, such as provided through Rust and similar languages, and FM to support affirmative claims—backed by evidence—regarding particular security properties.63

The solution of these problems and application and further development of these techniques acutely influences hard problems 1, 2, and 3 but also 8, 9, and 10.

Cyber-Physical Systems

A recent President’s Council of Advisors on Science and Technology (PCAST) report64 drew attention to critical CPS (which they call “operational technology systems”). These systems include the “integrated digital and physical resources that are crucial to

___________________

58 C. Thompson, 2023, “How Rust Went from a Side Project to the World’s Most-Loved Programming Language,” MIT Technology Review, February 14, https://www.technologyreview.com/2023/02/14/1067869/rust-worlds-fastest-growing-programming-language.

59 L. Lamport, 2024, “The TLA+ Home Page,” August 13, https://lamport.azurewebsites.net/tla/tla.html.

60 Amazon Web Services, “Provable Security Resources,” Cloud Security, https://aws.amazon.com/security/provable-security/resources, accessed February 6, 2025.

61 N. Rungta, 2022, “A Billion SMT Queries a Day,” Amazon Science, https://www.amazon.science/blog/a-billion-smt-queries-a-day.

62 P. Lincoln, W. Scherlis, and W. Martin, 2022, Formal Methods at Scale: 2019 Workshops Report, Computing-Enabled Networked Physical Systems Interagency Working Group, May, https://www.nitrd.gov/pubs/Formal-Methods-at-Scale-Workshops-Report.pdf.

63 ONCD, 2024, Back to the Building Blocks.

64 President’s Council of Advisors on Science and Technology (PCAST), 2024, Strategy for Cyber-Physical Resilience: Fortifying Our Critical Infrastructure for a Digital World, Executive Office of the President, February, https://bidenwhitehouse.archives.gov/wp-content/uploads/2024/02/PCAST_Cyber-Physical-Resilience-Report_Feb2024.pdf.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Americans’ daily lives, including the electrical grid, public water systems, internet and telecommunications, air traffic control and much more.” IoT and automated manufacturing systems that incorporate sensors or actuators, and indeed any CPS, can be included in this category as well. The PCAST report goes on to note:

Cyber-physical risk is high, while protections are disproportionately low. America’s infrastructure systems were created and operated long before they acquired cyber dependencies, with sensing, computing, and networking dependencies developing in different ways over time. There is no systemic, pervasive protection against cyber risk since our protections evolved over time.65

CPS (e.g., a car, a laboratory instrument, or a medical device in a doctor’s office) may have very long lifetimes. Future systems have to be shaped by cyber-informed engineering. Much of the technology that underpins cyber systems and CPS was engineered without appropriate consideration of security needs. Security and resilience elements are tacked on after systems are deployed, often imperfectly and at considerable expense.

As observed above, many CPS security practices lag far behind those of IT systems. These systems use old software versions and are often not upgraded, they are not managed under user policy, and there is seldom support for critical hardware features like “root of trust” nodes. Finally, they are especially vulnerable since they are exposed to physical attacks in addition to network-based attacks.

This is cyber hard problem 8, but it is affected by almost all the other “producer” cyber hard problems, including problem 10 (operational security).

OPERATIONAL CYBER HARD PROBLEMS

Operational cyber hard problems address securely operating a scale infrastructure, including responding to attacks. This is cyber hard problem 10. Earlier discussion has already described the importance of continuous updating, resilient deployment and operations, monitoring, and situational awareness. These are the core elements of resilient secure operation, and large cloud providers have made great progress in this area. However, customer insight into the effectiveness of already introduced measures is modest, and small providers as well as “in house operations” often suffer in comparison to the effectiveness of the operational security of a large cloud provider. This includes the operational infrastructure of CPS.

___________________

65 PCAST, 2024, Strategy for Cyber-Physical Resilience, p. 12.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Understanding Human Attributes of Attackers

It is important to acknowledge that there are human attackers who make decisions and can be influenced to defenders’ advantage. These attackers have a profound influence on the secure operation of a cyber system.

For much of the history of cybersecurity, defenders have thought about attackers as an abstract and homogenous group. This meant that defenses were applied to technical attributes of attacks, such as techniques and tools, rather than through understanding the humans behind the attacks. Maturing threat models identifying threat groups can differentiate capabilities, objectives, and likely victims. A cyber hard problem, however, remains incorporating insights about the human attributes of attackers to create tailored defenses.

Attackers routinely exploit the human weaknesses of their victims; however, defenders lack sufficient insights to effectively incorporate human factors into defense. Historically, deception, such as honeypots, has been used to manipulate attackers to gain intelligence, but these techniques are neither widespread nor generally evaluated for effectiveness. In addition, it is difficult to persuade engineering organizations, which are mainly rewarded for adding new features to work on adversarial engineering projects. Recent research on adversarial human factors is starting to identify and analyze the human attributes of attackers. More research funding is needed. Industry partnership will likely be critical for developing commercial defensive capabilities that apply adversary human factors. Together, industry and the research community have to develop metrics for evaluating the effectiveness of the approach.

Understanding the human attributes of attackers is crucial for several reasons. Primarily, it allows for more sophisticated and targeted defense mechanisms that go beyond merely blocking attacks or deactivating users to anticipating and mitigating them. This knowledge can significantly improve the ability to prevent breaches and reduce the impact of successful attacks. Governments, businesses, and individuals all have a stake in this issue because the consequences of cyberattacks can range from financial loss and reputational damage to national security threats. By understanding the motivations, psychological traits, economic incentives, and behavioral patterns of attackers, defenders can craft strategies that are more likely to deter or disrupt malicious activities.

The complexity of this problem lies in the inherent variability and adaptability of human behavior. Attackers come from diverse backgrounds, possess different skill levels, and are driven by a range of motives, including financial gain, political activism, espionage, or personal vendettas. The dynamic nature of human attributes makes it challenging to create a one-size-fits-all defense. Moreover, ethical and privacy considerations must be balanced when researching and using human factors in cybersecurity.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Lawyers taking a position that even attackers have privacy rights limits defender access to their internal communications and communication with fraud victims. The incentives for attackers are high and constantly evolving, while defenders often face significant resource constraints and organizational inertia, making it difficult to implement and adapt new strategies to increase the cost seen by attackers.

Potential approaches to this problem include interdisciplinary research that combines insights from psychology, sociology, and behavioral economics with cybersecurity. Developing comprehensive attacker profiles and predictive models can help in understanding likely behaviors and vulnerabilities. Simulation and gaming techniques can be employed to study attacker behavior in controlled environments. Additionally, leveraging machine learning (ML) and AI to analyze patterns in large data sets can provide deeper insights into attacker traits and tactics. Collaboration between academia, industry, and government will be essential to foster innovation and translate research findings into practical applications. Note, however, that attackers will be increasingly autonomous (bots) and will also employ AI to probe systems and people.

Key players who can act include cybersecurity researchers, defense agencies, technology companies, and policy makers. Cybersecurity firms and tech companies can incorporate human factors insights into their security products and services. Governments can fund research initiatives and create frameworks that encourage information sharing and collaboration. Policy makers can help by enacting regulations that support ethical research, while researchers can make progress in blinding and masking personal information to make it less easy to reidentify, to protect individual privacy. Education and training programs for cybersecurity professionals need to also emphasize the importance of understanding attacker psychology and behavior.

Success in this endeavor can be measured through several indicators. A reduction in the frequency and severity of successful cyberattacks would be a primary indicator. Improved response times and more effective mitigation strategies in the face of attacks would also signify progress. An increase in the cost of credentials or breach data offered on dark markets would be a direct observable measure of success. Additionally, the development and widespread adoption of new defense technologies and methodologies that incorporate human attributes of attackers would demonstrate advancement in this field. Regular assessments and refinements based on real-world data and feedback will be necessary to ensure the ongoing effectiveness of these approaches.

This informs cyber hard problem 10.

Personal and Societal Impacts in Design and Operation

There are few tools or techniques to predict how cyber systems may be used in the future, especially as relates to unforeseen use that can threaten an individual’s security,

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

the security of an organization, or the nation. Attacks that seem to have been largely unforeseen in the past include doxing, stalkers transitioning to the digital domain, false reporting to cause account deactivation, and disinformation and influence operations that aim to attack democratic processes.

The legacy of online anonymity, including the ease of creation of opaque personas in different forums, facilitates not just free speech but also an important category of adversarial information operations, the use of bots, artificial amplification of adverse memes, and impersonation of individuals and organizations. In certain online contexts, by contrast, identity has to be firmly established, such as for banking, health records access, and business process execution.

The challenge is to identify multifaceted approaches to identity that could thwart the adversarial behaviors that depend on lack of provenance while nonetheless supporting the many contexts where it is needed, including free speech and augmenting protections of the identity of victims and potential targets of abuse and attack (e.g., human rights workers).

Anonymity also facilitates socially destructive behaviors such as stalking, bullying, sexual harassment, and doxing (piercing the veil of the privacy of other people). What are technical approaches or potential remedies that do not require abandoning the free-speech benefits of anonymity?

Human-Centered Design of all Human-in-the-Loop Security-Related Interactions

Computer users are called on to make decisions and choices with cybersecurity consequences that they do not have sufficient knowledge and resources to make accurately or securely.

Decisions and choices may be explicit or implicit. Users may be using their own computer or a computer owned by their organization. They may be acting as an individual or as a member of an organization, such as an employee or even an administrator. They may be using special purpose tools as experts in a particular discipline (i.e., code development or threat hunting). The resources they lack to address the decisions thrust upon them, some of which may be irrevocable, are legion—information, knowledge, understanding, context, time, memory, attention, desire, and incentive.

The Verizon 2024 Data Breach Investigations Report66 calls out that approximately two-thirds of breaches involve a (non-malicious) human element. This reflects the tendency, as noted in the IAM discussion above, for all sorts of cyber products to “kick the can down the road” to required human decision making, when the humans involved do not have sufficient resources to do so. In the IAM case above, what is lacking is

___________________

66 Verizon, 2024, 2024 Data Breach Investigations Report.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

transparency into the system and an understanding of the system’s complexity, which seems obviously like a too-heavy lift to require of a user.

Even in the most straightforward context, research on expert advice about what security practices users, as consumers, need to follow shows that experts do not agree on the most important.67 The amount of expert advice is overwhelming and not necessarily coherent.

The responsibility for users to make impossible security decisions creates vulnerabilities that attackers discover and exploit using social engineering. The security impact from such attacks may be to the users themselves, their computers, their organization, or to the Internet at large. The unmet expectations placed on users deteriorate presumed protection levels (which may go undiscovered until it is too late). This can contribute to a state of habituation, or worse, learned-helplessness, for users—Why follow the security advice if I am not able to do it all, or even know which matters?

Testing with appropriately representative users is both expensive and difficult. There are no tools, frameworks, principles, or automation to do the required testing without humans in an adequate and rigorous fashion. There are still gaps between the best-of-breed testing methodologies in research and what’s available to testers, who are often entry-level employees.

While the techniques for testing are largely known, choice of testing subject needs to consider the user’s skills, knowledge, and context. Since those vary, testing needs to include a range of such subjects. In addition, because some of the cybersecurity-related decisions may come about from an error in the system, or an attack, the suite of interactions needing testing can be difficult to identify, and difficult to replicate for testing.

When users interact with devices and systems in unmanaged environments (personal use), they acquire ad hoc habits that are ineffective or insecure, and these are difficult to unlearn. Because there is often no discernable difference between effective security advice and ineffective security habits, users become accustomed to ignoring security advice and experiencing no known harms.

Builders may “punt to the user” to make decisions in uncertain design situations, which compounds the issue at hand. This shows up as either warnings presenting choices or new configuration settings, which are often hidden. If used correctly (with omniscience), perhaps the security would be improved, however usability of said features goes untested.

Security and privacy mechanisms today reflect current knowledge on how to expose and control, which often differs from the user’s cognitive model of the system.

___________________

67 E.M. Redmiles, N. Warford, A. Jayanti, A. Koneru, S. Kross, M. Morales, R. Stevens, and M.L. Mazurek, 2020, “A Comprehensive Quality Evaluation of Security and Privacy Advice on the Web,” Proceedings of the 29th USENIX Security Symposium, August 12–14, https://www.usenix.org/system/files/sec20-redmiles.pdf.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Builders build what they know how to build. Users use what they are given. The resulting gap present a symmetry of ignorance that must be overcome.68

Helping Security Professionals Help Security

This is part of cyber hard problem 6, “human–system interactions,” and cyber hard problem 2, “secure development.”

Even if human-centered design were addressed, support for human security workers is still insufficient to ensure optimal security outcomes. The workforce is core to cybersecurity, in practice, but there is not necessary support in place to help security workers thrive and perform, with education, tools, and processes. This includes workers of all sorts, including developers, designers, architects, IT, and specialists such as security architects, chief security officers, blue teams, content moderators, and fact checkers.

There are two interconnected facets to this problem—inadequate training and the inability to provide cybersecurity without that support. For generalists who are also security workers, the support needed includes training, resources, and incentives baked into the job, not exogenous. Education available to learn to code would teach best practices for coding securely. Specialized security worker education would cover the myriad of topics needed to be a security subject-matter expert—from architecture, to design, to coding security functions, to specialized security testing, to security in deployment and use.69

The issue of inadequate tools and techniques for security workers to develop secure code is addressed in the cyber hard problems above.

The pool of trained security workers is limited, and training in one job can lead to the ability to move to a better paying job, creating an ongoing training need. Jobs that involve operational security vigilance are high pressure and can lead to burnout. Lack of certification means it is impossible to enforce education, training, and standards of professional conduct.

Since the 1996 New Security Paradigms Workshop,70 there has been substantial work, both in research and practice, on usable security and privacy, particularly for individuals and consumers. However, there is less work on the many other humans involved in creating, maintaining, operating, receiving, and even attacking, security and privacy, and how they can be supported, or repelled.

___________________

68 O. Pieczul, S. Foley, and M.E. Zurko, 2017, “Developer-Centered Security and the Symmetry of Ignorance,” Proceedings of the 2017 New Security Paradigms Workshop.

69 As examples, NIST and the SANS Institute have training materials available for all levels from end users to cybersecurity professionals. See NIST, “Free and Low Cost Online Cybersecurity Learning Content,” Applied Cybersecurity Division, https://www.nist.gov/itl/applied-cybersecurity/nice/resources/online-learning-content, accessed February 6, 2025.

70 M.E. Zurko and R.T. Simon, 1996, “User-Centered Security,” New Security Paradigms Workshop, https://www.nspw.org/papers/1996/nspw1996-zurko.pdf.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Research in practical secure deployment is even sparser and tends to focus on individuals, largely overlooking organizational issues. Some research exists on why, how, and when individuals will accept security-related updates. Studies and measurements exist showing the rate of patch deployment. Mistakes are made, and attacks take advantage of those gaps. Research on expert advice about what security practices to follow shows that experts do not agree on the most important.71

Designing for Resilience

Resilience is more difficult to retrofit into existing systems as an afterthought. It requires thoughtful architectural design from the outset, considering factors such as graceful degradation and partitioning of mission-critical functions to minimize the impact of breaches. The goal is for systems to operate securely, albeit in degraded fashion, even when some components are compromised. This attribute is increasingly important as large-scale systems are interconnected into even larger-scale systems, with the larger goal of organizational resilience and the ability to operate essential business functions even when systems are impaired.72

The inherent uncertainty of cyberattacks often leads companies to delay investments despite the broader societal benefits of resilient systems. Therefore, integrating resilience into the initial design phase is far more effective than attempting to retrofit it, highlighting the importance of preparedness and proactive planning. In addition to the benefit of training, resilient design is still a matter of active research, especially regarding measuring the resilience properties of a composed system.

As cyber threats evolve, the need for resilience in maintaining operational continuity becomes increasingly urgent. Resilience is the key to ensuring that vital services, such as health care, finance, and critical infrastructure, can withstand and recover from attacks. This issue is paramount to the broader community, including businesses, consumers, and governmental bodies, as it directly impacts economic stability, public safety, and national security.

Potential approaches to enhancing resilience include both technical and organizational strategies. Again, as mentioned earlier, adopting architectural choices that support resilience in designs, including for distributed systems, as well as architecting for minimal trusted computing bases, can be beneficial. Formal or semi-formal verification of design and implementation can ensure that critical systems meet high standards of resilience.

On an organizational level, fostering a culture of preparedness and operational excellence is essential. This includes thorough risk assessments, continuous improvement

___________________

71 E.M. Redmiles, N. Warford, A. Jayanti, A. Koneru, S. Kross, M. Morales, R. Stevens, and M.L. Mazurek, 2020, “A Comprehensive Quality Evaluation of Security and Privacy Advice on the Web,” Proceedings of the 29th USENIX Security Symposium, August 12–14, https://www.usenix.org/system/files/sec20-redmiles.pdf.

72 CISA, “Secure by Design,” https://www.cisa.gov/securebydesign, accessed February 6, 2025.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

of security configurations, and minimizing the “blast radius” of potential attacks by effectively partitioning critical and non-critical functions. Information-sharing models, akin to those used by the Federal Aviation Administration or anti-spam initiatives, can enhance collective resilience by disseminating threat intelligence and best practices across organizations. Finally, preparation requires deliberate practice to ensure that people and plans are effective, agile, and ready to respond to real-world incidents.

Advancing resilience is a collective effort that involves multiple stakeholders. Government agencies play a crucial role by setting standards and incentivizing resilient design, while industry groups can contribute by developing and promoting best practices. Companies, especially those operating critical infrastructure, have to make resilience a priority in their design and operational processes. It is important that resilience strategies be practical, scalable, and widely adopted.73

Various indicators can be used to “measure” success in designing for resilience. For example, these include the system’s ability to maintain functionality during attacks, local and network outages and natural disasters, the speed and effectiveness of recovery processes, and the overall reduction in the impact of cyber incidents. Regular testing, simulation of attack scenarios, and continuous improvement based on feedback and threat intelligence will help gauge progress. Ultimately, a resilient cybersecurity posture will not only mitigate the damage from attacks but also instill greater confidence in the security and reliability of all systems.

This is cyber hard problems 1, 2, and 3, as well as its effect on problem 10.

Situational Awareness of Defenders

Growing software, system, and network size, complexity, and usage offer attackers increasing opportunities for both successful penetration (i.e., larger attack surface) and the ability to remain undetected and operate within the compromised environment (i.e., larger persistence volume). The former concern is generally addressed through secure system design and implementation, while the latter is addressed by intrusion detection and digital forensics. Although some progress has been made in software and system hardening (at least against certain classes of easier-to-exploit vulnerabilities), it appears that the dwell time of non-ransomware-focused sophisticated attackers remains high,74 despite significant investment in the collection, monitoring, and analysis of security-relevant events. Essentially, the duality of detection and evasion in cybersecurity

___________________

73 The Global Resilience Federation, a nonprofit, offers framework concepts for operational resilience for business. See Global Resilience Federation, “The Operational Resilience Framework,” https://www.grf.org/orf, accessed February 6, 2025.

74 The committee excludes ransomware because it inherently exhibits a very obvious and “noisy” behavior relatively soon after infection. Examples of more recent, long-dwell threat actors include TRIANGULATION and Volt Typhoon.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

continues to trend in ways favorable to the sufficiently motivated and well-resourced attacker.

As a result, successful defense and remediation requires timely identification and disruption of malicious activities, whether proactively or reactively. The current state of practice relegates defenders to playing “whack-a-mole” while peering through a keyhole, with in-band network management tools that are potentially influenced by the attacker. The current cybersecurity tools and practices are not sufficiently precise to reliably identify the activities of slow-and-stealthy attackers over extended periods of time, nor efficient or fast enough to identify and stop rapid-moving attacks. Furthermore, a lot of emphasis has been placed on exfiltration detection, with significantly less on detecting other types of cyberattacks such as scheduled system-level denial of service.

Several technical factors contribute to confounding the ability of defenders to achieve a sufficient level of situational awareness. These include, but are not limited to, (1) improved threat actor tactics (low “signal”) that increasingly take advantage of native features and resources75 of the targeted environment; (2) high volume of benign system events (high “noise”) as a function of system size and complexity; (3) high volume of low-sophistication attack events (high “background radiation”) that lead to alert fatigue and misprioritization of response resources; (4) more complex and diverse system capabilities that, at least from an observability perspective, partially overlap with attacker capabilities, objectives, and behaviors (e.g., built-in screen recording, system-wide document search); (5) continued reliance on human-driven analysis (threat hunting), with the corresponding limitations on volume and pace of analysis; (6) the inability of analytics to keep up with ever-growing, security-driven telemetry data volumes; and (7) diminishing returns (low “gain”) but high, continuous, fixed cost for any additional type of telemetry collected and used, due to the rich set of pathways attackers can exploit to meet a given objective. Sociotechnical factors that also negatively contribute to the problem include the high cost for security-data storage and processing, the friction of information and data sharing and analysis across intra-organizational boundaries, lack of trained personnel, and (a sometimes real) conflict with other legal, privacy, or regulatory requirements.

No single solution appears sufficiently powerful to fully or substantively address the detection problem on its own. However, with appropriate investment for further scientific investigation, the following practices and technologies could play a positive role in addressing the problem:

  • Software, system, and network partitioning76 and tailored system monitoring can increase observability and improve attack detection, if actually analyzed

___________________

75 This practice is often referred to as “living off the land.”

76 Such partitioning appears to offer several security benefits, at the potential cost of overall complexity.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • (and so carefully engineered to avoid data overload). What is needed are solid principles and tools for enabling system architects, implementers, and (critically) administrators to design, build, and manage such deployments.
  • Certain types of system structure are thought to make change detection and anomaly detection easier. For example, it is anecdotally believed that certain types of cloud computing offerings (in particular, software-as-a-service environments) are easier to defend due to the uniformity of the normal behavior of the computing and software infrastructure. Past government-sponsored efforts, such as the DARPA Transparent Computing program,77 explored aspects of system and analytic co-design for observability and cybersecurity reasoning leveraging the concept of determinism. What is needed are further studies, tools, and methodologies supporting data- and metric-driven decisions with respect to designing and implementing complex systems for observability.
  • The application of AI to the problem of information summarization, knowledge extraction, and presentation and interaction—tailored to the problem of situational awareness—offers the potential of reducing the increasing volume of telemetry to concise, contextualized, understandable, and actionable material. Existing work, such as MITRE ATT&CK,78 provides an initial framework for organizing information and knowledge, but further work is needed for determining how best to use AI in this setting.
  • Beyond direct enablement of human analysts, AI could address the problem of high cost and diminishing returns of security telemetry by continuously and dynamically orchestrating, evaluating, and acting on selective telemetry sources in response to updated models of hypothesized attacker activity based on prior observables.
  • The use of AI-driven agents for conducting large-scale red teaming can help with human analyst training; better tuning of detection models; identification of blind spots; and, at a strategic level, going beyond anecdote-driven security practices toward using better global metrics on the effectiveness of the various technologies and approaches in the field.

Most of the above practices and technologies would be significantly aided by the availability of open, high-fidelity, experimental test beds. These would need to go beyond the typical goal of offering representative topologies, systems, and software to include realistic (ideally real) background data and activity sufficient to simulate actual environments and scenarios.

___________________

77 DARPA, 2014, “Transparent Computing,” https://www.darpa.mil/research/programs/transparent-computing.

78 MITRE, “ATT&CK,” https://attack.mitre.org, accessed February 6, 2025.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

The fundamental and second-order metrics for success79 remain relevant and appropriate for evaluating individual technologies and, in some cases, combinations thereof. The primary challenges that need to be addressed are (1) providing higher-quality and unbiased evaluation of these technologies80 that goes beyond sparse, anecdotal empirical testing (e.g., human-driven red teaming) and (2) translating specific-technology in-lab effectiveness measurements to real-world impact. With respect to the latter, when better and consistent metrics of effectiveness become the norm, sharing of system and network security architecture patterns along with measurements would go a long way toward establishing a proper engineering discipline in this space.

Damage Assessment and System Reconstitution

Accurately determining the initial vector and subsequent impact of a cyberattack has always been a time-consuming and difficult task. A complete analysis would identify several actionable aspects of the attack, including the method of compromise; the software, systems, and users through which initial infection occurred; the full set of systems and data accessed, exfiltrated, or modified by the attacker; and any new software introduced, existing software modified, configuration changes made, or upstream and downstream services accessed during the attack—while at the same time filtering out the typically much larger volume of benign, legitimate activities that may be overlapping and interleaved with attacker activities in both time and space (i.e., in the same systems during the same time period). These are necessary for determining how an attacker was able to gain initial access (to prevent reinfection), what data were lost (e.g., to determine what intellectual property or customer personally identifiable information [PII] was stolen), what assets (e.g., critical infrastructure components) were tampered with, whether the attacker has been completely evicted, what latent access vectors an attacker may have introduced (again, to prevent reinfection), and what residual risk must be dealt with (and potentially through what methods). In many cases, the full extent of the damage incurred is often revealed only after significant time has elapsed since the initiation or even the discovery of the attack. Relying on reported extrinsic observables (e.g., reported financial fraud, identified damage on devices, or cyber-physical processes) negates much of the potential for timely intervention and prevention (or at least minimization) of said damage. For high-stakes events, teams of specialist forensic analysts must work manually over several weeks or months to produce an impact assessment. In the meantime, critical systems may remain exposed or even knowingly left compromised to avoid service disruption.

___________________

79 An incomplete list includes false-positive rate, false-negative rate, accuracy, precision, recall, mean-time-to-detection, and mean-time-to-remediation.

80 A fundamental limitation in the evaluation of almost all detection technologies remains the determination of false-negative rates.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

A related problem is determining the set of actions necessary to restore the integrity and trustworthiness of a system or network (along with the relevant data) after a compromise. To the extent that recovery is driven by damage assessment, there is an obvious dependency. Although one could theoretically imagine a fully agnostic system and network reconstitution (e.g., a full data and system recovery from a combination of full backup and reinstallation), several factors make this impractical at scale. These include critical external dependencies (e.g., credentials for external services), system and business availability constraints, and friction related to legacy or embedded devices (e.g., out of support devices) and failure of parts under stress (e.g., network saturation due to recovery traffic). Perhaps the biggest issue is the uncertainty in how far back to recover from,81 especially as it pertains to data. As software systems become increasingly interdependent in both direct (e.g., cloud-enabled multi-device synchronization) and subtle ways (e.g., credential caching), the traditional fallback approach of reformatting and reinstalling becomes both untenable and insufficient.

The size and continuous piecemeal evolution of software, systems, and networks inhibits a sufficiently detailed understanding of their composition and functionality (even under attack-free conditions), which is a necessary step to identifying the aspects of system operation that were (or could be) accessed or tampered by an attacker. Combined with the inherent stealthiness of attacker activities, a timely reconstruction of a reasonably complete timeline of said activities and relevant system assets is currently infeasible except in limited situations. The telemetry or logging necessary to achieve the necessary degree of visibility, strongly correlated with but potentially more detailed than that needed for attack detection, can be cost- and performance-prohibitive to collect, store, and analyze in a timely fashion, even putting aside concerns about the integrity and reliability of the telemetry data in the presence of a sophisticated attacker. Furthermore, the need to restore or maintain system operations practically limits the time and resources that can be committed to the assessment analysis. In the (typically informal) risk analysis that drives the relevant parameters for the system recovery (i.e., which systems and data, how far back), this biases toward a focus on minimization of attack footprint, allowing for undiscovered latent access and other leave-behind artifacts introduced by the attacker. At the very least, it is important to capture in the final after-action report any specific assumptions made relative to the conclusions. For example, if the system was restored from a checkpoint created on a certain date, the inherent assumption is that the compromise occurred after that date.

In terms of potential solutions in the space of damage assessment, forensic reconstruction of attacker activities would benefit from the same type of solution as is needed for attack detection (see the section above on situational awareness), albeit with the

___________________

81 Using an older backup is less likely to contain attacker artifacts, at the cost of lost data.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

need for higher fidelity. Other relevant knowledge and capability gaps that need to be addressed include the following:

  • Practical mechanisms are needed for creating high-integrity hot replicas or hot standbys that address common usage scenarios. Currently, the mechanisms and techniques that exist almost exclusively focus on server scenarios; expanding to desktop and mobile situations would greatly ease system reconstitution.
  • Even if wholesale network recovery from scratch is impractical or infeasible, designing systems and architecting networks such that rapid reconstitution from scratch of key components and/or of large numbers of enterprise devices (e.g., by limiting or eliminating local storage and configuration) can help focus attention and resources to fewer locations. Relatedly, the development of techniques for incremental recovery is needed (e.g., prioritizing recovering and operating the parts of a system that are strictly necessary for business continuity).
  • Determining and mitigating credential exposure, especially for those that are used by automated processes, is of high importance in reducing the potential for residual and lateral access. While credential management in general remains problematic, reconstitution in the absence of a human user presents its own particular challenges.

Several metrics can be used to gauge progress in this space. System-wide, the goal is to reduce mean time to second compromise and mean time to system restoration, and to achieve higher completeness in attack path reconstruction relative to ground truth (potentially in the context of red team–based evaluations, where ground truth can be made available).

This is a core aspect of cyber hard problem 10.

NEW TECHNOLOGY CYBER HARD PROBLEMS

New technology is bringing new cyber hard problems. A prime example is AI, another is CPS. The challenge in securing AI applications and CPS is a core contributor to cyber hard problem 9 but also affects 1, 2, 3, 4, and 10. It also profoundly affects cyber hard problem 7.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Security of Artificial Intelligence

It is difficult to assure the security of AI applications, particularly for design patterns that leverage generative AI. While the following discussion highlights AI-specific security challenges, it is important to note that AI systems are themselves software systems and thus susceptible to the full range of traditional cyberattacks.

The adoption of AI as an integral component in modern applications has been among the most disruptive innovations in computing this century. Many of the largest software companies have transitioned to using generative AI, as has become evident in public statements by Microsoft, Google, Meta, Salesforce, and others. Although traditional application security principles—when appropriately adopted—can safely accommodate the inclusion of AI components in software systems, there are unique attributes of AI that make securing forthcoming AI applications a hard problem.82

At the component level, both predictive and generative ML models are “non-smooth” systems that may produce very different outputs for similar inputs. Generative AI models are stochastic systems that can produce different inputs for the same input. Their non-smooth and sometimes stochastic nature may present a reliability challenge when using AI as a component in a repeatable system since their function cannot be formally guaranteed, nor behavior be fully characterized. Since the models themselves are not readily interpretable, this makes their safety and security difficult to assure. Remediation in AI components is difficult since the weaknesses which arise from training cannot be patched directly in code as it might in a traditional software component.

Applications using generative large language models (LLMs) typify several AI challenges. In a basic AI chatbot application, the user interacts with an LLM that iteratively predicts the next token (word chunk) from a growing input consisting of the original system instructions, user input, and previously predicted tokens. The initial and subsequent set of predictions is heavily influenced by the system instructions, which are designed to guide toward—but cannot robustly guarantee—predictions conforming to a preferred style or topic. Because LMs are instruction-following machines, attackers may attempt either indirectly or directly to lead the application away from the intended use. This can be especially problematic in agentic systems, in which the LLM output is connected to services that act on behalf of an (untrusted) user or respond to context fetched from external (untrusted) sources by the agent components.

While AI systems are fundamentally software systems, their characteristics—supply chains that include data sets and training code and runtime nondeterminism and non-smoothness—necessitate new approaches to risk assessment and vendor trust evaluation.

___________________

82 A. Vassilev, A. Oprea, A. Fordyce, and H. Anderson, 2024, “Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations,” NIST Computer Security Resource Center, January, https://csrc.nist.gov/pubs/ai/100/2/e2023/final.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

To that end, many emerging regulations have begun to specifically call for “AI red teaming” requirements. Regulations in the European Union83 and proposed regulations in the United Kingdom84 legislate requirements for model assessment, with an emphasis on safety and societal harms. However, these requirements still lack acceptable standards across the industry in what should be assessed, what are acceptable assessment outcomes, and how and to whom to disseminate the results of an assessment.

Although there are fundamentally still software systems that include software and third-party services in applications, the supply chain of AI applications also includes data and third-party pre-trained or fine-tuned models. In addition to the possibility that attackers may develop model or data deserialization-based file formats (e.g., pytorch, pickle, and numpy) to execute arbitrary code,85 the possibility exists that models may contain backdoor functionality encoded in the model’s architecture or model weights. Technology and processes to measure and mitigate risk in these supply-chain components are nascent. Specific challenges in supply chain include (see more at NIST Adversarial ML Taxonomy86) the following:

  • Model assurance—inherited vulnerabilities and weaknesses in third-party models, including the potential for deliberately backdoored models that rely on model weights rather than code for triggering mechanisms; and
  • Data assurance—poisoning of open web–scale data sets used to pre-train or fine-tune models that can result in targeted or indiscriminate integrity violations.

A key challenge for auditability in the AI supply chain is that there is not yet a standard for reporting the equivalent of an SBOM, although efforts to address this have emerged.87 The addition of model and data components can be accommodated by SBOM to include traditional static elements of component identification, dependency information, licensing, and versions. But for AI models, the behavioral reports should also be included that report on potentially risky runtime behaviors that have been

___________________

83 European Commission, 2024, “AI Act,” https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai.

84 Department for Science, Innovation and Technology and the Office for Artificial Intelligence, 2023, “AI Regulation: A Pro-Innovation Approach,” March 29, https://www.gov.uk/government/publications/ai-regulation-a-pro-innovation-approach.

85 Common Weakness Enumeration, “CWE-52: Deserialization of Untrusted Data,” https://cwe.mitre.org/data/definitions/502.html, accessed February 6, 2025.

86 A. Vassilev, A. Oprea, A. Fordyce, and H. Anderson, 2024, “Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations,” NIST AI 100-2 E2023, January, https://csrc.nist.gov/pubs/ai/100/2/e2023/final.

87 J. Bressers, 2023, “SBOM Everywhere and the Security Tooling Working Group: Providing the Best Security Tools for Open Source Developers,” Open Source Security Foundation (blog), June 30, https://openssf.org/blog/2023/06/30/sbom-everywhere-and-the-security-tooling-working-group-providing-the-best-security-tools-for-open-source-developers.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

observed. Unfortunately, unlike a binary set of attributes or functions, the set of risky model behaviors may be incomplete, imprecise, and less actionable than in traditional software. Thus, it is still important to employ third-party audits and third-party guardrails to discover and control runtime behavior.

Defense Against Offensive Artificial Intelligence

Defenders are unprepared for a dramatic increase in scale and complexity of cyber operations from offensive AI tools—when attackers leverage AI for traditional cybersecurity operations. The risks presented by using AI for offensive purposes are offset at least to some degree by the potential for defenders to leverage AI to implement compensatory security controls and mitigations, but these are not addressed here.

As highlighted in the National Security Commission on Artificial Intelligence Final Report,88 digital infrastructure may be increasingly indefensible against escalating, offensive, AI-enabled cyber capabilities without offsetting defensive controls. Threat actors are beginning to leverage AI for various malicious use cases, including offensive copilots, scaling social engineering attacks, and enhancing offensive operations.

Offensive AI is still nascent, but researchers are developing AI for various offensive purposes that will challenge defensive systems and processes. AI-driven offensive capabilities can increase the potency and speed of cyber campaigns and present significant threats to both digital infrastructure and human targets.

AI can expedite traditional cyber campaigns against digital infrastructure in several ways. For example, using LLMs, attackers can expedite the discovery, development, and delivery of exploits through automated code reversing, vulnerability discovery, and instrumentation of exploits for vulnerabilities. AI systems that reduce the time required for threat actors to execute attacks by automating labor-intensive tasks represent a sort of “offensive copilot” that can decrease the time to impact in cyber operations.

AI-powered tools can also assist attackers in more rapid maneuvering during hands-on parts of offensive campaigns to scale offensive operations. By integrating generative AI agentic frameworks with existing tools, attackers can orchestrate complex operations that cover large portions of an attack life cycle in a way that was not previously possible.

The impact on human targets using AI presents a formidable challenge. Disinformation campaigns that leverage deepfakes have already become part of public awareness due to several incidents involving elections89 and digital warfare that now requires

___________________

88 E. Schmidt, R. Work, S. Catz, E. Horvitz, S. Chien, A. Jassy, M. Clyburn, et al., 2021, Final Report, National Security Commission on Artificial Intelligence, released March 1, https://reports.nscai.gov/final-report.

89 E. Sayegh, 2024, “The Battle for Truth in Election Seasons: AI-Generated Deepfakes,” Forbes, May 14, https://www.forbes.com/sites/emilsayegh/2024/05/14/the-battle-for-truth-in-election-seasons-ai-generated-deepfakes.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

news consumers to question the validity of reports.90 Since bad news tends to travel faster than good news, correcting disinformation is an asymmetric challenge.

These tools can also be used for fraud. Highly realistic and interactive social engineering attacks for fraud are now possible with generative AI. In this setting, attackers can create convincing impersonations or scenarios to manipulate individuals in a way that feels customized.91 AI’s potential to scale such attacks is a developing threat vector, where generative AI can create deepfakes and other convincing forms of fake identities for automated and interactive phishing or scamming operations.

The human challenges that this presents have been called out in other cyber hard problems. The key ingredient that AI brings is the sophistication and potential for scale. While fundamental security practices can ward off many of these attacks, the increased scale and sophistication allows attackers and fraudsters to affect a much broader set of victims. In a setting of fixed resource constraints of defenders, remediation and response can become intractable.

Enforceable Policies for Data in Distributed Systems

Many cyber and cyber-enabled systems include a data component, either creating new data, processing existing data, or transmitting data to achieve a particular purpose. The designers of the system or application build the service with specific security and privacy properties to mitigate the occurrence and impact of adverse events—that is, uses of the data that go beyond the intended purpose. The user of the system, and relevant regulatory or law enforcement entities, desire the ability to hold the data steward accountable for upholding the properties as promised while also ensuring that unexpected uses of the data are not possible (i.e., “the software does what it says with the data; no more, no less”).

In the absence of a solution to this problem, there is little choice but to trust that data are collected, used, and stored appropriately without much assurance. Prior to deployment or adoption, there is a requirement to convince the user and relevant authorities that the promised properties are sufficient and correctly implemented. Post-deployment, the data steward may need to modify the data use or protection terms and need to re-consent the data subject or owner, updating the presentation of the use and proposed protections and accurately recording the update.

___________________

90 D. Klepper, 2023, “Fake Babies, Real Horror: Deepfakes from the Gaza War Increase Fear About AI’s Power to Mislead,” Associated Press, November 28, https://apnews.com/article/artificial-intelligence-hamas-israel-misinformation-ai-gaza-a1bb303b637ffbbb9cbc3aa1e000db47.

91 H. Chen and K. Magramo, 2024, “Finance Worker Pays Out $25 Million After Video Call with Deepfake ‘Chief Financial Officer,’” CNN World, February 4, https://www.cnn.com/2024/02/04/asia/deepfake-cfo-scam-hong-kong-intl-hnk/index.html.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Retrospectively, there is a requirement to be able to determine as much as possible about what went wrong (e.g., whether the security properties were inadequate, if the security properties were incorrectly implemented, or if the system were modified in some way that impacted the security properties).

  • Example 1: A social network requests a phone number to be used only for MFA. How is the developer to build and prove that the phone number is used only for MFA and not targeted advertising or other contrary behavior?
  • Example 2: The sustainability task force for a local municipality deploys a network of smart streetlights to reduce energy usage. The streetlight uses a video camera to determine lighting needs based on traffic patterns and natural light. How can citizens be sure that the local police have not used data collected by the cameras?
  • Example 3: A personal genomics company sells a DNA testing kit and allows customers to opt-in to donating data to research trials. How can customers rest assured that the data are used only for the research trials they have opted into and that the results will not be subsequently used by health insurance companies?

A sizable portion of technical innovation is rooted in advancing the state of the art of what can be done with data, yet the technical mechanisms for setting and enforcing policies throughout the data life cycle (e.g., data at rest, in transit, and in use) have not kept pace. It is extremely time-consuming and difficult to identify, mitigate, and prevent the misuse of data without policy-aware data systems. In the past, we accepted possession and access to data as a proxy for permission to use the data. The proliferation of devices that collect data; the inherent complexity of the software, hardware, and network ecosystem; and also the ease with which data can be transmitted to another party make fine-grained control over the use of data untenable for the future. The misuse of data and our inability to make verifiable claims about how data will be used degrades trust in IT systems and hampers future innovations.

Making progress on this hard problem will support better outcomes on avoiding adverse outcomes for end users related to the misuse of data, increasing trustworthiness of personal devices, and perhaps decreasing disinformation. Large-scale change is needed to evolve IT systems to be policy-aware when processing data. At a minimum, such change requires the following:

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • The design and implementation of a policy language that is suitably expressive (e.g., what is the agreed-upon set of “nouns” “verbs” and “attributes”) and
  • Designing policy authoring tools that meet the needs of diverse stakeholders, allowing them to write, review, negotiate, manage, and audit policies.

One open question is whether the desired policies are actually expressible (i.e., what kinds of policies are expressible and enforceable, and are these what people care about?). Although this has been tackled in traditional IAM systems (see above), it has not been done for shared data.

Technology-Enabled Disinformation and Fraud

The emergence of cyber-mediated, human-targeted attacks of various sorts has a history of being considered cybersecurity or privacy problems. Examples include inducing the receiver of a malicious email to download and run an attachment, click on a link (to deliver malware), clinking on a link and type things in (to steal identity), and the recognition of stalkerware as a category and problem. Targeting individuals through spear phishing and catfishing are recognized cybersecurity attacks.

Although propaganda, disinformation, and military deception have a long history, technology-enabled creation and dissemination of disinformation is a newer and growing problem. Everything on the computer is mediated by technology. The Internet and World Wide Web, along with social media, expands the reach of disinformation. Automation and AI expand the scale and precision of disinformation, to bots, deepfakes, and written and spoken text that can increasingly mimic anyone trustworthy. In parallel, “broadcast” sources of journalism (television, radio, newspapers) are being replaced with peer-to-peer communications with poor or missing authentication. What people see is determined in whole or in part by algorithms that (typically) optimize for engagement.

Per the Verizon 2024 Data Breach Investigations Report,92 deepfake-like technology has already been used in many reported cases of fraud and misinformation. As generative AI only increases in abilities, scope, and scale, AI-generated fakes as weapons of disinformation will move beyond “deepfake” pictures of humans and audio fakes, to more complex scenes compellingly attesting to events that never occurred, compelling quotes, speeches, and “fake news” articles, and full videos. Generative AI that undermines artists of all professions today can become tomorrow’s tools of disinformation.

Technology-enabled disinformation can be used to undermine individual reputations and emotional well being (e.g., deepfake revenge porn), create conspiracy theories

___________________

92 Verizon, 2024, 2024 Data Breach Investigations Report.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

targeted at public figures, and attack core democratic processes such as elections. Much of the U.S. economy relies on the reputations of the strength of businesses and financial infrastructure, making this a potential weapon against U.S. economic stability.

What makes this a hard problem is the following:

  • Core principles important to democracy, including free speech and privacy protections, preclude certain approaches to detecting these or fighting fire with fire. Censorship of U.S. citizens is directly counter to free speech.
  • Notwithstanding the variously successful Internet censorship regimes, there are no geographical boundaries that serve as a natural defensible “perimeter.”
  • Misaligned incentives—for example, for ranking algorithms, “engagement” versus “truth.” The former can be monetized. Unlike preventing cybersecurity attacks, preventing the use of a platform for disinformation is not unequivocally counter to the business concerns or agendas of those who own and run such a platform. Nor is it necessarily counter to the desires of those targeted (see The Weekly World News tabloid for a benign analogy and conspiracy theorists for a less benign example).
  • Bots are often and usually indistinguishable from people, creating perceptions of scale and diversity that can be false and misleading.
  • Human attributes and limitations are targeted: confirmation bias, the tendency to anthropomorphize, motivations, and desires to believe the best or worst in others.
  • While there are experiments in the wild or with crowdsourcing and panels, there are no tools or frameworks to systematically consider how a deployed technology may be effectively used for disinformation, including threat modeling and abuse cases to consider the cybersecurity of a system.
  • Using cyber means to maliciously manipulate individuals, communities, and societies for adversarial purposes is a cybersecurity problem, but not only a cybersecurity problem. It can involve social science, psychology, ethics, policy, business, law, and political science. It can overlap with other categories of potential cyber hard problems; social science of human capabilities and limitations and mitigations to account for challenges, engineering to mitigate human limitations and weaknesses, AI reliability, deepfakes, phones as a privacy problem, authentication and access control in the context of the global commons, and information and data provenance. It may involve so many disciplines and related issues that it is sui generis as a cybersecurity problem.
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Information and Data Provenance

Content on the Internet (and later curated as part of another data set or AI function) often does not come with any trustworthy indication of the source or provenance of that information.

Much of the information people receive comes through someone else, either directly or through a communication artifact (e.g., books, newspapers). All the information received through computer interactions is the latter. Even a video call is intermediated by sophisticated software that can change backgrounds and faces. Information comes to consumers from or through a source, and their reaction to that information is potentially colored by knowledge of that source, from news to education to books, from religion to civics to politics. The reaction may be to the identity of the source, such as an individual (Walter Cronkite) or an organization (Fox News), or the reaction may be to the process and assumptions around the source’s communication type (autobiography, medical advice).

The pseudonymity promised by “no one know[ing] you’re a dog” is rapidly extending to all contexts on the Internet. “Fake news” is shared by people you know, and anyone can stand up a website claiming to be a news or publishing source. Identity, identity attributes, and source creation context are all at risk of being inaccurately relayed or assumed. Immersive virtual environments make alternate realities the norm.

Civil and societal institutions rely on some shared understanding of the authoritativeness of various kinds of information. Examples include news about communities, states, and nations, or results of the electoral process. The stability and safety of people’s economic supports and investments rely on reliable information about them.

Building blocks for enabling a more trustworthy information ecosystem might include digital signatures (including source devices that apply signatures at the point of capture), imperceptible signals in media streams (watermarking), and widespread and reliable conveyance of provenance information through social media channels.93 An alternate approach is centralized or decentralized fact checking and “community notes.”

What makes this a hard problem is the following:

  • Information flowing on today’s Internet is almost always transformed in transit (everything from user-initiated cut and paste to transcoding to enable optimal use of bandwidth).
  • Balancing strong provenance with privacy.
  • News, as a money-making business, has almost disappeared, and there are now very few sources that are universally (or at least widely) trusted. News has also partly been replaced by entertainment. (Arguably, the news business

___________________

93 Coalition for Content Provenance and Authenticity, “Overview,” https://c2pa.org, accessed February 6, 2025.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
  • was always funded by advertising, but now news organizations are not benefitting as much from the advertising revenue.)
  • AI systems will be indistinguishable from humans for most online interactions and can scale to be formidable misinformation bots (e.g., AI systems summarizing source information). Weak authentication on the Internet, which makes it difficult or impossible to ascertain the source of a piece of information, exacerbates this problem.

POLICY CYBER HARD PROBLEMS

Many of the overarching cyber challenges described earlier are expressly amplified by missing or misaligned policies. It is difficult to design effective policy for complex systems that does not increase cost disproportionately to its benefit. Resilient system design and operations, even when properly guided by policy, can slow progress. Competing interests often retard policy solutions even when candidate solutions exist. Jurisdictional questions, including globalization and failures to provide federal preemption, further complicate policy solutions and effective remedies that would apply to providers and users in a single legal jurisdiction.94

Regulatory policy and economic incentives can be confounded by competing policy goals. For example, the desire for rapid problem identification can often be achieved by comprehensive authentication; however, this can often interfere with users’ privacy. Furthermore, policy that encourages disclosure in support of principled risk assessment can threaten providers’ intellectual property if done carelessly.

There has been policy progress that has helped ensure vulnerability and breach disclosure, but this is a fairly crude measure of resilience and safety.

The lack of effective policy (economic and regulatory) is one of the most dogged and influential of hard problems.

Policy hard problems need to be addressed by laws, policies, regulations. As described in the committee’s overarching problem framework, they affect almost all cyber hard problems.

Misaligned Incentives

Misaligned incentives in cybersecurity are a significant challenge, manifesting in the varied and often conflicting priorities of stakeholders such as vendors, consumers, insurers, and regulators.

___________________

94 Harmonizing policies across jurisdictions (state, federal, and international) is a super-hard problem.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Vendors prioritize speed to market and cost efficiency over security and say additional security measures would slow the pace of innovation, while consumers often choose products based on price rather than security features. Insurers, who have the potential to influence better security practices through underwriting conditions, struggle with accurately assessing risks and enforcing effective mitigations. Competitors, despite facing similar threats, are often unwilling to share valuable threat intelligence, undermining collective defense efforts. This misalignment results in suboptimal decisions that increase overall vulnerability and delay the benefits of addressing other cyber hard problems. Without progress on incentives, benefits from solving the other cyber hard problems will be disadvantaged or delayed.

Solving the issue of misaligned incentives is crucial for enhancing the overall security posture of the digital ecosystem. It matters to a wide array of stakeholders, including businesses that suffer financial losses from breaches, consumers whose personal information is compromised, and national and homeland security agencies tasked with protecting critical infrastructure. The economic impact of cyber incidents is substantial, with costs extending beyond immediate financial losses to include reputational damage, loss of consumer trust, and long-term recovery expenses. However, to date this has been inadequate to spur changes needed. Therefore, realigning incentives to promote better security practices is essential for reducing these risks and enhancing resilience against cyber threats.

The difficulty in addressing misaligned incentives stems from several factors. Economic and competitive pressures often discourage businesses from investing adequately in cybersecurity, as the benefits are not always immediately observable.95 The tendency to prioritize short-term gains over long-term security investments is pervasive, and the lack of standardized metrics for measuring cybersecurity return on investment or cyber-coverage quality complicates decision making. Organizations may sometimes feel that the most cost-effective method for limiting damage for faulty products is public relations, especially for categories of weaknesses that are not readily assessed or repaired. Regulatory and policy efforts to realign incentives have been slow and fragmented, with various proposals such as grants, tax incentives, and liability considerations failing to achieve widespread implementation. The complexity of the cyber threat landscape and the rapid evolution of attack techniques further exacerbate these challenges.

Potential approaches to realigning incentives include policy reforms and innovative economic models. Governments can play a pivotal role by introducing and enforcing regulations that mandate minimum security standards, introducing “safe havens”

___________________

95 This is related to myopic loss aversion. See R.H. Thaler, A. Tversky, D. Kahneman, and A. Schwartz, 1997, “The Effect of Myopia and Loss Aversion on Risk Taking: An Experimental Test,” The Quarterly Journal of Economics 112(2):647–661.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

for good faith efforts accompanied by product design transparency, and by offering tax incentives or subsidies for businesses that invest in robust cybersecurity measures. Public–private partnerships can facilitate better information sharing and collective defense initiatives. An example of a public–private partnership that does such a thing successfully is the National Cyber-Forensics and Training Alliance, which brings together the business sector and law enforcement to disrupt cybercrime. Additionally, developing standardized metrics for assessing cybersecurity investments and outcomes can help businesses make more informed decisions. The implementation of mechanisms like the U.S. Cyber Trust Mark, which provides a recognizable standard of cybersecurity for consumers of wireless IoT devices, is a step in the right direction.

Those who can take action to realign incentives span across sectors. Policy makers and regulators can introduce and enforce laws that require graded security standards, depending on the kind of device and its use environment, and incentivize compliance. For example, autos that can be easily stolen because their keyfobs use weak cryptography and pervasive back doors in network-connected devices seem like areas that need to be addressed. Absent this, industry leaders and business executives will not prioritize cybersecurity as a critical component of their operational strategy and allocate appropriate resources. Insurers can refine their risk assessment models and offer premium reductions for policyholders that adopt best practices. Consumers can influence the market by demanding more secure products and services. Additionally, cybersecurity researchers and advocacy groups can continue to highlight the importance of aligned incentives and drive awareness.

Success in realigning incentives can be measured through several indicators. A notable decrease in the frequency and severity of cyber incidents would suggest that stakeholders are making more security-conscious decisions. Increased investment in cybersecurity by businesses, moving closer to the recommended 10 percent of budgets, would also be a positive sign. Ultimately, success will be reflected in a more resilient and secure digital ecosystem where the costs and benefits of cybersecurity investments are better aligned across all stakeholders.

Cybersecurity Poverty

An often-overlooked consequence of technology’s spread is the difficulty that organizations and individuals have in securing it. Originally described in 2011,96 the “security poverty line” is a concept that delineates the “haves” from the “have nots”: whether it is economically or technically feasible to implement what is generally assumed to be effective security, given real-world conditions. Just as with economic poverty, cybersecurity

___________________

96 W. Nather, 2011, “T1R Insight: Living Below the Security Poverty Line,” 451 Research, May 26, https://web.archive.org/web/20140203193523/https:/451research.com/t1r-insight-living-below-the-security-poverty-line.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

poverty results from many complex dynamics and factors. This problem exacerbates the effects of cyber hard problems 1, 2, and 4.

Because there is no simple prescriptive blueprint for building secure systems, some have tried to measure effective mitigation of carefully scoped attack scenarios, such as MITRE’s Engenuity evaluations97; others have tried to calculate the projected cost of security technology according to a given compliance framework98 or simply following security professionals’ recommendations.99 Although peer benchmarking and trends reports describe how much some organizations spend on cybersecurity, the reports do not address whether the spending is effective or appropriate. Spending formulas, such as the percentage of IT budget, do not necessarily scale up or down, nor do they have any link to positive or negative outcomes. Some increasingly critical controls (e.g., logging) are not included in the minimum baseline edition of products but are premium priced.

Another confounding factor for organizations is expertise. Cybersecurity expertise is not simply education or training; it also includes the experience of securing new technology and diagnosing and responding to new vulnerabilities and attacks. Organizations find themselves competing for this expertise against the deeper pockets of security providers (according to Glassdoor, the total salary in 2024 for a senior cybersecurity analyst is $156,000–$234,000 per year).

Constraints within the environment also affect an organization’s capability to secure itself. For example, conventional best practice in cybersecurity calls for a system to be designed to fail safe rather than open; this is not an option in a safety-focused sector such as health care, where medical staff may never be barred access to equipment or data needed to treat patients in an emergency. Software that integrates with hundreds of different systems under a variety of countries’ regulatory environments can take months or years to update. Onsite upgrades for thousands of point-of-sale systems mean that retailers must choose carefully when to incur that downtime and expense, and certainly not during the heaviest shopping times of the year. Every cybersecurity risk framework or practice may need to be adapted substantially to work around these obstacles.

Finally, in an era where cybersecurity controls are spread among third-party providers (see the section “Supply Chain” below), organizations have to rely on the cooperation of other entities with whom they may have little to no legal or commercial influence. With a sufficiently large amount of money at stake or the possibility of negative public relations, a provider may be incentivized to meet the security requirements of a customer, but smaller organizations lacking this kind of influence cannot necessarily receive

___________________

97 MITRE, 2024, “Our ATT&CK Evaluations Methodology,” https://attackevals.mitre-engenuity.org.

98 Center for Internet Security (CIS), 2023, “The Cost of Cyber Defense,” CIS Controls Implementation Group 1, August, https://www.cisecurity.org/insights/white-papers/the-cost-of-cyber-defense-cis-controls-ig1.

99 A. Shimel, 2013, “What Is the Real Cost of Security?” NetworkWorld, April 4, https://www.networkworld.com/article/744780/opensource-subnet-what-is-the-real-cost-of-security.html.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

the emergency services they need during an incident, force the timely remediation of an identified vulnerability, or reject provider conditions that may result in increased risk (such as allowing overly broad network access). Regulations will not work unless the regulated parties have access to tools that are affordable (sustainable) and actually reduce risk. Without uniform cybersecurity regulations or other incentives, most small- and medium-sized businesses, nonprofits, and local public-sector entities (including law enforcement) have to make do with the equivalent of security scraps, with support only available piecemeal from managed service providers, from a provider specified by their cyber insurer, or volunteer efforts such as the University of California, Berkeley–led Cybersecurity Clinics.100

Supply Chain

Fully organic development shops such as Google and Apple have the advantage of full (internal) transparency in their software code bases. This facilitates direct analysis at scales ranging from lines of code in small components to design choices for APIs and architectural features. This also facilitates comprehensive assured refactoring; for example, updating a service API with potentially hundreds of clients, all incompatibly updated in an atomic action. Additionally, it facilitates a fully explicit linking of design models, implementation artifacts, test cases, analysis tooling, and any supporting elements.101,102

In other words, full transparency facilitates ongoing acceptance evaluation, rapid adaptation, and repairs without creating technical debt (i.e., expedient decisions that would later need to be revised in order to permit continued evolution of a system).

By contrast, large enterprise and mission systems are generally integrated from diversely sourced components and services (“system elements”), some of which are kept opaque to their clients in order to retain competitive advantage, protect sensitive data and algorithms, and enable update and enhancement without unwanted dependencies on (hidden) implementation choices. This means that even when one layer is revealed in a complex system, there can be multiple opaque layers beneath, analogous to “turtles all the way down.”

The integrated systems model poses challenges, however. One set of challenges relates to acceptance evaluation, due to opacity of system elements and uncertainty regarding compatibility of elements. Another set of challenges relates to update and evolution, deriving from compatibility issues as individual elements are on uncorrelated update cycles. (Services, for example, can be updated several times per day, while

___________________

100 Consortium of Cybersecurity Clinics, “Cybersecurity for the Public Good,” Center for Long-Term Cybersecurity, https://cltc.berkeley.edu/program/consortium-of-cybersecurity-clinics, accessed February 6, 2025.

101 “Why Google Stores Billions of Lines of Code in a Single Repository,” posted September 14, 2025, by @scale, YouTube, https://www.youtube.com/watch?v=W71BTkUbdqE.

102 H. Wright, 2019, “Lessons Learned from Large-Scale Refactoring,” 2019 IEEE International Conference on Software Maintenance and Evolution, December 5, https://ieeexplore.ieee.org/document/8919159.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

open-source components may be updated every few weeks.) Evolution is also impaired by opacity, and many systems sustainment teams must engage in active reverse engineering to assess and document for repair security vulnerabilities, for example. There are also challenges related to overall systems architecting and design. Architectural decisions focused on reducing coupling, localizing variabilities, enhancing key quality attributes (particularly resiliency), and the like may need to be compromised to support compatibility of APIs, data representations, and service interfaces among elements that are meant to interoperate.

An extreme example is the incorporation of vendor components as original equipment manufacturer elements into integrated systems, such as commercial desktop systems into medical devices such as imaging systems. The end user, and possibly the IT support team, might not have sufficient visibility to be aware of the incorporated desktop as other than part of an appliance, and so that desktop may not, over a period of years, receive necessary updates and security patches. The resulting vulnerabilities have been exploited in ransomware attacks.103,104

An additional consequence of supply chain opacity is hidden dependencies, where a deeply embedded vulnerable system element can trigger disruptions in the event of compromise or, in the case of open source, loss of configuration control. Attacks on embedded supply chain elements can have broad consequences, and so these elements are a favored target by attackers. Examples include Blackbaud, a service provider to financial services and other organizations including critical nonprofits.105 A research report by the Cyentia Institute (a subsidiary of Mastercard) and RiskRecon106 identified ripple effects impacting between 800 and 1,000 downstream organizations. The network security vendor SolarWinds unintentionally delivered a malware payload embedded in a signed system update that was automatically distributed.107 A more recent extended global outage, caused by an automatically deployed update to CrowdStrike security software on Windows systems, affected millions of systems from banks to commercial aviation, health care, and critical infrastructure.

This can be an issue even when the embedded system element is a tiny fragment of code. One example from 2016 is leftpad, which is an 11-line module of code in the

___________________

103 L. Hautala, 2020, “Hospital Devices Exposed to Hacking with Unsupported Operating Systems,” CNET, March 10, https://www.cnet.com/health/medical/hospital-devices-exposed-to-hacking-with-unsupported-operating-systems.

104 C. Van Alstin, 2023, “RSNA 2023: Hospital Imaging Systems May Be Gateways for Ransomware, Expert Warns,” HealthImaging, November 30, https://healthimaging.com/topics/professional-associations/radiology-associations/radiological-society-north-america-rsna/rsna-2023-ransomware-medical-devices.

105 L. Fair, 2024, “FTC Says Blackbaud’s Lax Security Allowed Hacker to Steal Sensitive Data—and That’s Just the Beginning,” Federal Trade Commission (blog), February 1, https://www.ftc.gov/business-guidance/blog/2024/01/ftc-says-blackbauds-lax-security-allowed-hacker-steal-sensitive-data-thats-just-beginning-story.

106 Riskrecon, “New Report: Ripples Across the Risk Surface,” Riskrecon by Mastercard, https://www.riskrecon.com/ripples-across-the-risk-surface, accessed February 6, 2025.

107 L. Fair, 2024, “FTC Says Blackbaud’s Lax Security Allowed Hacker to Steal Sensitive Data.”

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

million-element open source NPM ecosystem widely used for web applications. The developer of this small component chose to delete it and other elements from the library due to a dispute over names for a software package. The deletion lasted only 2 hours but caused widespread disruption because of its pervasive use deep in the supply chain supplying web applications.108

There are also supply-chain issues in IoT, CPS, and networking infrastructure generally. One example was an attack on small office and home office (SOHO) routers, identified by Black Lotus Labs at Lumen Technology.109 In this case, more than 600,000 routers belonging to a single internet service provider were completely disabled, forcing the entire customer base to have their equipment physically replaced. In other cases, these SOHO routers are not regularly updated or are no longer supported by the vendor so that no security updates are available; the accumulating residue of vulnerabilities makes a perfect platform for attackers to take over infrastructure and use it for botnets or proxying services.

On the one hand, the approach to security vulnerabilities has been to encourage organizations to patch early and often, and preferably automatically. But as these examples show, automatic updates gone wrong can also cause catastrophic events. Victims of attacks and outages are caught in the middle between conflicting imperatives, and the cybersecurity industry owes them a better answer than to say “just patch.”

The recent mandates regarding use of an SBOM can be seen as transforming what is sometimes a full opacity into a kind of “translucency” where some information is provided downstream (i.e., to client users, integrators, and end customers) in order to overcome some of these challenges and, additionally, create some incentives within the supply chain to address security attributes more aggressively. An SBOM, representing something akin to a food ingredient list, can empower organizations to make better procurement decisions, but only when they have feasible alternatives.

This is, of course, a supply-chain cyber hard problem also affecting hard problems 1, 2, and 3.

Liability

Liability for faulty code or hardware represents a critical and complex issue in cybersecurity. Under “contracts of adhesion” vendors often sell software “as is,” disclaiming responsibility for defects that may lead even to significant breaches or failures. Cyber systems cannot be evaluated based on a quick inspection (like a vacuum cleaner) or even a diligent inspection. Choices for equivalent functionality in other products, are often

___________________

108 Ibid.

109 Black Lotus Labs, 2024, “The Pumpkin Eclipse,” Lumen, May 30, https://blog.lumen.com/the-pumpkin-eclipse.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

very limited and, as discussed, there is a market failure that does not practically enable consumers to “select the model with the security they want.” Except for copyright and patent infringement and designated systems like medical devices and automotive, or software used in other regulated industries, where limitation of liability is circumscribed by law, most software is marketed with the “understanding” that some level of imperfection is acceptable. Thus “legal” remedies are largely ineffective even with expensive litigation. This becomes particularly problematic in cybersecurity where software faces sophisticated, evolving attackers. The challenge is compounded by the displacement of loss onto consumers rather than the companies responsible for the vulnerabilities.

The consequences of faulty code extend far beyond mere inconvenience. Consumers, businesses, and governments all suffer from the fallout of software failures. Establishing liability would incentivize vendors to prioritize security and quality, potentially reducing the frequency and severity of breaches. However, current practices and economic realities pose significant hurdles. Large companies, even after significant breaches, rarely face existential threats, and the costs are often borne by consumers and smaller entities.

Addressing this issue requires overcoming several barriers. First, there is a need for a cultural and operational shift within the software industry. The mantra that “we did the best we could” (even if true) must give way to more rigorous standards and accountability. Introducing liability necessitates robust metrics and frameworks to evaluate software safety and security, akin to those in place for other regulated industries. However, creating these standards is not straightforward. The dynamic nature of software development, coupled with the continuous evolution of threats, makes it difficult to establish a static baseline for safety. Furthermore, the global nature of the software supply chain complicates the assignment of liability, as many stakeholders—from developers to suppliers—are involved in the creation and maintenance of software products.

Potential approaches to this problem include both voluntary and mandatory assessment mechanisms. These standards would need to be continuously reviewed and updated to remain relevant. Another approach is to create a “safe harbor” for vendors who follow best practices, thus incentivizing compliance while recognizing the inherent challenges of achieving absolute security.110

Government agencies can establish and enforce regulatory standards, while industry groups can develop and promote good practices, with incremental adoption. Companies, particularly large enterprises with significant market influence, can lead by example, incorporating security into their development processes and advocating for broader industry changes. Collaboration between public and private sectors is essential

___________________

110 Lawfare (https://www.lawfaremedia.org/topics/cybersecurity-tech) has a number of relevant notes on this topic.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

to ensure that standards are practical and effective. Although other domains such as medicine and civil and mechanical engineering have successfully employed professionalization standards, the software field generally does not possess the sort of widely accepted, comprehensive principles as the other fields.111

Success in this endeavor would be indicated by a measurable reduction in the frequency and impact of software-related breaches. Metrics could include the number and severity of vulnerabilities discovered and exploited, time to patch after exploit announcements, the financial and operational damage from breaches, and the rates of compliance under established standards. Ultimately, creating a more secure software ecosystem will require sustained effort and cooperation across the industry, but the potential benefits for all stakeholders make it a goal worth pursuing.

Third-Party Intervention

Most cyberattacks rely on communication over the Internet. The Internet, globally and even within countries, is not owned, operated, or controlled by a single legal or technical entity or by a closed set of governments, major corporations, or technology institutions; nor is it subject to a single set of policies. Rather, the Internet is an aggregate whose owners, operators, participants, and technologies function independently of one another but interoperate. As a result, long before they touch their target networks and endpoints, cyber operations traverse and in some cases leverage infrastructure, technology, and services owned, operated, and offered by different Internet infrastructure providers (IIPs). These companies include (without counting computing hardware companies) operating system developers, cybersecurity firms, Internet service providers, mobile telecommunications companies, cloud and virtual private server providers, content delivery networks, Domain Name System service providers, hosting providers, domain registrars, and a variety of Internet technology platforms, such as browser, e-mail, and search platforms.

These companies see themselves as neutral providers of global Internet services but not as critical infrastructure assets with a key role to play in systemic security and resilience. Although most of these companies invest substantially in cybersecurity measures, and occasionally cooperate operationally to degrade specific threats, they lack regulatory, financial, or other incentives to systematically address malicious actors’ use of their technology and services in cyber operations that do not directly and immediately impact them. Their cybersecurity efforts and cross-company cooperation are further hamstrung by a complex legal, privacy, regulatory, antitrust, government, and business environment.

___________________

111 Circumscribed certifications, like network administration, can be useful but they are scarcely comprehensive.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Internet infrastructure companies have unique (and in many cases the only legal) vantage points to observe attacks and detect and frustrate malicious cyber activity on a systematic, nationwide basis (and, for some of the large IIPs, even globally). If these entities were to take on the responsibility of discovering and limiting malicious cyber activities and implement successful regimes to do so, malicious actors’ ability to conduct cyber operations against U.S. targets would be substantially attenuated. Coordination across providers is required because attackers straddle their infrastructure across multiple providers for survivability against takedowns.

There have been several attempts at private-sector coordination (with or without U.S. government participation), primarily at the threat information sharing (TIS) level, over many years. Perhaps the biggest inhibitors to such coordination and threat information sharing revolve around liability concerns and the lack of a business case. While setting industry-wide standards and the exchange of best practices is non-controversial, the (at best) federated nature of the infrastructure landscape means that operational coordination that leads to systemic action (e.g., coordinated takedowns of malicious infrastructure) is the exception rather than the norm, and only occurs as a knee-jerk reaction to high-visibility events. Therefore, in practice, most coordination has taken the form of TIS. This need not be a handicap, if the information exchanged is accurate, timely, and conveys sufficient context to provide the necessary confidence so each participant can take action. Unfortunately, that is typically not the case, and critical information is often missing due to lack of collaboration across key industry verticals.

One of the potential liability concerns expressed is that TIS represents or can lead to collusion between companies (creating the perception of collusion to the public and to regulators), leading to privacy and antitrust concerns (and associated litigation). Making the government part of any threat information sharing arrangement focused solely on technical factors is one possible approach, but for the global firms, there will also be concerns on other countries’ reactions to such direct U.S. government participation. The most common liability concern comes from actions following high-profile attacks. All such incidents have been accompanied with various types of litigation, including litigation by different agencies of the U.S. government. For companies, the most damaging form of litigation alleges “willful neglect,” meaning the commercial entity knew or should have known about the specific problem but failed to act. This creates a perverse incentive for a deliberate lack of knowledge in some area as a cost-effective form of liability protection. There will be no long-term continuous incentive for TIS or any other meaningful collaboration without addressing these concerns.

There are at least four different concerns over liability: customer and third-party impacts; notification shortfalls; regulatory fines; and civil and class-action lawsuits. In many cases, this involves different agencies such as the Department of Justice, the

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

Securities and Exchange Commission, the Federal Trade Commission, the Department of Homeland Security Cybersecurity and Infrastructure Security Agency (CISA); and, in certain cases, the Department of Defense. While CISA has made progress with information sharing, the fear of liability still exists because it is not possible for CISA to fully protect companies reporting a breach from litigious actions taken by other government agencies due to different authorities and regulations. In addition, CISA cannot protect companies from civil lawsuits from other companies, customers, and third parties.

Accordingly, and starting from a recognition that its greatest strength is as a convener for collaboration on a technical level, the U.S. government could provide a safe, non-litigious forum for technical collaboration and data sharing without fear of liability from both the U.S. government and private industry, perhaps after the model of the Information Sharing and Analysis Centers.112 Such a forum would create potential for end-to-end visibility across the domestic Internet infrastructure. Such a forum could be coordinated with the U.S. Cyber Command, and potentially be tipped by the Intelligence Community—but, critically, it should not be used as a source for data by the latter. To the extent that technologies for private TIS exist or can be developed, they will play a significant role in countering narratives of collusion and negative public perception.

One option is to require all companies over a certain size to have cybersecurity insurance—and allow companies to work together to bundle end-to-end coverage to further sharing of information. This also creates the incentive for companies to follow standards and adopt technologies for lower insurance rates. Working through the insurance providers may be a more tractable proposition. However, cyber insurance coverage is very limited and for all the reasons mentioned earlier, insurance companies themselves are in no position to judge the security of the systems they insure. Careful mechanism and incentive design are needed to avoid simply transferring the intractable problem from the company to an insurance firm.

Individuals and organizations have a legal right to pursue those who violate their service agreements with civil courts. Creating a special cyber court and providing it with the necessary technical resources to fully pursue cyber criminals and threat actors (even nation-state sponsored ones, to the extent that it is not desirable to treat these as an act of war) may be appropriate given the technical understanding required by the U.S. government and all parties.

Another option is to create a “cyber fire department,” with broad authority to act on third-party, including private cyberinfrastructure. This could be operated by the U.S. government, by contracted private entities, or in a federated or localized manner to reflect constraints of specific sectors, geographical areas, or other considerations. It is

___________________

112 National Council of ISACs, “Information Sharing and Analysis Centers (ISACs),” https://www.nationalisacs.org, accessed February 6, 2025.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.

almost certain that significant new authorities conferred by legislation (including limited liability waiver) would have to be granted to such an entity.

Ultimately, if the United States wishes for the Internet infrastructure providers to play better defense, it will have to either create the right financial incentives for markets to value security more than they currently do, or directly pay for such better practices. Tax credits, bounties, fees on connectivity bills, security investment programs through the Small Business Administration, and subsidized cyber insurance (combined with heightened terms and conditions for such policies) are only some of the ways such financial support can be extended.

Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 47
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 48
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 49
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 50
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 51
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 52
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 53
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 54
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 55
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 56
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 57
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 58
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 59
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 60
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 61
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 62
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 63
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 64
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 65
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 66
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 67
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 68
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 69
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 70
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 71
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 72
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 73
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 74
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 75
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 76
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 77
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 78
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 79
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 80
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 81
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 82
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 83
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 84
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 85
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 86
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 87
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 88
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 89
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 90
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 91
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 92
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 93
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 94
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 95
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 96
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 97
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 98
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 99
Suggested Citation: "4 The Producer Perspective." National Academies of Sciences, Engineering, and Medicine. 2025. Cyber Hard Problems: Focused Steps Toward a Resilient Digital Future. Washington, DC: The National Academies Press. doi: 10.17226/29056.
Page 100
Next Chapter: 5 Toward Community Coordination and Progress
Subscribe to Email from the National Academies
Keep up with all of the activities, publications, and events by subscribing to free updates by email.