{"id":698,"date":"2016-01-12T14:44:17","date_gmt":"2016-01-12T13:44:17","guid":{"rendered":"http:\/\/www.abnormaldistribution.org\/?p=698"},"modified":"2016-01-12T14:44:17","modified_gmt":"2016-01-12T13:44:17","slug":"risk","status":"publish","type":"post","link":"https:\/\/abnormaldistribution.org\/index.php\/2016\/01\/12\/risk\/","title":{"rendered":"Risk"},"content":{"rendered":"<p>There are a few different notions of risk used in dependability engineering.<\/p>\n<p>One notion, used in finance and in engineering safety, is from <a href=\"http:\/\/www.jstor.org\/stable\/1403045\">De Moivre (1712, De Mensura Sortis in the Proceedings of the Royal Society)<\/a> and is<br \/>\n<center><br \/>\n(A) the expected value of loss (people in engineering say &#8220;combination of severity and likelihood&#8221;).<br \/>\n<\/center><\/p>\n<p> A second notion, used in quality-control and project management, is<br \/>\n<center><br \/>\n(B) the chances that things will go wrong.<br \/>\n<\/center><\/p>\n<p>Suppose you have a \u20ac20 note in your pocket to buy stuff, and there is an evens chance that it will drop out of your pocket on the way to the store. Then according to (A) your risk is -\u20ac10 (= \u20ac20 x 0.5) and according to (B) your risk is 0.5 (or 50%). Notice that your risk according to (A) has units which are the units of loss (often monetary units) whereas your risk according to (B) has no units, and is conventionally a number between 0 and 1 inclusive.  <\/p>\n<p>(A) and (B) are notions 2 and 3 in the <a href=\"https:\/\/en.wikipedia.org\/wiki\/Risk\">Wikipedia article on Risk<\/a>, for what it&#8217;s worth.<\/p>\n<p>The International Standards Organisation (ISO) and the International Electrotechnical Commission (IEC) put out guides to the inclusion of common aspects in international standards. One is on <a href=\"http:\/\/isotc.iso.org\/livelink\/livelink\/fetch\/2000\/2122\/4230450\/8389141\/ISO_IEC_Guide_51_2014%28E%29_%2D_Safety_aspects_%2D%2D_Guidelines_for_their_inclusion_in_standards.pdf?nodeid=8389248&#038;vernum=-2\">Safety Aspects (Guide 51, 2014 edition)<\/a> and one is on <a href=\"http:\/\/www.iso.org\/iso\/catalogue_detail?csnumber=44651\">Risk Management (Guide 73, 2009 edition)<\/a>. The Guide 51 definition of <i>risk<\/i> is the <i>combination of probability of occurrence of harm and the severity of that harm<\/i>, where <i>harm<\/i>is <i>injury or damage to the health of people, or damage to property or the environment<\/i>. The Guide 73 definition of <i>risk<\/i> used to be <i>change or probability of loss<\/i>, i.e. (B),  but has changed in the 2009 edition to the <a href=\"https:\/\/www.iso.org\/obp\/ui\/#iso:std:iso:guide:73:ed-1:v1:en\"><i>effect of uncertainty on objectives<\/i><i><\/i><\/a>. <\/p>\n<p>The 2013 edition of ISO\/IEC 15026 Systems and Software Engineering &#8211; Systems and Software Assurance, Part 1: Concepts and Vocabulary (formally denoted ISO\/IEC 51026-1:2013), defines <i>risk<\/i> to be <i>the combination of the probability of an event and its consequence<\/i>, so (A).<\/p>\n<p>The IEEE-supported <a href=\"http:\/\/www.computer.org\/web\/swebok\">Software Engineering Body of Knowledge (SWEBOK)<\/a> says, in Section 2.5 on Risk Management,<\/p>\n<blockquote><p><i><br \/>\nRisk identification and analysis (what can go wrong, how and why, and what are the likely consequences), critical risk assessment (which are the most significant risks in terms of exposure, which can we do something about in terms of leverage), risk mitigation and contingency planning (formulating a strategy to deal with risks and to manage the risk profile) are all undertaken. Risk assessment methods (for example, decision trees and process simulations) should be used in order to highlight and evaluate risks.<br \/>\n<\/i><\/p><\/blockquote>\n<p>Notice <i>what can go wrong<\/i> is hazard identification, <i>how and why<\/i> is analysis, along with <i>what are the likely consequences<\/i>, which is severity assessment, also part of hazard analysis. What is missing here is an assessment of likelihood, which is common to both (A) and (B), the Guide 51 definition and the Guide 73 definition.<\/p>\n<p>ISO\/IEC 24765:2010 Systems and Software Engineering &#8211; Vocabulary defines <i>risk<\/i> to be<\/p>\n<blockquote><p><i><br \/>\n<b>1.<\/b> an uncertain event or condition that, if it occurs, has a positive or negative effect on a project&#8217;s objectives. A Guide to the Project Management Body of Knowledge (PMBOK\u00ae Guide) \u2014 Fourth Edition. <br \/>\n<b>2.<\/b> the combination of the probability of an abnormal event or failure and the consequence(s) of that event or failure to a system&#8217;s components, operators, users, or environment. IEEE Std 829-2008 IEEE Standard for Software and System Test Documentation.3.1.30. <br \/>\n<b>3.<\/b> the combination of the probability of an event and its consequence. ISO\/IEC 16085:2006 (IEEE Std 16085-2006), Systems and software engineering \u2014 Life cycle processes \u2014 Risk management.3.5; ISO\/IEC 38500:2008, Corporate governance of information technology.1.6.14. <br \/>\n<b>4.<\/b> a measure that combines both the likelihood that a system hazard will cause an accident and the severity of that accident. IEEE Std 1228-1994 (R2002) IEEE Standard for Software Safety Plans.3.1.3. <br \/>\n<b>5.<\/b> a function of the probability of occurrence of a given threat and the potential adverse consequences of that threat&#8217;s occurrence. ISO\/IEC 15026:1998, Information technology \u2014 System and software integrity levels.3.12. <br \/>\n<b>6.<\/b> the combination of the probability of occurrence and the consequences of a given future undesirable event. IEEE Std 829-2008 IEEE Standard for Software and System Test Documentation.3.1.30<br \/>\n<\/i><\/p><\/blockquote>\n<p>ISO\/IEC 24765 thus acknowledges that there are different notions doing the rounds.<\/p>\n<p>The System Engineering Body of Knowledge (SEBOK) says in its <a href=\"http:\/\/sebokwiki.org\/wiki\/Risk_Management\">Wiki page on Risk Management<\/a> that<\/p>\n<blockquote><p><i><br \/>\nRisk is a measure of the potential inability to achieve overall program objectives within defined cost, schedule, and technical constraints. It has the following two components (DAU 2003a):<\/p>\n<p>the probability (or likelihood) of failing to achieve a particular outcome<br \/>\nthe consequences (or impact) of failing to achieve that outcome<br \/>\n<\/i><\/p><\/blockquote>\n<p>which is a version of (A).<\/p>\n<p>What are the subconcepts underlying (A) and (B), and other conceptions of risk?<\/p>\n<p>(1) There is <i>vulnerability<\/i>. Vulnerability is the hazard, along with the damage that could result from it, and the extent of that damage; this is often called &#8220;severity&#8221;. So: hazard + hazard-severity. This is close to Definition 1 of ISO\/IEC 24765.<br \/>\n(2) There is likelihood. This can be likelihood that the hazard is realised (assuming worst-case severity) or likelihood that a specific extent of damage will result. This is only meaningful when events have a stochastic character. This is (B), the former definition in ISO\/IEC Guide 73, and item 3 in the Wikipedia list.<\/p>\n<p>If you have (1) and (2), you have (A) and (B). If you have (A) and (B), you have (2) (=B) but you don&#8217;t have (1). But (1) is what you need to talk about security, because security incidents do not generally have a stochastic nature.<\/p>\n<p>Terje Aven, in his book <a href=\"http:\/\/eu.wiley.com\/WileyCDA\/WileyTitle\/productCd-0470683880.html\">Misconceptions of Risk<\/a> argues (in Chapter 1) that even notion (A) is inadequate to capture essential aspects of risk. He attributes to Daniel Bernoulli the observation that utility is important: just knowing expected value of loss is insufficient to enable some pertinent decisions to be made about the particular risky situation one is in.<\/p>\n<p>A third subconcept underlying risk is that of <i>uncertainty<\/i>. Aven <a href=\"http:\/\/www.sciencedirect.com\/science\/article\/pii\/S095183201000027X\">has argued recently that uncertainty is an appropriate replacement for probability<\/a> in the notion of risk. Uncertainty is related to what one knows, to knowledge, and of course the Bayesian concept of probability is based upon evaluating relative certainty\/uncertainty. <\/p>\n<p>It is worthwhile to think of characterising risk in terms of uncertainty where traditional probability is regarded as inappropriate. However, there are circumstances in which it can be argued that probabilities are objective features of the world; quantum-mechanical effects, for example. And if a system operates in an environment of which the parameters pertinent for system behavior have a stochastic nature, no matter how much of this is attributable to a lack of knowledge (a failure to observe or understand causal mechanisms, for example) and how much to objective variation, such probabilities surely must play a role as input to a risk assessment.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>There are a few different notions of risk used in dependability engineering. One notion, used in finance and in engineering safety, is from De Moivre (1712, De Mensura Sortis in the Proceedings of the Royal Society) and is (A) the expected value of loss (people in engineering say &#8220;combination of severity and likelihood&#8221;). A second [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[13,6],"tags":[],"_links":{"self":[{"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/posts\/698"}],"collection":[{"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/comments?post=698"}],"version-history":[{"count":0,"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/posts\/698\/revisions"}],"wp:attachment":[{"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/media?parent=698"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/categories?post=698"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/abnormaldistribution.org\/index.php\/wp-json\/wp\/v2\/tags?post=698"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}