Whatever can go wrong, will go wrong. So a solution is better the less possibilities there are for something to go wrong.
Although often cited like that, Murphy's Law actually is not a fatalistic comment stating “that life is unfair”. Rather it is (or at least can be seen as) an engineering advice to design everything in a way that avoids wrong usage. This applies to everything that is engineered in some way and in particular also to all kinds of modules, (user) interfaces and systems.
Ideally an incorrect usage is strictly impossible. For example this is the case when the compiler will stop with an error if a certain mistake is made. And in case of user interface design, a design is better when the user cannot make incorrect inputs as the given controls won't let him.
It is not always possible to design a system in such a way. But as systems are built and used by humans, one should strive for such “fool-proof” designs.
There are different kinds of possible errors that can and according to ML eventually will occur in some way: Replicated data can get out of sync, invariants can be broken, preconditions can be violated, interfaces can be misunderstood, parameters can be given in the wrong order, typos can occur, values can be mixed up, etc.
Note that Murphy's law also applies to every chunk of code. According to the law the programmer will make mistakes while implementing the system. So it is better to implement a simple design, as this will have fewer possibilities to make implementation mistakes. Furthermore code is maintained. Bugfixes will be necessary present functionality will be changed and enhanced, so every piece of code will potentially be touched in the future. So a design is better the fewer possibilities there are to introduce faults while doing maintenance work.
Systems are built and used by humans. And as humans always will make mistakes, there always will be some possibilities for a certain mistake. So if some mistake is possible, eventually there will be someone who makes this mistake. This applies likewise to system design, implementation, verification, maintenance and use as all these tasks are (partly) carried out by humans.
This means the fewer possibilities there are that a mistake is made, the fewer there will be. As mistakes are generally undesirable, a design is better when there are fewer possibilities for something to go wrong.
Note that ML does not claim that everything constantly fails unless there is no possibility to do so. It simply says that statistically in the long run a system will fail if it can.
This is a very general principle so there is a large variety of possible strategies to adhere more to this principle largely depending on the given design problem:
See section contrary principles.
The exact wording and who exactly coined the term, remains unknown. Nevertheless it can be stated that its origin is an experiment with a rocket sled conducted by Edward A. Murphy and John Paul Stapp. During this experiment some sensors have been wired incorrectly. A more accurate quote might read something like this: “If there's more than one possible outcome of a job or task, and one of those outcomes will result in disaster or an undesirable consequence, then somebody will do it that way.” A more detailed version of the history of the experiment and the law can be found in 2) and Wikipedia.
Furthermore every defect in any system is a manifestation of ML. If there is a fault then obviously something went wrong. The correlation between the number of possibilities for introducing defects and the actual defect count can be regarded trivially intuitive.
Suppose there are two methods of a string class
replaceAll() which replace the first or all occurrences of a certain substring, respectively.
The following method signatures are a bad choice:
Eventually someone will mix up the order of the parameters leading to a fault in the software which is not detectable by the compiler.
So it is better to make parameter lists consistent:
This is less error prone. When for example a call to
replaceFirst() is replaced by a call to
replaceAll(), one cannot forget to exchange the parameters anymore. This is how is is done in the Java API.
But here still one could mix up the two string parameters. Although this is less likely, as having the substring to look for first is “natural”, such a mistake is still possible. An alternative would be the following:
Here both methods expect a
Pattern object instead of a regular expression expressed in a string. Mixing up the parameters is impossible in this case as the compiler would report that error. On the other hand using these methods becomes a bit more complicated:
"This are a test.".replaceFirst(new Pattern("are"), "is");
3) instead of
"This are a test.".replaceFirst("are", "is");
The KISS-Principle is about this disadvantage.
Another example for the application of Murphy's Law would be the avoidance of typecasts:
This works but it makes a cast necessary and every cast circumvents type checking by the compiler. This means it is theoretically possible that during maintenance someone will make a mistake and store a value other than Integer in the list:
Murphy's Law claims that however unlikely such a mistake might seem, eventually someone will make it. So it is better to avoid it. In this case this could be done using Generics:
List<Integer> l = new ArrayList<Integer>(); l.add(5); return l.get(0) * 3;
Here this mistake is impossible as the compiler only allows storing integers.
Note that the typecast is rather a symptom than the actual problem here. The problem is, that the
List interface is not generic and the symptom is the typecast. The reason for this flaw is, that the
List interface predates the introduction of generics in Java.
In Java the classes ''Date'' as well as the newer ''Calendar'' are mutable which means the reference semantics of Java objects may cause unintended alternations of date values. Eventually someone will copy the reference to a date object instead of copying the object itself, which is usually a mistake when programming with dates.
Date date1 = new Date(2013, 01, 12); Date date2 = date1; System.out.println(date1); // Sun Feb 12 00:00:00 CET 3913 System.out.println(date2); // Sun Feb 12 00:00:00 CET 3913 date1.setMonth(2); System.out.println(date1); // Sun Mar 12 00:00:00 CET 3913 System.out.println(date2); // Sun Mar 12 00:00:00 CET 3913
Furthermore as can be seen in the code above, the month value counterintuitively is zero-based, which results in 1 meaning February. This obviously is another source for mistakes. Also the order of the parameters can be mixed up easily. And lastly this does not refer to a date in 2013 but to one in 3913! The year value is meant to be “two-digit”, so 1900 is added to it. So there are plenty of possibilities for making mistakes. And sooner or later someone will make them.
Because of these and several other flaws in the design of the Java date API, most of the methods in
Date are deprecated and also the newer
Calendar API will be replaced by a new API in Java 8.
Discuss this wiki article and the principle on the corresponding talk page.