Intelligence analysis

From Citizendium
(Redirected from Analytic tradecraft)
Jump to navigation Jump to search
This article may be deleted soon.
To oppose or discuss a nomination, please go to CZ:Proposed for deletion and follow the instructions.

For the monthly nomination lists, see
Category:Articles for deletion.


Intelligence analysis is the process of taking known information about situations and entities of strategic, operational, or tactical importance, characterizing the known, and, with appropriate statements of probability, the future actions in those situations and by those entities. Known information may include that which is collected by intelligence organizations, as well as historical, geographic, and other basic references. The descriptions are drawn from what may only be available in the form of deliberately deceptive information; the analyst must correlate the similarities among deceptions and extract a common truth. Although its practice is found in its purest form inside intelligence agency|intelligence agencies, such as the Central Intelligence Agency (CIA) in the United States or the Secret Intelligence Service (SIS, MI6) in the UK, its methods are also applicable in fields such as business intelligence about competitors.

Intelligence analysis accepts it has to operate with imperfect information about highly ambiguous situations, with the ambiguity often very deliberately created by highly intelligent people with mindsets very different from the analyst's. Consumers of intelligence, like patients of physicians, may not realize that the best judgments are probabilities rather than certainties. Many analysts prefer the middle-of-the-road explanation, rejecting high or low probability explanations. Analysts may use their own standard of proportionality as to the risk acceptance of the opponent, rejecting that the opponent may take an extreme risk to achieve what the analyst regards as a minor gain. Above all, the analyst must avoid the special cognitive traps for intelligence analysis projecting what she or he wants the opponent to think, and using available information to justify that conclusion. Albert Einstein said "God is subtle, but he is not malicious," but the opponents of intelligence agencies may be have very different value systems. To assume that one's enemies try to confuse is not being paranoid but realistic, especially in the areas of intelligence cycle security and its subdiscipline counterintelligence. The WWII German term of counterintelligence art, funkspiel or radio game [1], is not a game in the sense of playing fields, but still drawing from game theory and the goal of confusing one's opponents.

Analysts must be good problem-solvers, blending specific knowledge with trained intuition. Since the other side may be hiding their intention, the analyst must be tolerant of ambiguity, of false leads, and of partial information far more fragmentary than faces the experimental scientist. According to Dick Heuer [2], in an experiment in which analyst behavior was studied, the process is one of incremental refinement: "...with test subjects in the experiment demonstrating that initial exposure to blurred stimuli interferes with accurate perception even after more and better information becomes available...the experiment suggests that an analyst who starts observing a potential problem situation at an early and unclear stage is at a disadvantage as compared with others, such as policymakers, whose first exposure may come at a later stage when more and better information is available.

Heuer refers to what others call the intelligence mosaic when he observes "The receipt of information in small increments over time also facilitates assimilation of this information into the analyst's existing views." As with a mosaic or a jigsaw puzzle, there are may brightly colored pieces, but not all are relevant. "No one item of information may be sufficient to prompt the analyst to change a previous view. The cumulative message inherent in many pieces of information may be significant but is attenuated when this information is not examined as a whole." The only declassified paragraph of the United States intelligence community review of its performance before the 1973 Arab-Israeli War noted [in the only declassified paragraph] that there was no one critical piece of information.[2]

"The problem of incremental analysis--especially as it applies to the current intelligence process--was also at work in the period preceding hostilities. Analysts, according to their own accounts, were often proceeding on the basis of the day's take, hastily comparing it with material received the previous day. They then produced in 'assembly line fashion' items which may have reflected perceptive intuition but which [did not] accrue from a systematic consideration of an accumulated body of integrated evidence."

Writers on analysis [3] [4] have suggested reasons why analysts come to incorrect conclusions, by falling into Cognitive traps for intelligence analysis. Without falling into the trap of avoiding decisions by wanting more information, analysts also need to recognize that they always can learn more about the opponent.

Analytic Tradecraft

No intelligence analyst is perfect, but the best ones learn from their own mistakes and positive experience, as well as the mistakes and experiences of others. Even if the top authorities are fully ethical and effective, just as generals don't succeed without great privates, the environment needs to encourage the junior analyst, and encourage growth in what is a profession, but sometimes seemingly a black art where intuition must be cherished. The body of specific methodology for making the best feasible judgments,from limited and possibly uncertain data, often is called analytic tradecraft[5]. The academic disciplines examining the art and science of intelligence analysis are most routinely referred to as "Intelligence Studies", and exemplified by institutions such as the Joint Military Intelligence College. The goal of the Analytic Tradecraft Notes of the Central Intelligence Agency's Directorate of Intelligence (DI) include the "Pursuit of expertise in analytic tradecraft is a central element of this plan. Our tradecraft enables analysts to provide "value-added" to consumers of intelligence by ensuring:

  • Reports are objective, including an honest assessment of confidence in the conclusion.
  • Delivery must be timely, or it is useful. Analysis does not take place in a vacuum; consumer feedback is the fundamental way to drive collection and general intelligence research.

Analytic tradecraft skills also serve as "force multipliers," helping us provide top-quality analysis:

  • The feedback our customers give us on our customized analysis clarifies for the analyst what questions most need answering.
  • Employing rules for evaluating information and making judgments helps analysts manage the deluge of information, discern trends, and identify attempts at deception.
  • Tradecraft standards can be used to iron out differences among experts who have complementary substantive specialties. Their interaction enhances teamwork, which allows the [Directorate of Intelligence] to be more productive.

Setting Goals for an Intelligence Analysis

Many new analysts find that getting started is the hardest part of their job. Stating the objective, from the consumer's standpoint, is an excellent starting point. If the analyst cannot define the consumer and his needs, how is it possible to provide analysis that complements what the consumer already knows. See Bureau of Intelligence and Research for the mindset of analysts that focuis on the needs of diplomats, rather than the broader audience of the CIA.

A senior diplomat ... seized the attention of the class of some 30 [intelligence community managers] by asserting that as a policy official he never read ... analytic papers. Why? "Because they were nonadhesive." As Blackwill explained, they were written by people who did not know what he was trying to do and, so, could not help him get it done:

"When I was working at State on European affairs, for example, on certain issues I was the Secretary of State. CIA]...analysts did not know that--that I was one of a handful of key decision makers on some very important matters....

This ambassador looks back at his early periods of service at the NSC Staff and in State Department bureaus as ones of "mutual ignorance": DI analysts did not have the foggiest notion of what I did; and I did not have a clue as to what they could or should do.[6]

Blackwill explained how he used his time efficiently, which rarely involved reading general CIA reports.

"I read a lot. Much of it was press. You have to know how issues are coming across politically to get your job done. Also, cables from overseas for preparing agendas for meetings and sending and receiving messages from my counterparts in foreign governments. Countless versions of policy drafts from those competing for the President's blessing. And dozens of phone calls. Many are a waste of time but have to be answered, again, for policy and political reasons.

Not all intelligence consumers do build a wide background.

"One more minute, please, on what I did not find useful. This is important. My job description called for me to help prepare the President for making policy decisions, including at meetings with foreign counterparts and other officials.... Do you think that after I have spent long weeks shaping the agenda, I have to be told a day or two before the German foreign minister visits Washington why he is coming?"

Be Bold and Honest

Wikis advise authors to "be bold". and so do mentors of intelligence analysis. This article is written with due regard to the principle that Wikipedia is not a how-to guide, but drawing a balance that analysis is something that one does and must (in a game theory| game-theoretic sense) have a strategy of approaching the problem. Weasel-wording is as problematic in intelligence analysis as in Wikipedia, but the analyst also must recognize that some things truly are uncertain. Arguably, when uncertainties are given with probabilities or at least some quantification of likelihood, they become less a case of weasel wording and more a case of reflecting reality as it is best understood.

While a good analyst must be able to consider, thoughtfully, alternative viewpoints, an analyst must be willing to stand by his or her position. This is especially important in specialized areas, when the analyst may be the only one that reads every field report, every technical observation on a subject.

Believe in your own professional judgments. Always be willing to listen to alternative conclusions or other points of view, but stand your ground if you really believe the intelligence supports a certain conclusion. Just because someone is your boss, is a higher grade, or has been around longer than you does not mean he or she knows more about your account than you do. You are the one who reads the traffic every day and who studies the issue. [7]

At the same time, Watanabe observes, "It is better to be mistaken than wrong...." Not willing to be wrong is also a disease of the highest policymaker levels, and why there needs to be a delicately balanced relationship, built of trust, between a policymaker and his closest intelligence advisors. Intelligence analyst is not a popularity contest. It is being a seeker of the best approximation of truth.

I recall a colleague who forwarded an analysis that called into question the wisdom behind several new US weapon systems. This analysis caused criticism of the CIA, of his office, and of himself. He stood his ground, however; the Agency supported him, and eventually he was proved right. He did not make a lot of friends, but he did his job.[7]

Intelligence analysts are expected to give policymakers opinions both support and reality checks [5]. The most effective products have several common features:

  • Opportunities and dangers for interests of the analyst's country, especially unexpected developments that may require a reaction.
  • Motives, objectives, strengths, and vulnerabilities of adversaries, allies, and other actors.
  • Direct and indirect sources of friendly parties' leverage on foreign players and issues.
  • Tactical alternatives for advancing stated national policy goals.

Reality checking is not to be underestimated. In WWII, the Allies launched an air offensive against a target system that they really did not understand: the V-1 cruise missile. Image:V1-20040830.jpg|thumb|left|V-1's were attacked before their role was understood Their rationale to attack it was that if the enemy apparently valued it, it must be worth attacking [8]. This may have been rational when there were large numbers of aircraft and pilots, but it might not be warranted today, at least until the analysts have a chance to verify the target system is not a decoy. If the system is real, then it might have been warranted to defer attack until a massive one could have been delivered.

Agreement on Content

In good working relationships, the analytic process is interactive with the customer. For example, the first IMINT of Soviet missiles during the Cuban Missile Crisis was verified and quickly taken to the President and Secretary of Defense. The highest level of authority immediately requested more detail, but also wanted a perspective on the Soviet strategy, which was not available from photography. Image:Cuban missiles.jpg | thumb | Photographs cannot show what is in the opponent's mind As the White House requested more CIA and Navy support for photography, it simultaneously searched for HUMINT and SIGINT from Cuba, as well as diplomatic HUMINT. Until John F. Kennedy was briefed by excellent briefers, such as Dino Brugioni, he probably did not understand the capabilities of IMINT[9].

Frequently, the intelligence service will organize the production process and its output to mirror the customer organization. Government production by the single-source intelligence agencies is largely organized geographically or topically, to meet the needs of all-source country, region, or topic analysts in the finished-intelligence producing agencies.

In terms of intended use by the customer, both business and government producers may generate intelligence to be applied in the current, estimative, operational, research, science and technology, or warning context. Serendipity plays a role here, because the collected and analyzed information may meet any or all of these criteria.

A good example is warning intelligence[10]. Military and political analysts are always watching for predefined indication that an emergency, such as outbreak of war, or a political coup, is imminent. When an indicator is approved, policymakers are alerted and a crisis team is often convened, with the mission of providing time-sensitive intelligence on the situation can be issued to all relevant customers.

Orienting oneself to the Consumers

Experienced analysts recommend seeing oneself as a specialist on a team, with 5-10 key players. Learn something about each of them, both in terms of how they express themselves, and how you can reinforce their strengths and support their weaknesses. The analyst must constantly ask himself, "what do they want/need to know? How do they prefer to have it presented? Are they still trying to select the best course of action, or have they committed and now need to know the obstacles and vulnerabilities on their chosen path?"

Others on the team may know how to handle the likely challenges. The analyst's contribution is in recognizing the unlikely, or providing connections that are not obvious. Consumers must get information in a timely manner, not after they commit to a decision they might not have made having rougher information available sooner.

Sometimes, when the producer is struggling with how to meet the needs of both internal and external customers, the solution is to create two different types of products, one for each type of customer. An internal product might contain detail of sources, collection methods, and analytic techniques, while an external product is more like journalism. Remember that journalists always address:

  1. Who
  2. What
  3. When
  4. Where
  5. Why

"How" is often relevant to journalists, but, in intelligence, may wander into that delicate area of sources and methods, appropriate only for internal audiences. The external consumer needs to know more of potential actions. Actions exist in three phases:

  1. The decision to act
  2. The action
  3. Disengagement from the action [11]

Internal products contain details about the sources and methods used to generate the intelligence, while external products emphasize actionable target information. Similarly, the producer adjusts the product content and tone to the customer’s level of expertise.

Orienting Yourself to Peers

Even in professional sports, where there are strict anti-fraternization rules on the playing field, players often have deep friendships with counterparts on opposing teams. They might have been on a college team together, or are simply aware that the team they oppose today might be the team to which they might be traded tomorrow. If a technique is personal, rather than a proprietary idea of a coach, one professional might be quite willing to show a nominal opponent how he does some maneuver.

As an analyst, you have the advantage of knowing both what the consumer needs to know (sometimes better than the consumer knows himself) and which collectors can obtain the needed intelligence...An analyst needs to understand the general capabilities and limitations of collection systems. ...If the analyst is in a technical discipline, the analyst might have an insight about a collection system that the operators have not considered.[7]

Watanabe said "If you are not frequently tasking collectors and giving them feedback on their reporting, you are failing to do an important part of your job."[7]

Peers, both consumer and analyst, also have a psychological context. Johnston wrote [12] suggests the three major components of that context are:

  1. socialization
  2. degree of risk taking or risk aversion
  3. organizational-historical context

Devlin [13] observes that while traditional logical work does not consider socialization, work on extending logic into the real world of intelligence requires it. "The first thing to note, and this is crucial, is that the process by which an agent attaches meaning to a symbol always takes place in a context, indeed generally several contexts, and is always dependent on those contexts. An analytic study of the way that people interpret symbols comes down to an investigation of the mechanism captured by the diagram:

[agent] + [symbol] + [context] +. . . + [context] → [interpretation]

Things that are true about contexts include:

  1. Contexts are pervasive
  2. Contexts are primary
  3. Contexts perpetuate
  4. Contexts proliferate
  5. Contexts are potentially pernicious

The discipline of critical discourse analysis [14] can help organize the context. Michael Crichton, [15] in giving examples of physicians communicating with other physicians, points out that laymen have trouble following such discourses not only because there is specialized vocabulary in use, but the discourse takes place in an extremely high context. One physician may ask a question about some diagnostic test, and the other will respond with a result from an apparently unrelated test. The shared context was that the first test looked for evidence of a specific disease, while the answer cited a test result that ruled out the disease. The disease itself was never named, but, in the trained context, perfectly obvious to the participants in the discourse.

Intelligence analysis is also extremely high context. Whether the subject is political behavior or weapons capabilities, the analysts and consumers share a great deal of context. Intelligence consumers express great frustration with generic papers that waste their time by giving them context they already have internalized.

Organizing What You Have

Collection processes provide analysts with assorted kinds of information, some important and some irrelevant, some true and some false (with many shades inbetween), and some requiring further preprocessing before they can be used in analysis. Raw information reports use a Intelligence collection management#Ratings by the collection department | standard code for the presumed reliability of the source and of the information. The US intelligence community uses some formal definition of the kinds of information [5]

Term Definition Example
Fact Verified information; something known to exist or to have happened. A confirmed inventory of a resource of one's own service
Direct Information The content of reports, research, and analytic reflection on an intelligence issue that helps analysts and their consumers evaluate the likelihood that something is factual and thereby reduces uncertainty, Information relating to an intelligence issue under scrutiny the details of which can, as a rule, be considered factual, because of the nature of the source, the source's direct access to the information, and the concrete and readily verifiable character of the contents COMINT or OSINT quoting what a foreign official said; IMINT providing a count of the number of ships at a pier. HUMINT from a US diplomatic officer who directly observed an event.
Indirect Information Information relating to an intelligence issue the details of which may or may not be factual, the doubt reflecting some combination of the source's questionable reliability, the source's lack of direct access, and the complex character of the contents HUMINT from a reliable agent, citing secondhand what an informant said that a government official said. OSINT providing a foreign government document that gives the number of ships at a pier. Indirect OSINT from a US embassy officer. COMINT that contains a report by a foreign official to his government, about what something he cannot confirm, but states with a probability.
Direct Data Organized information that provides context for evaluating the likelihood that a matter under scrutiny is factual. A chronology of events based on observations by US officers
Indirect Data Organized information that provides context for evaluating the likelihood that a matter under scrutiny is factual. A chronology based on reports from a liaison intelligence service

Collation describes the process of organizing raw data, interpolating known data, evaluating the value of data, putting in working hypotheses. The simplest approaches often are an excellent start. With due regard for protecting documents and information, a great deal can be done with pieces of paper, a whiteboard, a table, and perhaps a corkboard. Maps often are vital adjuncts, maps that can be written upon.

There are automated equivalents of all of these functions, and each analyst will have a personal balance between manual and machine-assisted methods. Unquestionably, when quantitative methods such as computer simulation are appropriate, the analyst will want computer assistance, and possibly consultation from experts in methodology. Geospatial intelligence involves combining maps and imagery, especially different kinds of imagery, tools needed to superimpose different maps and images, and to suppress certain details and add others.

Outlining, possibly in a word processing program; using the most advanced scientific visualization or a simple sketch can give structure; file folders and index cards and can be moved in a way hard to do in software. Data bases analyzed with statistical techniques can give insight; consulting statisticians can be very helpful.

Some analysts speak of a Zen-like state in which they allow the data to "speak" to them. Others may meditate, or even seek insight in dreams, hoping for an insight such as that from which a noted chemist inferred the structure of benzene, as snakes chasing snakes.

Krizan[4] took criteria from [16]. Regardless of its form or setting, an effective collation method will have the following attributes:

  1. Be impersonal. It should not depend on the memory of one analyst; another person knowledgeable in the subject should be able to carry out the operation.
  2. Not become the “master” of the analyst or an end in itself.
  3. Be free of bias in integrating the information.
  4. Be receptive to new data without extensive alteration of the collating criterion.

Analysts have found that the principles of the philosophical discipline of ontology [17] have Intelligence collection ontology|applications in intelligence

The more interactive that the relationship between producer and consumer becomes, the more important will be tools [18]:

  • Collaboration tools. These include all media: voice, video, instant messaging, electronic whiteboards, and shared document markup
  • Databases. Not only will these need to be interoperable, they need to reflect different models, when appropriate, such as the semantic web. There may no longer be a clear line between databases and web applications.
  • Analytic tools. These will cover a wide range of pattern recognition and knowledge organization.

The Nature of Analysis

An analysis is not a nicely arranged scrapbook of raw data. It should have a summary of the key characteristics of the topic, followed by the key variables and choices. Increasingly deep analysis can explain the internal dynamics of the matter being study, and eventually to prediction, known as estimation.

The purpose of intelligence analysis is to reveal to a specific decision maker the underlying significance of selected target information. Analysts should begin with confirmed facts, apply expert knowledge to produce plausible but less certain findings, and even forecast, when the forecast is appropriately qualified. Analysts should not, however, engage in fortunetelling that has no basis in fact.

Image:FoodChain.png | thumb | Food chain in intelligence analyst: the bigger the fish, the more unlikely it is

The mnemonic “Four Fs Minus One” may serve as a reminder of how to apply this criterion. Whenever the intelligence information allows, and the customer’s validated needs demand it, the intelligence analyst will extend the thought process as far along the Food Chain as possible, to the third “F” but not beyond to the fourth.

Types of Reasoning

Objectivity is the intelligence analyst’s primary asset in creating intelligence that meets the Four Fs Minus One criterion. To produce intelligence objectively, the analyst must employ a process tailored to the nature of the problem. Four basic types of reasoning apply to intelligence analysis: induction, deduction, abduction and the scientific method.[4]

Induction: Seeking Causality

The induction process is one of discovering relationships among the phenomena under study. It may come from human pattern recognition ability, looking at a seemingly random set of events, perhaps writing them on cards and shuffling them until a pattern emerges[4].

An analyst might notice that when Country X's command post with call sign ABC sent out a message on frequency 1 between Thursday and Saturday, an air unit will move to a training range within one week. The acknowledgement will take one day, so the analyst should recommend intensified COMINT monitoring of the appropriate frequencies between Friday and Sunday. Another kind of causality could come from interviews, in which soldiers might describe the things that warn them of an impending attack, or how the ground might look when an improvised explosive device has been emplaced.

While induction, for human beings, is usually not at a fully rational level, do not discount the potential role of software that uses statistical or logical techniques for finding patterns. Induction is subtly different from intuition: there usually is a pattern that induction recognizes, and this pattern may be applicable to other situations.

Deduction: applying the General

Deduction, is the classic process of reasoning from the general to the specific, a process made memorable by Sherlock Holmes: "How often have I said to you that when you have eliminated the impossible, whatever remains, however improbable, must be the truth?" Deduction can be used to validate a hypothesis by working from premises to conclusion[4].

The pattern of air maneuvers described above may be a general pattern, or it may be purely General X's personal command style. Analysts need to look at variables, such as personalities, to learn whether a pattern is truly general doctrine, or simply idiosyncratic.

Not all intelligence officers regard this as a desirable approach. At his confirmation hearing for CIA Director, Gen. Michael V. Hayden said he believes that intelligence analysis should be done by "induction," under which "all the data" are gathered and general conclusions determined, rather than by "deduction," under which you have a conclusion and seek out the data that support it [19]

Trained Intuition

Analysts need to harness trained intuition: the recognition that one has come to a spontaneous insight. The steps leading there may not be apparent, although it is well to validate the intuition with the facts and tools that are available[4].

Polish cryptanalysts first were reading German Enigma machine|Enigma ciphers in 1932, although the commercial version may have been broken by the British cryptanalyst, Dilwyn Knox, in the 1920s. Poland gave critical information to the French and British in 1939, and production British cryptanalysis was well underway in 1940. The Enigma, with German military enhancements, was quite powerful for a mechanical encryption device, and it might not have been broken as easily had the Germans been more careful about operating procedures. Throughout the war, Germany introduced enhancements, but never realized the British were reading the traffic almost as fast as the Germans.

Image:Enigma.jpg|thumb|left|Even though the Germans improved Enigma machine|Enigma, poor procedure compromised it

US cryptanalysts had broken several Japanese diplomatic ciphers, but, without ever seeing the Purple machine until after the war, they deduced the logic. Purple was actually mechanically simpler than Enigma, but the US Army team struggled with a mechanical reproduction until Leo Rosen had the unexplained insight that the critical building block in the Purple machine was a telephone-type stepping switch rather than the rotor used in Enigma and in more advanced US and UK machines. Rosen, Frank Rowlett, and others of the team recognized Rosen's insight as based on nothing but a communication engineer's intuition.

Experienced analysts, and sometimes less experienced ones, will have an intuition about some improbable event in a target country, and will collect more data, and perhaps send out collection requests within his or her authority. These intuitions are useful just often enough that wise managers of analysts, unless the situation is absolutely critical, allow them a certain amount of freedom to explore.

Scientific Method

Astronomers and nuclear physicists, at different ends of the continuum from macroscopic to microscopic, share the method of having to infer behavior, consistent with hypothesis, not by measuring phenomena to which they have no direct access, but by measuring phenomena that can be measured and that hypothesis suggests will be affected by the mechanism of interest. Other scientists may be able to set up direct experiments, as in chemistry or biology. If the experimental results match the expected outcome, then the hypothesis is validated; if not, then the analyst must develop a new hypothesis and appropriate experimental methods[4].

In intelligence analysis, the analyst rarely has direct access to the observable subject, but gathers information indirectly. Even when the intelligence subject at hand is a technical one, analysts must remain aware that the other side may be presenting deliberately deceptive information.

From these gathered data, the analyst may proceed with the scientific method by generating tentative explanations for a subject event or phenomenon. Next, each hypothesis is examined for plausibility and compared against newly acquired information, in a continual process toward reaching a conclusion. Often the intelligence analyst tests several hypotheses at the same time, whereas the scientist usually focuses on one at a time. Furthermore, intelligence analysts cannot usually experiment directly upon the subject matter as in science, but must generate fictional scenarios and rigorously test them through methods of analysis suggested below.

Methods of Analysis

As opposed to types of reasoning, which are ways the analyst drafts the product, the following methods are ways of validating the analyst's results of reasoning.

Opportunity Analysis

Opportunity analysis identifies for policy officials opportunities or vulnerabilities that the customer’s organization can exploit to advance a policy, as well as dangers that could undermine a policy. Lawyers apply the test cui bono (who benefits?) in a rather similar way.

To make the best use of opportunity analysis, there needs to be a set of objectives for one's own country, preferably with some flexibility to them. The next step is to examine personalities and groups in that target country to see if there are any with a commonality of interest. Even though the different sides might want the same thing, it is entirely possible that one or the other might have deal-breaking conditions. If that is the case, then ways to smooth that conflict need to be identified, or no more work should be spent on that alternative.

Conversely, if there are elements that would be utterly opposed to the objectives of one's side, ways of neutralizing those elements need to be explored. They may have vulnerabilities that could render them impotent, or there may be a reward, not a shared opportunity, that would make them cooperate.

Linchpin Analysis

Linchpin analysis proceeds from information that is certain, or with a high probability of being certain. In mathematics and physics, a similar problem formation, which constrains the solution by certain known or impossible conditions, is the Boundary value problem | boundary value condition.

By starting from knowns (and impossibilities), the analyst has a powerful technique for showing consumers, peers, and managers that a problem has both been thoroughly studied and constrained to reality[20]. Linchpin analysis was introduced to CIA by Deputy Director for Intelligence (1993-1996) Doug MacEachin, as one of the "muscular" terms he pressed as an alternative to academic language, which was unpopular with many analysts. He substituted linchpin analysis for the hypotheses driving key variables. MacEachin required the hypotheses -- or linchpins -- needed to be explicit, so policymakers could be aware of coverage, and also aware of changes in assumptions.

This method is an "anchoring tool" that seeks to reduce the hazard of self-inflicted intelligence error as well as policymaker misinterpretation. It forces use of the checkpoints listed below, to be used when drafting reports:

  1. Identify the main uncertain factors or key variables judged likely to drive the outcome of the issue, forcing systematic attention to the range of and relationships among factors at play.
  2. Determine the linchpin premises or working assumptions about the drivers. This encourages testing of the key subordinate judgments that hold the estimative conclusion together.
  3. Marshal findings and reasoning in defense of the linchpins, as the premises that warrant the conclusion are subject to debate as well as error.
  4. Address the circumstances under which unexpected developments could occur. What indicators or patterns of development could emerge to signal that the linchpins were unreliable? And what triggers or dramatic internal and external events could reverse the expected momentum?

Analysis of Competing Hypotheses

See also: Analysis of Competing Hypotheses

Dick Heuer spent years in the CIA Directorate of Operations (DO) as well as the DI, and worked on methodology of analysis both in his later years and after retirement[2]. Some of his key conclusions, coming from both experience and an academic background in philosophy, include:

  1. The mind is poorly "wired" to deal effectively with both inherent uncertainty (the natural fog surrounding complex, indeterminate intelligence issues) and induced uncertainty (the man-made fog fabricated by denial and deception operations).
  2. Even increased awareness of cognitive and other "unmotivated" biases, such as the tendency to see information confirming an already-held judgment more vividly than one sees "disconfirming" information, does little by itself to help analysts deal effectively with uncertainty.
  3. Tools and techniques that gear the analyst's mind to apply higher levels of critical thinking can substantially improve analysis on complex issues on which information is incomplete, ambiguous, and often deliberately distorted. Key examples of such intellectual devices include techniques for structuring information, challenging assumptions, and exploring alternative interpretations.

In 1980, he wrote an article, "Perception: Why Can't We See What Is There to be Seen?" which suggests to Davis[20] that Heuer's ideas were compatible with linchpin analysis. Given the difficulties inherent in the human processing of complex information, a prudent management system should

  1. Encourage products that (a) clearly delineate their assumptions and chains of inference and (b) specify the degree and source of the uncertainty involved in the conclusions.
  2. Emphasize procedures that expose and elaborate alternative points of view--analytic debates, devil's advocates, interdisciplinary brainstorming, competitive analysis, intra-office peer review of production, and elicitation of outside expertise.

According to Heuer, analysts construct a reality based on objective information, filtered through complex mental processes that determine which information is attended to, how it is organized, and the meaning attributed to it. What people perceive, how readily they perceive it, and how they process this information after receiving it are all strongly influenced by past experience, education, cultural values, role requirements, and organizational norms, as well as by the specifics of the information received. To understand how the analysis results, one must use good mental models to create the work, and understand the models when evaluating it. Analysts need to be comfortable with challenge, refinement, and challenge. To go back to linchpin analysis, the boundary conditions give places to challenge and test, reducing ambiguity. '

More challenge, according to Heuer, is more important than more information. He wanted better analysis to be applied to less information, rather than the reverse. Given the immense volumes of information that modern collection systems produce, the mind is the limiting factor. Mirror-imaging is one of Heuer's favorite example of a Cognitive traps for intelligence analysis | cognitive trap, in which the analyst substitutes his own mindset for that of the target. "To see the options faced by foreign leaders as these leaders see them," according to Heuer, " one must understand [the foreign leaders'] values and assumptions and even their misperceptions and misunderstandings. ... Too frequently, foreign behavior appears "irrational" or "not in their own best interest." Projecting American values created models that were inappropriate for the foreign leader.

A significant problem during the Vietnam War is that Secretary of Defense Robert S. McNamara, an expert on statistical decision making, assumed that Ho Chi Minh, Vo Nguyen Giap, and other North Vietnamese officials would approach decision making as he did. For example, in McNamara's thinking, if the US did not attack SA-2 anti-aircraft missiles, the enemy would interpret that as "restraint" and not use them against US aircraft [21] The North Vietnamese leadership, not privy to McNamara's thinking, were unaware of the "signaling" and did their best to shoot down US aircraft with those missiles.

Heuer's answer was making the challenge of Analysis of Competing Hypotheses (ACH) the core of analysis. In ACH, there is competition among competing hypotheses of the foreign leader's assumptions, which will reduce mirror-imaging even if they do not produce the precise answer. The best use of information, in this context, is to challenge the assumption the analyst likes best.

One of the key motivations for ACH, according to Heuer, is to avoid rejecting deception out of hand, because the situation looks straightforward. Heuer observed that good deception looks real. "Rejecting a plausible but unproven hypothesis too early tends to bias the subsequent analysis, because one does not then look for the evidence that might support it. The possibility of deception should not be rejected until it is disproved or, at least, until a systematic search for evidence has been made and none has been found."

The steps in ACH areCite error: Invalid <ref> tag; invalid names, e.g. too many:

  1. Identify the possible hypotheses to be considered. Use a group of analysts with different perspectives to brainstorm the possibilities.
  2. Make a list of significant evidence and arguments for and against each hypothesis.
  3. Prepare a matrix with hypotheses across the top and evidence down the side. Analyze the "diagnosticity" of the evidence and arguments--that is, identify which items are most helpful in judging the relative likelihood of the hypotheses.
  4. Refine the matrix. Reconsider the hypotheses and delete evidence and arguments that have no diagnostic value.
  5. Draw tentative conclusions about the relative likelihood of each hypothesis. Proceed by trying to disprove the hypotheses rather than prove them.
  6. Analyze how sensitive your conclusion is to a few critical items of evidence. Consider the consequences for your analysis if that evidence were wrong, misleading, or subject to a different interpretation.
  7. Report conclusions. Discuss the relative likelihood of all the hypotheses, not just the most likely one.
  8. Identify milestones for future observation that may indicate events are taking a different course than expected.

Keith Devlin has been researching the use of mathematics and formal logic in implementing Heuer's ACH paradigm.[13]

Analogy

Analogy is common in technical analysis, but engineering characteristics seeming alike do not necessarily mean that the other side has the same employment doctrine for an otherwise similar thing. Sometimes, the analogy was valid for a time, such as the MiG-25 aircraft being designed as a Soviet counter to the perceived threat of the high-altitude, supersonic B-70 bomber. The Soviets could have cancelled the MiG-25 program when the US changed doctrines to low altitude penetration and cancelled the B-70 program, but they continued building the MiG-25.

One of the Soviet variants was a high-speed, high-altitude reconnaissance aircraft (MiG-25RB), which, for a time, was thought comparable to the US SR-71 aircraft. Several additional points of data, however, showed that an analogy between the SR-71 and MiG-25RB. HUMINT revealed that a single Mach 3.2 flight of the MiG wrecked the engines beyond hope of repair, and the cost of replacement was prohibitive unless there was no other way to get the information. The SR-71, however, could make repeated flights with the same engines. The dissimilarity of engine life was not only expensive, but meant that the MiG-25RB could operate only from bases with the capability to change engines.

Image:MiG-25RB camera 2 JNO.jpg | thumb | Reconnaissance MiG-25RB assumed disposable engines

The US had applied "reverse engineering" to the MiG, essentially saying "if we had an aircraft with such capabilities, what would we do with it?" In the fighter-interceptor role, however, the US gives the pilot considerable flexibility in tactics, where the Soviets had a doctrine of tight ground control. For the US doctrine, the aircraft was too inflexible for American fighter tactics, but made sense for the Soviets as an interceptor that could make one pass at a penetrating bomber, using an extremely powerful radar to burn through jamming for final targeting.

Many of these assumptions fell apart after Viktor Belenko flew his MiG-25 to the West, where TECHINT analysts could examine the aircraft, and doctrinal specialists could interview Belenko [22].

The Analytic Process

While the exact process used from analyst to analyst will differ somewhat across the intelligence community, there are a series of steps it is recommended that analysts useCite error: Invalid <ref> tag; invalid names, e.g. too many:

Define the Problem

Policy makers will have questions based on their intelligence requirements. Sometimes questions are clear and can easily be addressed by the analyst. Sometimes however, clarification is required due to vagueness, multiple layers of bureaucracy between customer and analyst, or due to time constraints. Just as analysts need to try to understand the thinking of the adversary, analysts need to know the thinking of their customers and allies.

Generate Hypotheses

Once the problem is defined, the analyst is able to generate reasonable hypotheses based on the question. For example, a business may want to know whether a competitor will lower their prices in the next quarter. From this problem, two obvious hypotheses are:

  1. The competitor will lower prices or
  2. The competitor will not lower prices.

However, with a little brainstorming, additional hypotheses may become apparent. Perhaps the competitor will offer discounts to long term customers, or perhaps they may even raise prices. At this point, no hypothesis should be discarded.

Determine information needs & gather information

In intelligence, Intelligence collection management|collection usually refers to the step in the formal Intelligence cycle|intelligence cycle process. In many cases, the information needed by the analyst is either already available or is already being sought by collection assets (such as spies, imagery satellites, etc). If not, the analyst may request collection on the subject, or if this is not possible identify this information gap in their final product. The analyst will generally also research other sources of info, such as open source (public record, press reporting), historical records, and various databases.

Evaluate sources

Information used for intelligence analysis (whether national, criminal, or business) has been obtained from people or organizations that are actively seeking to keep the information from the analyst, or who are providing misleading information. Adversaries do not want to be analyzed correctly by competitors. This withholding of information is known as counterintelligence, and is very different from similar fields of research, such as Scientific method|science and History|history where information may be misleading, incomplete or wrong, but rarely does the subject of investigation actively deny the researcher access. So, the analyst must evaluate incoming information for reliability (has the source reported accurate information in the past?), credibility (does the source reasonably have access to the information claimed? Has the source lied in the past?), and for possible denial and deception (even in the source is credible and reliable, they may have been fooled).

Evaluate (Test) Hypotheses

This is the step most would consider "actual analysis". Here the analyst compares the evidence gathered against his or her hypotheses, using various analytic tools and methods such as Analysis of Competing Hypotheses or link charts. Many hypotheses may be quickly discarded, while for others there may simply not be enough information one way or the other to evaluate them.

Production and Packaging

Once hypotheses have been evaluated, the intelligence product must created for the consumer. Three key features of the intelligence product are:

  • timeliness. Timeliness includes not only the amount of time required to deliver the product, but also the usefulness of the product to the customer at a given moment.
  • scope. Scope involves the level of detail or comprehensiveness of the material contained in the product.
  • periodicity. Periodicity describes the schedule of product initiation and generation.

Government intelligence products are typically packaged as highly structured written and oral presentations, including electrical messages, hardcopy reports, and briefings. Many organizations also generate video intelligence products, especially in the form of live daily “newscasts,” or canned documentary presentations.

Analysts should understand the relationship between the analyst's and the consumer's organization. There may be times that while the ultimate consumer and originating analyst simply want to pass information, a manager in either chain of command may insist on a polished format. Analysts should also be aware of any staff through which the material will have to pass before reaching the end consumer. If staff may return return a report unless it meets some artificial criterion.

Peer Review

If the information is not disseminated, it is useless, and dissemination includes people who need it within the intelligence agency. "Coordination with peers is necessary...If you think you are right, and the coordinator disagrees, let the assessment reflect that difference of opinion and use a footnote, called a reclama [23], inside the US intelligence community if necessary. But never water down your assessment to a lowest common denominator just to obtain coordination.When everyone agrees on an issue, something probably is wrong. "As an example, following the collapse of the Soviet Union, there was an almost unanimous belief that large numbers of Russian ballistic missile specialists would flood into the Third World and aid missile programs in other states (the so-called brain drain). ...As it turned out, there was no [expected] mass departure of Russian missile specialists, but Russian expertise was supplied to other states in ways that had been ignored due to the overemphasis on the brain drain.

In large intelligence establishments, analysts have peers at other agencies. The practical amount of coordination, indeed inside one's own agency, will depend on the secure collaboration tools available (wikis, analyst webpages, email), the schedule and availability of the other analysts, any restrictions on dissemination of the material, and the analyst's ability to play nicely with others. Extremely specialized issues might have very few people who could meaningfully look at it.

An intelligence community document, as opposed to a spot report from a single agency, is expected to be coordinated and reviewed. For example, in reports on the Iraqi WMD program, given a field report that aluminium tubes were on order, which might have been received both at the geographic desk and the Counterproliferation Center, someone might have thought they were for use in uranium separation centrifuges. It has been reported that some analysts thought they might be used for rocket casings, which apparently was the correct interpretation. The question needs to be asked "did the original analyst contact a technical specialist in Isotope separation#Centrifugal effect | separation centrifuges, perhaps at Department of Energy intelligence?"

Such an analyst might have mentioned that while aluminium has been used, maraging steel is the material of choice for Zippe-type centrifuges. The alternative, the Helikon vortex separation process , has no moving parts and thus less demand on the tubes, but takes much more energy. If the Helikon had been under consideration, the consultation could have gone farther, perhaps to IMINT analysts familiar with power generation in the area or Electro-optical_MASINT#Infrared_MASINT | infrared MASINT specialists who could look for the thermal signature of power generation or the cascade itself. Both Zippe and Helikon techniques take a great deal of energy, and often have been placed near hydroelectricity | hydroelectric dam power plants so power will be nearby.

Customer Feedback and Production Evaluation

The production phase of the intelligence process does not end with delivering the product to the customer. Rather, it continues in the same manner in which it began: with interaction between producer and customer. For the product to be useful, the analyst and policymaker need to hear feedback from one another, and they refine both analysis and requirements.

Feedback procedures between producers and customers should include key questions, such as: Is the product usable? Is it timely? Was it in fact used? Did the product meet expectations? If not, why not? What next? The answers to these questions will lead to refined production, greater use of intelligence by decisionmakers, and further feedback sessions. Thus, production of intelligence actually generates more requirements in this iterative process.

Never Forget the End User

The end user wants information that he needs to know, but, unless he happens to have an interest in a particular subject, does not care how much you know about the topic. William Donovan, the head of the WWII OSS, began to get FDR's ear because he gave vividly illustrate, well-organized briefings that would be common today, but were unprecedented in WWII. Today, there is danger of becoming too entranced with the presentation and less with its subject. This is also a delicate dance of overemphasizing the subjects that interest high officials, and what they want to hear declared true about them, rather than hearing what the analysts believe is essential.

Most consumers do not care how attractive a report looks or whether the format is correct. I have lost count of the number of times consumers have told me they do not care if an assessment has a CIA seal on it, if it is in the proper format, or even if it has draft stamped all over it; they just want the assessment in their hands as soon as possible, at least in time to help make a decision." Unfortunately, a number of mid-level managers get overly worried about form, and wise top-level intelligence officials make sure that does not happen.[7]

References

  1. Lee, Bartholomew (2006), Radio Spies - Episodes in the Ether Wars
  2. 2.0 2.1 2.2 Heuer, Richards J. Jr. (1999), Psychology of Intelligence Analysis. Chapter 2. Perception: Why Can't We See What Is There To Be Seen?. Retrieved on 2007-10-29
  3. North Atlantic Treaty Organization (November 2001), NATO Open Source Intelligence Handbook. Retrieved on 2007-10-23
  4. 4.0 4.1 4.2 4.3 4.4 4.5 4.6 Krizan, Lisa (June 1999), Intelligence Essentials for Everyone, Joint Military Intelligence College. Retrieved on 2007-10-23 Cite error: Invalid <ref> tag; name "Krizan1999" defined multiple times with different content Cite error: Invalid <ref> tag; name "Krizan1999" defined multiple times with different content Cite error: Invalid <ref> tag; name "Krizan1999" defined multiple times with different content
  5. 5.0 5.1 5.2 Central Intelligence Agency, Directorate of Intelligence (February 1997), A Compendium of Analytic Tradecraft Notes. Retrieved on 2007-12-03
  6. Davis, Jack (1995), "A Policymaker's Perspective On Intelligence Analysis", Studies in Intelligence (no. 5)
  7. 7.0 7.1 7.2 7.3 7.4 Watanabe, Frank (1997), "Fifteen Axioms for Intelligence Analysts: How To Succeed in the DI [Directorate of Intelligence]", Studies in Intelligence. Retrieved on 2007-10-23
  8. Kalisch, Robert B. (July-August 1971), "Air Force Technical Intelligence", Air University Review. Retrieved on 2007-10-27
  9. May, Ernest R. & Philip D. Zelikow (1996), The Kennedy Tapes: Inside the White House During the Cuban Missile Crisis. Retrieved on 2007-10-23
  10. Morgan, Brent A. (September 1995), Employment of Indications and Warning Intelligence Methods to Forecast a Potentially Hostile Revolution in Military Affairs., US Naval Postgraduate School
  11. Fred Ikle (2005). Every War Must End. Columbia University Press. 
  12. Johnston, Rob (2005), Analytic Culture in the US Intelligence Community: An Ethnographic Study, Center for the Study of Intelligence, Central Intelligence Agency. Retrieved on 2007-10-29
  13. 13.0 13.1 Devlin, Keith (July 15, 2005), Confronting context effects in intelligence analysis: How can mathematics help?, Center for the Study of Language and Information, Stanford University
  14. Luke, Allan, Introduction: Theory and Practice in Critical Discourse Analysis, International Encyclopedia of the Sociology of Education
  15. Crichton, Michael (1970). Five Patients. Ballantine Books. ISBN 0345354648. 
  16. Robert Mathams (1995), The Intelligence Analyst’s Notebook, in Douglas Dearth and R. Thomas Goodden, Strategic Intelligence: Theory and Application, Joint Military Intelligence Training Center
  17. Ontology. A Resource Guide for Philosophers, September 8, 2008
  18. National Intelligence Production Board (2001), Strategic Investment Plan for Intelligence Community Analysis, NIPB-2001. Retrieved on 2007-10-28
  19. Priest, Dana & Walter Pincus (May 19, 2006), "Nominee Has Ability To Bear Bad News: Some Senators Unsure He Will Use It With Bush", Washington Post. Retrieved on 2007-10-28
  20. 20.0 20.1 Davis, Jack (1999), Improving Intelligence Analysis at CIA: Dick Heuer's Contribution to Intelligence Analysis, Davis 1999. Retrieved on 2007-10-27 Cite error: Invalid <ref> tag; name "Davis1999" defined multiple times with different content
  21. McMaster, H. R. (1998), Dereliction of Duty: Johnson, McNamara, the Joint Chiefs of Staff, and the Lies That Led to Vietnam, Harper Perennial
  22. Barron, John (1983), Mig Pilot: The Final Escape of Lt. Belenko, ISBN 0380538687
  23. US Department of Defense (12 July 2007), Joint Publication 1-02 Department of Defense Dictionary of Military and Associated Terms. Retrieved on 2007-10-01