RemNote Community
Community

Study Guide

📖 Core Concepts Intelligence Collection Management (ICM) – The coordinated process of acquiring raw intelligence from all sources; it does not analyze significance, only validates accuracy. Validation vs. Analysis – Validation: check raw data for factual correctness. Analysis: interpret validated data to produce actionable insight. Collection Disciplines – Specialized domains (e.g., Cyber, FININT, GEOINT, HUMINT, IMINT, MASINT, OSINT, SIGINT, TECHINT) each exploiting a different type of source or sensor. Collection Guidance – Director‑level tasking that matches mission requirements (what, priority, secrecy) with asset capabilities (what each sensor/collector can do). Utility Assessment – A model that scores each possible asset (or asset mix) for how well it satisfies the mission; the highest‑utility option is selected. Priority Intelligence Requirements (PIRs) – NATO’s top‑level questions that drive all subsequent Information Requirements and discipline selection. Source Rating System (U.S.) – Two‑part: Reliability (letter A–F) + Information Validity (number 1–6). Source Types – Primary: direct knowledge; Secondary: information twice‑removed. Proximity and Appropriateness further qualify reliability. Plausibility & Expectability – Plausibility: certainty level (certain, uncertain, impossible) after checking for deception. Expectability: does the info logically follow from known facts? --- 📌 Must Remember Rating format: A‑1 = highly reliable source + highly reliable info; E‑5 = questionable source + doubtful info. Directive = short, prioritized order; used when one customer dominates and a single method suffices. Request = most common; involves a requester‑collector relationship; may be solicited (collector asks for a tailored request). Inventory of Needs = standing list of unmet intelligence gaps; not addressed to any specific collector. Asset Utility Rule – If multiple assets meet the requirement, pick the one with the higher utility score (greater relevance, lower cost, lower risk). Alternative Discipline Decision – Substitute platform only after evaluating weather, terrain, enemy counter‑measures, and orbital constraints. High‑Volume Platform Rule – Provide raw data only if network bandwidth and analyst capacity can handle it; otherwise deliver analytic summaries. --- 🔄 Key Processes Mission Requirement Definition Identify data type, priority, secrecy level. Asset Capability Matching Asset managers list sensor/collection capabilities. Utility Assessment Model Compare mission specs vs. asset specs → utility score. Asset Selection Choose single asset or asset mix with highest utility; rank alternatives if needed. Source Anonymization Split report into: (a) true source identity → removed, (b) pseudonym/code name, (c) content. Source Rating Evaluation Assess reliability (history, proximity, expertise) → letter. Assess validity (corroboration, consistency) → number. Plausibility Assessment Determine if info is certain, uncertain, or impossible; check for deception. Expectability Check Ask: “Does this logically follow from what we already know?” Confirmation Decision Assign responsibility (analyst, collector, or both) to verify the report. --- 🔍 Key Comparisons Validation vs. Analysis – Validation = “Is it true?”; Analysis = “What does it mean?” Directive vs. Request – Directive = top‑down, single‑customer order; Request = collaborative, most common. Primary vs. Secondary Source – Primary = direct observation; Secondary = information twice removed. A Rating vs. E Rating – A = fully trusted; E = questionable reliability. Raw Data vs. Analytic Report – Raw data = full volume, needs bandwidth; Analytic report = distilled, lower bandwidth. --- ⚠️ Common Misunderstandings “Collection = Analysis” – Collection stops at validated data; analysis is a separate phase. “Higher source rating always means accurate info.” – Rating reflects source trustworthiness, not the truth of a specific piece (that's the validity number). “All platforms can be used regardless of weather.” – Weather, terrain, and enemy defenses can render a platform ineffective. “Only one asset is ever needed.” – Missions often require a combination of assets for full coverage. --- 🧠 Mental Models / Intuition Match‑Making Model – Think of ICM as a dating app: mission requirements are the “profile” and collection assets are the “candidates”. The utility score is the “compatibility rating”. Two‑Axis Rating Grid – Plot reliability (A‑F) on the vertical axis and validity (1‑6) on the horizontal; the quadrant tells you overall confidence. Plausibility Triangle – Vertices = Certain, Uncertain, Impossible; any new info lands inside this triangle after checking deception. --- 🚩 Exceptions & Edge Cases Preferred Platform Unavailable – Substitute satellites, MASINT sensors, or HUMINT after evaluating counter‑measures and environmental limits. High‑Volume Data Overload – If receivers/analyst teams cannot handle the flow, switch to analytic summaries instead of raw feeds. Adversary Counter‑Measure – If a broken cryptosystem is known, the adversary may change the system or inject disinformation. --- 📍 When to Use Which Discipline Selection – Use the discipline whose sensor best matches the PIR and current conditions (weather, terrain, enemy defenses). Directive vs. Request – Issue a directive when a single commander controls the mission and a specific, limited method suffices; otherwise use a request. Raw Data vs. Analytic Report – Choose raw data when bandwidth > 10 Mbps and analyst capacity is high; otherwise deliver analytic reports. Single Asset vs. Asset Mix – If a single asset meets ≥ 80 % of utility criteria, use it; if not, combine assets to cover gaps. --- 👀 Patterns to Recognize Multi‑Source Convergence – Traffic analysis + IMINT + SIGINT pointing to the same event → high confidence of an imminent action. High Reliability + Low Proximity – An “A‑4” rating (trusted source but distant) warns that info may be accurate but not timely. Bandwidth Bottleneck – Large‑volume platforms (e.g., SAR satellites) paired with limited receivers → likely to trigger “analytic summary” guidance. --- 🗂️ Exam Traps Distractor: “A source rating of A guarantees the information is correct.” → Wrong; validity number still required. Distractor: “If weather is bad, only HUMINT can be used.” → Incorrect; MASINT or satellite with different wavelengths may still work. Distractor: “Directives are always better than requests.” → Not true; directives are only optimal with a single dominant customer and limited method. Distractor: “Plausibility means the information is true.” → Plausibility only assesses certainty level, not proven truth. Distractor: “All collection assets should always send raw data.” → Bandwidth and analyst overload may force summary delivery.
or

Or, immediately create your own study flashcards:

Upload a PDF.
Master Study Materials.
Start learning in seconds
Drop your PDFs here or
or