Foothold OSINT
OSINT Methodology for Security Professionals

OSINT Methodology for Security Professionals

A structured OSINT methodology for security professionals — from requirements through evidence packaging, source diversity, and confidence vocabulary across pentest, bug bounty, and threat intel.

A Practical OSINT Methodology Framework for Security Professionals

Open-source intelligence is not a tool. It is not a plugin, a platform subscription, or a one-liner that pipes Shodan results into a Slack channel. It is a repeatable operating model — a disciplined sequence of decisions about what to look for, where to look, how to weigh what you find, and how to communicate it to the people who need to act on it.

This post lays out that operating model in full. Whether you are conducting pre-engagement reconnaissance for a red team engagement, mapping attack surface for bug bounty, attributing a threat actor campaign, or enriching indicators of compromise for a detection engineering team, the underlying methodology is the same. The tools and sources change. The mental model does not.


Why Methodology Matters Before Tools

Most practitioners start with tools. They run Amass, check theHarvester, query VirusTotal, and produce a spreadsheet of hostnames. That is reconnaissance activity. It is not intelligence.

Intelligence requires a requirement — a specific question the collection is meant to answer. Without a requirement, you collect everything and understand nothing. You burn time on data that will never influence a decision, miss data that would, and deliver a report that nobody acts on.

A formal methodology forces you to answer three questions before you open a single browser tab:

  1. What decision does this intelligence need to support?
  2. What is the minimum confidence threshold that decision requires?
  3. What would change my assessment?

With those three questions answered, every subsequent collection and analysis step has a purpose.


The Operating Model: Six Phases

Phase 1 — Requirements Definition

The intelligence cycle begins with a requirement. In practice, requirements come in two forms: standing requirements (ongoing monitoring of a persistent threat or attack surface) and tasked requirements (a specific question tied to a time-bounded engagement or incident).

Write the requirement as a complete sentence. “Conduct OSINT on Acme Corp” is not a requirement. “Identify externally exposed administrative interfaces and associated credential exposure for Acme Corp’s production infrastructure ahead of a black-box external penetration test” is a requirement.

A well-formed requirement has:

Documenting the requirement formally also protects you legally and ethically. It creates a record that collection was authorized and scoped.

Phase 2 — Source Planning

Before you query anything, map your source categories against your requirement. This is the source-diversity principle: no single-source finding should drive a high-confidence assessment. Corroboration across independent source types is what converts a data point into intelligence.

Practical source categories for security OSINT include:

CategoryExamples
Passive DNS / certificate transparencySecurityTrails, crt.sh, Censys
Internet-wide scan dataShodan, Censys, FOFA, BinaryEdge
Code and configuration repositoriesGitHub, GitLab, Bitbucket, npm
Breach and credential datasetsHave I Been Pwned API, aggregated breach data (where legally accessible)
Domain and registrant historyWHOIS history, DomainTools (paid), RiskIQ
Social and professional networksLinkedIn, X/Twitter, GitHub profiles
Paste and leak sitesPastebin, BreachForums (monitoring only, within legal boundaries)
Dark web forums and marketsTor-accessible sources (legal monitoring context only)
Malware repositories and sandboxesVirusTotal, MalwareBazaar, Any.run, Hybrid Analysis
Threat intelligence feedsMISP communities, OTX AlienVault, isac feeds
News, filings, and official recordsSEC EDGAR, Companies House, press releases, job postings

For each source category, note whether it provides direct evidence (the thing itself, such as a hostname in a certificate SAN) or circumstantial evidence (a pattern that implies something, such as a job posting that implies a technology stack). Direct evidence has higher baseline weight.

Also note recency. A Shodan banner from eighteen months ago tells you something different than one from last week. Timestamp every finding during collection.

Phase 3 — Structured Collection

Collection should be systematic, not opportunistic. Create a collection matrix before you start: rows are your information requirements, columns are your source categories. Work through cells deliberately. This prevents the common failure mode of spending four hours going deep on one interesting thread while leaving entire source categories untouched.

Practical collection hygiene:

Phase 4 — Analysis and the Confidence Vocabulary

This is where most practitioner methodologies break down. Raw findings get copy-pasted into a report with no analytical layer. The reader cannot tell what is confirmed, what is inferred, and what is speculation.

A structured confidence vocabulary solves this. The vocabulary below draws from intelligence community practice adapted for operational security contexts.

The Five-Layer Model

Observable Something directly witnessed in collected data, with a specific source and timestamp. No inference required. Example: “The certificate at SAN mail.acme-corp.com was issued on 2024-11-14, as observed in crt.sh on 2025-01-28.” Observables are facts about what was collected. They do not make claims about what those facts mean.

Likely Association A connection between two or more observables that is supported by evidence but involves one inferential step. Confidence is moderate. Example: “The registrant email address observed in historical WHOIS for acme-legacy[.]net matches the address observed in a 2023 credential breach dataset, suggesting the same individual manages both domains.” The word suggesting is load-bearing. Use it intentionally.

Confidence Assessment A statement about what the totality of collected evidence indicates, with an explicit confidence level. Use a labeled scale — not percentages, which imply false precision, but named levels such as Low / Moderate / High with defined criteria:

Alternatives Explicit statement of competing hypotheses you considered and why you assessed them as less likely. This is the most commonly skipped step and the most analytically valuable. If you cannot articulate an alternative explanation for your key findings, you have not analyzed them — you have confirmed your assumptions.

Collection Gaps Explicit statement of what you did not collect, could not collect, or chose not to pursue, and how those gaps affect confidence. Example: “No dark web forum monitoring was conducted for this engagement. The absence of credential leak findings should not be interpreted as evidence of no exposure.” This protects the consumer from acting on absence-of-evidence as evidence-of-absence.

Documenting all five layers for every significant finding is what separates an intelligence product from a data dump.

Phase 5 — Production: The Evidence Package

The output of an OSINT engagement is an evidence package, not a list of hostnames. A complete evidence package contains:

  1. Requirement restatement — confirms scope and authorization
  2. Collection log — timestamped record of sources queried, what was found, and what was not found
  3. Key findings — significant observables, each with source citations and timestamps
  4. Analytical assessments — findings interpreted through the confidence vocabulary, with alternatives and gaps
  5. Recommendations — specific, actionable guidance tied to the decision context defined in the requirement
  6. Raw evidence artifacts — screenshots, cached pages, certificate details, WHOIS records, stored locally and hashed for integrity

Hash your artifacts. SHA-256 the screenshots and exported data files at the time of collection. If the finding becomes relevant to an incident, legal proceeding, or client dispute, you need to demonstrate that the evidence has not been modified.

Phase 6 — Dissemination and Feedback

Deliver the evidence package to the defined consumer and capture their feedback. Did the intelligence answer the requirement? Were confidence levels calibrated correctly relative to what the consumer needed to decide? Were the collection gaps significant in practice?

This feedback loop is what makes methodology improve over time. Without it, each engagement starts from scratch.


Applying the Methodology Across Use Cases

Pentest Pre-Engagement Reconnaissance

The requirement here is typically: identify the realistic external attack surface and credential exposure before the authorized test begins.

Source priorities: certificate transparency, passive DNS, internet-wide scan data, code repositories, breach datasets.

The confidence vocabulary matters because your findings inform scope decisions. If you assess with high confidence that a specific IP range belongs to the target, the test proceeds. If your confidence is moderate because the IP is associated with a shared hosting provider, that uncertainty needs to surface before active testing begins. Misattributing in-scope infrastructure has legal consequences.

Collection gaps to always document: internal network assets are by definition inaccessible to external OSINT. What you cannot see from outside is a gap, not an absence.

Bug Bounty Reconnaissance

The requirement shifts to: identify assets within the defined program scope that may present exploitable attack surface.

Bug bounty reconnaissance introduces a specific constraint: scope rules. The confidence vocabulary is critical here for scope attribution. Just because a subdomain resolves to an IP associated with the target organization does not place it in scope. The association needs to be high-confidence and cross-validated before you test against it.

Source priorities are similar to pentest pre-engagement, with additional weight on acquisition records (a company that acquired another may have inherited its infrastructure) and job postings (technology stack inference).

Threat Actor Attribution

Attribution is where the confidence vocabulary does the most work, because the stakes of getting it wrong are highest. Misattribution drives incorrect defensive posture, potentially implicates innocent parties, and can have policy implications.

The requirement here is: assess with what confidence we can associate observed malicious activity with a specific threat actor or cluster.

Source priorities: malware repositories and sandbox reports, passive DNS and certificate infrastructure, code repository artifacts (leaked tools, distinctive compilation artifacts), breach forum activity (monitoring context), prior threat intelligence reporting from independent vendors.

The alternatives layer is non-negotiable for attribution. The well-documented false-flag problem — where adversaries deliberately use TTPs associated with other actors — means that any attribution assessment must explicitly address: “What evidence would look exactly the same if a different actor deliberately mimicked these TTPs?”

Use the tiered language carefully. “We assess with moderate confidence that this infrastructure cluster overlaps with previously reported APT infrastructure” is a defensible statement. “This is [named actor]” without the confidence vocabulary is not intelligence — it is a claim.

IOC Enrichment

The requirement here is operational: determine whether a given indicator — IP address, domain, hash, email address — is associated with malicious activity and with what confidence, so detection and response decisions can be made.

Source priorities: VirusTotal and other multi-AV aggregators, passive DNS, WHOIS history, malware sandbox detonation reports, threat intelligence feed cross-reference, MISP communities.

The confidence vocabulary maps directly onto detection engineering decisions. A high-confidence malicious indicator justifies an automatic block. A moderate-confidence indicator might justify an alert-and-monitor posture. A low-confidence indicator with significant alternative explanations should not be used to block production traffic.

Collection gaps are especially relevant for IOC enrichment: many IP addresses and domains are reused across multiple actors and campaigns. Absence of prior reporting does not mean clean. Always document what you did not check.


Common Methodology Failures

Confirmation bias in collection. Collecting more sources that confirm an initial hypothesis while under-collecting sources that might challenge it. The structured collection matrix and the alternatives step are the primary mitigations.

Timestamp neglect. Treating historical data as current. A domain that resolved to a malicious IP eighteen months ago may now be pointing to legitimate infrastructure. Every finding needs a recency assessment.

Single-source high confidence. Using confident language for a finding supported by only one source. The confidence vocabulary definitions above make this structurally impossible if followed correctly.

Gap non-disclosure. Delivering an evidence package without collection gaps documented. Consumers make decisions based on what they assume you checked. If you did not check it, they need to know.

Tool-first thinking. Running every tool you have against a target without a requirement to guide collection. This produces data, not intelligence, and creates noise that obscures the significant findings.


Building the Habit

The methodology described here adds overhead relative to unstructured reconnaissance. That overhead pays for itself in three ways: it produces better intelligence, it protects you legally and professionally, and it builds transferable skills that do not depreciate when a specific tool gets blocked or a data source goes offline.

Start with the requirement. Write it down. Define your confidence threshold before you collect. Use the five-layer vocabulary on every significant finding. Document your gaps. Hash your evidence.

Do that consistently across engagements and you will have built something more valuable than a toolchain — you will have built an OSINT methodology for security professionals that produces reliable, defensible, actionable intelligence regardless of which specific sources and tools the next engagement requires.


This post is descriptive and comparative in nature. Nothing here constitutes legal advice. Always ensure OSINT collection activities are authorized and conducted in compliance with applicable laws and the terms of service of data sources.