Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Log analysis is the backbone of both proactive security monitoring and reactive forensic investigation. When you're analyzing a breach, responding to an incident, or hunting for threats, logs are your primary evidence source. They tell the story of what happened, when, and often how. You'll be tested on understanding not just what each technique does, but how they work together in a complete security workflow, from collection and normalization to correlation and forensic reconstruction.
Don't just memorize definitions here. Know which techniques address data preparation challenges versus threat detection goals versus legal and compliance requirements. Understanding the purpose behind each technique helps you answer scenario-based questions like "What should an analyst do first when investigating X?" or "Which technique would reveal Y type of attack?"
Before you can analyze anything, logs must be collected, formatted consistently, and time-aligned. These foundational techniques transform raw, chaotic data into usable evidence.
Logs come from dozens of sources: firewalls, routers, servers, applications, endpoint agents, and security tools. Collection and aggregation funnels all of these into a single repository so nothing gets overlooked.
Different devices produce logs in wildly different formats. A Cisco firewall log looks nothing like a Windows Event Log or an Apache access log. Normalization solves this.
If your firewall says an event happened at 14:02:03 but your server says the related event happened at 14:05:47, you can't tell which came first unless both clocks are accurate. Time synchronization, typically through NTP (Network Time Protocol), keeps all system clocks aligned.
Compare: Log normalization vs. time synchronization: both prepare data for analysis, but normalization addresses format consistency while synchronization addresses temporal accuracy. An FRQ might ask which you'd prioritize when logs come from systems in different countries (time sync) versus different vendors (normalization).
Once data is prepared, these techniques identify malicious activity by recognizing patterns, filtering noise, and connecting related events. This is where raw data becomes actionable intelligence.
These are two distinct but complementary approaches to finding threats:
' OR 1=1, or known malware callback domains. It's fast and reliable for known threats.The tradeoff: signature-based detection has low false-positive rates but misses novel attacks. Anomaly detection catches novel attacks but generates more false positives.
Raw logs contain enormous amounts of noise: routine heartbeats, informational status messages, and known-good traffic. Filtering cuts through this.
source_ip="10.0.0.5" AND event_type="authentication_failure"This is one of the most powerful techniques in log analysis. Individual log entries rarely tell a complete story, but correlation connects them.
Compare: Pattern recognition vs. event correlation: pattern recognition identifies individual suspicious events, while correlation connects multiple events into attack narratives. If an FRQ describes a multi-stage attack, correlation is your answer. If it asks about detecting a specific exploit signature, pattern recognition applies.
When incidents occur, these techniques support detailed reconstruction and legal proceedings. Forensic analysis requires both technical accuracy and evidentiary integrity.
A forensic timeline merges timestamps from all relevant log sources into a single chronological sequence. This is different from simply reading one log file top to bottom.
Building a forensic timeline typically follows these steps:
Sophisticated attackers often target logs to cover their tracks. If an attacker gains root access, one of their first moves may be clearing or modifying log files.
Network logs capture communication between systems, which is essential for detecting lateral movement and data exfiltration.
While network logs show traffic between hosts, system and application logs reveal what happened on each individual host.
Compare: Network traffic logs vs. system logs: network logs show communication patterns between hosts, while system logs reveal what happened on individual hosts. A complete investigation requires both. Network logs might show data leaving the network; system logs show what process sent it and under which user account.
Specialized security tools generate their own logs that require dedicated analysis approaches. IDPS logs provide high-fidelity threat data but need context from other sources to be truly useful.
IDS/IPS sensors sit on network segments or hosts and generate alerts based on signature matches and protocol anomalies.
A SIEM is the enterprise platform that brings together most of the individual techniques covered in this guide.
Compare: SIEM vs. individual log analysis techniques: a SIEM combines collection, normalization, correlation, and alerting into one platform, while individual techniques can be performed manually or with specialized tools. Understanding the component techniques helps you troubleshoot when SIEM correlation rules produce unexpected results or miss threats.
These techniques ensure log analysis meets organizational and legal requirements while communicating findings effectively.
Not all logs need to be instantly searchable, but they do need to exist when an investigation or audit demands them.
Compliance isn't optional. Failure to maintain required logs can result in fines, legal liability, and loss of certifications.
Raw log data is powerful for analysts, but findings need to be communicated to people who won't read through thousands of entries.
Compare: Retention policies vs. integrity controls: both support forensic validity, but retention ensures logs exist when needed, while integrity ensures logs are trustworthy. Compliance questions often test whether you understand that having logs isn't enough; they must also be authentic and unmodified.
| Concept | Best Examples |
|---|---|
| Data Preparation | Log collection, normalization, time synchronization |
| Threat Detection | Pattern recognition, anomaly detection, event correlation |
| Search Optimization | Log filtering, Boolean queries, field-specific searches |
| Forensic Reconstruction | Timeline creation, system log analysis, network traffic analysis |
| Evidence Integrity | Tamper detection, cryptographic hashing, write-once storage |
| Enterprise Platforms | SIEM systems, IDPS integration |
| Governance | Retention policies, compliance requirements, audit trails |
| Communication | Visualization dashboards, executive reporting |
Which two techniques must be completed before event correlation can be effective? Explain why each is a prerequisite.
An attacker compromised a system and deleted local logs before exfiltrating data. Which techniques would have prevented or detected this anti-forensic activity?
Compare and contrast pattern recognition with anomaly detection. When would each be more effective at identifying a novel attack?
A forensic investigator notices a 15-minute gap in firewall logs during the suspected breach window. Which technique addresses this concern, and what might the gap indicate?
Your organization must demonstrate to auditors that authentication logs from the past 18 months are complete and unmodified. Which three techniques or controls would you reference in your response?