← Back to Resources
Nov 8, 2025 · 16 min read · Alex Rodriguez

Real-Time vs Batch Transaction Monitoring: Choosing the Right Architecture

Should your AML system process transactions in real-time or in batches? This architectural decision profoundly impacts detection capabilities, operational costs, and regulatory compliance. Here's how to choose the right approach for your institution.

The Fundamental Architecture Question

When designing an anti-money laundering transaction monitoring system, one of the most critical architectural decisions is whether to process transactions in real-time or in batches. This choice cascades through every aspect of your system: infrastructure costs, detection capabilities, operational workflows, and regulatory compliance posture.

The traditional approach has been batch processing: collect transactions throughout the day, run detection algorithms overnight, and present alerts to analysts the next morning. But as technology has evolved and money laundering schemes have become more sophisticated, real-time monitoring has become increasingly attractive. The question is no longer whether real-time monitoring is possible, but whether it's the right choice for your specific requirements.

Key Insight

The real-time vs batch decision isn't binary. Most sophisticated AML systems use a hybrid approach, processing some scenarios in real-time while batching others. The art is knowing which transactions require immediate attention and which can wait.

Understanding Batch Processing

Batch processing analyzes transactions in groups at scheduled intervals—typically daily, but sometimes hourly or weekly depending on the scenario. This approach has dominated AML systems for decades, and for good reasons:

Advantages of Batch Processing

  • Complete Data Access: Batch systems can analyze an entire day's transactions together, detecting patterns that emerge across the full dataset. Structuring schemes that involve multiple transactions throughout the day become apparent when you see the complete picture
  • Computational Efficiency: Processing in bulk allows for optimization techniques like vectorization and parallel processing across entire datasets. You can run expensive machine learning models that would be too slow for per-transaction processing
  • Resource Predictability: Infrastructure costs are predictable because you know exactly when processing will occur. You can spin up compute resources during off-hours when cloud costs are lower
  • Simpler Architecture: Batch systems are conceptually simpler to build and maintain. There's no need for complex streaming infrastructure, state management, or exactly-once processing guarantees
  • Easier Testing: Batch jobs are deterministic and reproducible. You can rerun yesterday's batch with modified rules to test changes before deploying to production

Disadvantages of Batch Processing

  • Detection Latency: By definition, batch processing introduces delay. A suspicious transaction at 9 AM won't generate an alert until the next day, giving criminals a 24+ hour head start
  • Limited Intervention Capability: You can't stop a suspicious transaction if you don't know about it until after it's completed. For high-risk scenarios, this delay can be unacceptable
  • Batch Window Constraints: All processing must complete within the batch window. As transaction volumes grow, you may struggle to finish processing before the next batch arrives
  • Resource Spikes: Batch jobs create significant infrastructure load during processing windows, requiring over-provisioning to handle peak capacity that sits idle most of the time
18-36h
typical detection latency
for batch systems
60%
lower infrastructure cost
vs equivalent real-time system
99.9%
batch processing reliability
mature, proven technology

Understanding Real-Time Processing

Real-time (or near-real-time) processing analyzes each transaction as it occurs, generating alerts within seconds or minutes. This approach requires fundamentally different architecture but offers unique capabilities:

Advantages of Real-Time Processing

  • Immediate Detection: Suspicious activity is identified within seconds, enabling rapid response. For scenarios like account takeover or mule account detection, this speed is crucial
  • Transaction Intervention: Real-time systems can block or delay transactions before they complete, preventing fraud and money laundering rather than just detecting it after the fact
  • Continuous Resource Utilization: Compute resources are used smoothly throughout the day rather than spiking during batch windows. This can be more cost-effective at scale
  • Better Customer Experience: For legitimate transactions incorrectly flagged (false positives), real-time systems can provide immediate feedback, allowing quick resolution rather than surprising customers days later
  • Incremental Pattern Detection: Some patterns are easier to detect in real-time. For example, velocity checks ("too many transactions in the past hour") are natural in streaming architectures

Disadvantages of Real-Time Processing

  • Architectural Complexity: Real-time systems require sophisticated streaming infrastructure (Kafka, Flink, Kinesis), state management, and careful handling of out-of-order events
  • Limited Context Window: When processing transaction N, you can't see transaction N+1 that happens five minutes later. Some patterns that are obvious in batch become harder to detect in streams
  • Computational Constraints: Every model must complete within your latency SLA (typically < 100ms for inline processing). Complex machine learning models may be too slow for per-transaction execution
  • Higher Infrastructure Costs: Real-time systems must provision for peak transaction volumes 24/7, potentially increasing infrastructure costs significantly
  • Stateful Complexity: Maintaining state across distributed streaming systems is challenging. Features like "number of transactions this week" require careful state management and windowing
// Example: Real-time velocity check in streaming architecture
// Using Apache Flink-style processing

stream
  .keyBy(txn => txn.accountId)
  .window(TumblingEventTimeWindows.of(Time.hours(1)))
  .aggregate(new VelocityAggregator())
  .filter(velocity => velocity.count > threshold)
  .map(createAlert)
  .addSink(alertSink);

// The same logic in batch:
SELECT 
  account_id,
  COUNT(*) as txn_count,
  SUM(amount) as total_amount
FROM transactions
WHERE timestamp >= NOW() - INTERVAL '1 hour'
GROUP BY account_id
HAVING COUNT(*) > threshold;

Hybrid Architectures: The Best of Both Worlds

In practice, most sophisticated AML systems use a hybrid approach, combining real-time and batch processing based on scenario requirements. This tiered architecture processes transactions through multiple layers:

Tier 1: Inline Real-Time (< 100ms)

  • Simple rule-based checks (sanctions screening, amount thresholds)
  • Velocity checks using recent state (transactions per hour)
  • Known bad actor detection (blocklists, previous fraudsters)
  • Basic anomaly flags that inform downstream processing
  • Can block transactions before completion

Tier 2: Near-Real-Time (1-15 minutes)

  • Sophisticated ML models (too slow for inline)
  • Graph analysis of transaction networks
  • Cross-account pattern detection
  • Behavioral anomaly detection
  • Generates alerts for immediate investigation

Tier 3: Batch Processing (Daily/Weekly)

  • Complex temporal pattern detection
  • Comprehensive network analysis
  • Computationally expensive deep learning models
  • Historical comparison and trending
  • Regulatory reporting and analytics

This tiered approach lets you apply the right processing model to each scenario. High-risk scenarios that benefit from immediate intervention get real-time treatment, while complex pattern detection that requires full context runs in batch. Most transactions flow through all three tiers, with each layer adding progressively more sophisticated analysis.

Decision Framework: When to Use Each Approach

Choosing between real-time and batch processing requires analyzing your specific requirements across multiple dimensions:

Use Real-Time Processing When:

  • Intervention is Valuable: If blocking or delaying suspicious transactions provides significant value, real-time is essential. This includes fraud prevention, sanctions compliance, and high-risk customer monitoring
  • Speed Matters Operationally: When analysts need to investigate and resolve alerts while customers are still engaged (e.g., at a branch or on the phone), real-time processing enables better customer service
  • Regulatory Requirements: Some jurisdictions or scenarios mandate real-time checks, particularly for sanctions screening and terrorist financing prevention
  • Detection Logic is Simple: If your rules and models can execute in under 100ms, real-time processing is feasible without extensive optimization
  • Per-Transaction Patterns: When anomalies are detectable in individual transactions or very recent history (velocity, geolocation, device fingerprinting)

Use Batch Processing When:

  • Context Requires Complete Dataset: Detecting structuring, layering, or other complex schemes often requires seeing an entire day's or week's transactions to identify the pattern
  • Computationally Intensive: Deep learning models, graph neural networks, or comprehensive network analysis may require seconds or minutes per transaction—acceptable in batch but impossible in real-time
  • Retroactive Analysis: Regulatory reporting, trend analysis, and model training naturally fit batch processing since they analyze historical data
  • Cost Constraints: When infrastructure budget is limited, batch processing provides more detection capability per dollar spent
  • Intervention Isn't Needed: For many AML scenarios, detection delay of 24 hours is acceptable because you're documenting patterns for SAR filing, not preventing transactions

Architecture Pattern

A common successful pattern is to start with batch processing for comprehensive coverage, then progressively move specific high-value scenarios to real-time as you identify them.

  • Begin with batch processing for all scenarios
  • Identify high-risk patterns where speed matters
  • Implement real-time processing for those specific scenarios
  • Maintain batch processing for comprehensive backstop coverage

Technical Implementation Considerations

Successfully implementing either architecture requires careful attention to technical details that significantly impact performance and reliability:

Real-Time Implementation Challenges

Building production-grade real-time AML monitoring involves solving several technical challenges that don't exist in batch systems:

  • State Management: Streaming systems must maintain state (account balances, transaction counts, historical patterns) across distributed workers. This requires careful use of state stores, windowing, and watermarks
  • Exactly-Once Processing: Avoiding duplicate alerts when systems fail and restart requires idempotent processing and careful transaction handling
  • Late-Arriving Data: Transactions may arrive out of order or late. Your windowing strategy must handle this gracefully without creating gaps in detection
  • Backpressure Handling: When downstream systems slow down, the streaming pipeline must handle backpressure without dropping transactions or creating unbounded memory growth
  • Model Deployment: Updating ML models in a running streaming system without downtime or duplicated processing requires sophisticated deployment strategies

Batch Implementation Challenges

While conceptually simpler, batch systems have their own technical complexities:

  • Batch Window Management: As data volumes grow, completing processing within available windows becomes challenging. Strategies include incremental processing, data partitioning, and progressive optimization
  • Dependency Management: Batch jobs often depend on multiple upstream data sources. Orchestrating these dependencies and handling failures gracefully requires tools like Airflow or Dagster
  • Reprocessing: When you discover issues in historical processing, reprocessing large date ranges efficiently requires careful architecture for incremental updates
  • Data Freshness: Ensuring all required data has arrived before starting batch processing requires careful coordination, especially across time zones

Performance and Cost Analysis

The performance and cost characteristics of real-time versus batch processing differ significantly, and the optimal choice depends heavily on your transaction volume and patterns:

3-5x
infrastructure cost increase
real-time vs batch (typical)
99.5%
cost on compute
for batch processing
60%
cost on streaming infra
for real-time systems

Cost Breakdown: Batch Processing

For a financial institution processing 10 million transactions per day with batch AML monitoring:

  • Compute: $3,000/month (spin up large cluster for 4 hours nightly)
  • Storage: $800/month (90 days transaction history, 1 year alert history)
  • Orchestration: $200/month (Airflow managed service)
  • Total: ~$4,000/month or $0.012 per 1,000 transactions

Cost Breakdown: Real-Time Processing

For the same institution with real-time monitoring:

  • Streaming Infrastructure: $8,000/month (Kafka or Kinesis at scale)
  • Stream Processing: $5,000/month (Flink cluster, always-on)
  • State Store: $2,500/month (Redis or DynamoDB for real-time state)
  • Model Serving: $3,500/month (low-latency inference infrastructure)
  • Total: ~$19,000/month or $0.057 per 1,000 transactions

The 4-5x cost difference is typical, but the value proposition depends entirely on what that real-time capability enables. If blocking $10 million in fraudulent transactions per month, the ROI is obvious. If you're just moving alert generation from 24 hours to 5 minutes with no operational impact, it may not be justified.

Regulatory and Compliance Considerations

Different regulatory regimes have varying expectations around AML monitoring latency, which can influence your architectural choices:

Sanctions Screening

Most jurisdictions require real-time sanctions screening before transactions complete. This is typically non-negotiable and must be part of your inline processing. However, comprehensive sanctions screening can often be split:

  • Inline: Exact name matches and high-confidence fuzzy matches (< 100ms)
  • Near-real-time: Comprehensive fuzzy matching, phonetic algorithms (1-5 minutes)
  • Batch: Network-based sanctions detection, beneficial ownership analysis (daily)

Suspicious Activity Detection

For general AML monitoring, regulators typically expect 'timely' detection rather than real-time. What constitutes 'timely' varies by jurisdiction and risk profile:

  • High-risk customers: Near-real-time to daily monitoring expected
  • Medium-risk customers: Daily to weekly monitoring acceptable
  • Low-risk customers: Weekly to monthly monitoring often sufficient

This risk-based approach allows for a tiered architecture where high-risk scenarios get real-time treatment while lower-risk scenarios use batch processing, optimizing both compliance and cost.

Regulatory Perspective

Regulators care more about the effectiveness of your detection than the speed, with some exceptions. A batch system that detects 95% of money laundering is better than a real-time system that catches 60%.

  • Document your risk-based approach to monitoring frequency
  • Demonstrate that detection latency is appropriate for each scenario
  • Show that your architecture can scale with transaction growth
  • Prove you can investigate and file SARs within required timeframes

Migration Strategies

Many institutions are considering migrating from batch to real-time processing, or implementing hybrid architectures. Based on our experience helping dozens of financial institutions with this transition, here's a proven migration path:

Phase 1: Establish Streaming Infrastructure (3-6 months)

  • Deploy streaming platform (Kafka, Kinesis, Pulsar)
  • Implement transaction ingestion into streaming infrastructure
  • Build monitoring, alerting, and operational runbooks
  • Train operations team on streaming operations
  • Run in parallel with existing batch system (don't switch yet)

Phase 2: Shadow Mode Implementation (2-4 months)

  • Implement selected scenarios in real-time processing
  • Generate alerts but don't send to analysts (shadow mode)
  • Compare real-time alerts with batch alerts for same scenarios
  • Tune to achieve parity or better with batch performance
  • Validate latency, accuracy, and operational characteristics

Phase 3: Gradual Cutover (3-6 months)

  • Move first scenario to production real-time processing
  • Monitor carefully for issues, maintaining batch as backup
  • Progressively migrate additional scenarios
  • Keep batch processing running for comprehensive coverage
  • Document learnings and refine approach iteratively

Phase 4: Optimization and Enhancement (Ongoing)

  • Optimize performance and reduce costs
  • Add sophisticated real-time detection scenarios
  • Implement intervention capabilities
  • Enhance with real-time ML model updates
  • Continue batch processing for complex analytics

Case Study: Large European Bank Migration

A major European bank with 15 million customers and 50 million monthly transactions successfully migrated from pure batch to hybrid architecture over 18 months:

  • Starting Point: Daily batch processing with 24-36 hour alert latency, 3.2% alert rate, 8% true positive rate
  • Phase 1: Implemented real-time sanctions screening and velocity checks (6 months)
  • Phase 2: Added near-real-time ML models for account takeover and mule detection (6 months)
  • Phase 3: Maintained batch processing for complex pattern detection, enhanced with graph neural networks (6 months)
  • Results: Average alert latency reduced to 4 hours, alert rate decreased to 1.8%, true positive rate improved to 15%, blocked €45 million in fraud during first year

The hybrid approach proved optimal: real-time processing for scenarios where intervention has value, batch processing for comprehensive pattern detection that requires full context.

Future Trends: Convergence and Evolution

The future of AML monitoring architectures is increasingly sophisticated, with several trends emerging:

  • Micro-Batching: Processing small batches every few minutes combines benefits of both approaches—near-real-time latency with batch-style complete context
  • Adaptive Processing: Systems that dynamically choose processing strategy based on transaction risk score, using real-time for high-risk and batch for low-risk
  • Continuous Learning: ML models that update incrementally from streaming data rather than requiring batch retraining, enabling faster adaptation to new patterns
  • Unified Streaming/Batch: Frameworks like Apache Beam that allow writing processing logic once and deploying to either streaming or batch engines
  • Edge Processing: Moving some detection logic to the point of transaction origination for ultra-low latency intervention

Conclusion: It's Not Either/Or

The real-time versus batch question is not binary. The most effective AML systems use both approaches strategically, applying real-time processing where its capabilities provide clear value and batch processing where comprehensive analysis requires complete context.

Your decision should be driven by a clear-eyed assessment of your requirements: Where does detection speed matter? Where is intervention valuable? What are your computational and cost constraints? What does your regulatory environment require? The answer will be different for every institution and every scenario within that institution.

At nerous.ai, our platform supports both real-time and batch processing natively, allowing institutions to choose the right approach for each scenario without being locked into a single architectural pattern. This flexibility is crucial as money laundering schemes evolve and institutional requirements change over time.

The best architecture is not the most advanced or the most expensive—it's the one that best aligns with your specific operational, regulatory, and business requirements while remaining sustainable and scalable as your institution grows.

👨‍💻

Alex Rodriguez

Chief Technology Officer at nerous.ai

Alex leads nerous.ai's technology strategy and platform architecture. With 20 years of experience building large-scale data processing systems at companies like LinkedIn and Stripe, he specializes in architecting AML systems that balance detection effectiveness with operational practicality. He holds a PhD in Distributed Systems from MIT.

Explore Flexible AML Architecture

See how nerous.ai supports both real-time and batch processing in a unified platform. Schedule a demo to discuss the right architecture for your institution's specific requirements.

Schedule Demo →