International Journal For Multidisciplinary Research

E-ISSN: 2582-2160     Impact Factor: 9.24

A Widely Indexed Open Access Peer Reviewed Multidisciplinary Bi-monthly Scholarly International Journal

Call for Paper Volume 8, Issue 3 (May-June 2026) Submit your research before last 3 days of June to publish your research paper in the issue of May-June.

Real-Time Regulatory Reporting in Banking: From 24-Hour Batch to Sub-Two-Minute Streaming with Auditable Lineage

Author(s) Jeevan Krishna Paruchuri
Country United States
Abstract Regulatory reporting in retail and commercial banking has historically been built on overnight batch ETL: extract from the core banking system after end-of-day, transform overnight, submit before the morning regulatory deadline. This pattern is operationally simple and audit-friendly, and it imposes a hard floor of approximately 24 hours between when a transaction occurs and when a regulator could in principle see it. As regulatory regimes Basel IV's granularity expectations, DORA's operational resilience requirements, MiFID II's trade reporting timelines push toward shorter and more frequent reporting windows, the 24-hour floor becomes a problem. This paper presents a case study of migrating a banking regulatory reporting pipeline from a legacy overnight batch architecture to a streaming architecture with end-to-end latency under 2 minutes, while preserving and in fact strengthening the audit and lineage properties that make the legacy system defensible to regulators. The architecture is built on Attunity Replicate for mainframe CDC-based change data capture from the Oracle core banking system into per-table Kafka topics governed by a schema registry, Spark Structured Streaming with exactly-once semantics writing into Delta Lake with append-only audit semantics, and dual-region replication that drove disaster recovery time from 6 hours 20 minutes to 3 hours 45 minutes. Inter-region Kafka replication lag is under 1 minute in steady state. The pipeline handles a baseline of approximately 2 million transactions per day (~23 TPS baseline, ~230 TPS peak) with capacity for 10x spikes at month-end and during crisis scenarios. We document the regulatory requirements that drove the architecture, the technical patterns that emerged, the audit capabilities the new system provides point-in-time queries, full change history, immutable amendment records and the things that took multiple iterations to get right, most notably exactly-once semantics under Kafka offset management and schema evolution coordination across producer and consumer teams. We are honest about the operational complexity the streaming architecture introduces and about the months of work required to build compliance-team trust through hands-on demonstrations of the audit trail. The contribution is a practitioner-grounded blueprint for teams considering whether the cost of streaming regulatory reporting is justified by the regulatory and operational benefits, with explicit attention to the audit and lineage properties that compliance teams will demand.
Field Computer Applications
Published In Volume 6, Issue 5, September-October 2024
Published On 2024-10-10
DOI https://doi.org/10.36948/ijfmr.2024.v06i05.75350

Share this