Northeastern University, Boston, MA.
International Journal of Science and Research Archive, 2025, 16(03), 1263-1278
Article DOI: 10.30574/ijsra.2025.16.3.2668
Received on 16 August 2025; revised on 24 September 2025; accepted on 27 September 2025
Large language model (LLM)-powered AI has created a new paradigm for the operationalization and deployment of AI systems, resulting in a new field of study, LLMOps. This study provides empirical metrics of a simulated production scale deployment, with a P95 inference latency of 470ms, and a token throughput of 28,000 tokens per second from an LLM with a hallucination rate of 4.2 percent, demonstrating the viability of LLM in real piping. This study examines how the integration of LLMOps with real-time streaming data and complex event processing (CEP) pipelines can combine NLP and high-velocity event-driven architecture. In addition to helping expand the prospective value of LLM in event driven systems, this project will explore intelligent text understanding, real-time decision making based on the text understanding, and reflexive engagement to advance discussion on LLM through a review of MLOps into LLMOps. Some of the preliminary ideas, including AgentOps, observability, and real-time optimization, will be presented and/or discovered in the framework of data streams. Practical deployment opportunities in various verticals, such as finance, human health, and cyber security, and latency, scale, and possible observability attack vectors will also be considered by the review. The more adaptive, explainable and scalable solutions of current event based systems will be deployed in the future work through the use of LLMOps.
LLMOps; Streaming Data; Natural Language Processing (NLP); Complex Event Processing (CEP)
Preview Article PDF
Devarsh Hemantbhai Patel. LLMOps for Streaming Data: Bridging NLP and Event Pipelines. International Journal of Science and Research Archive, 2025, 16(03), 1263-1278. Article DOI: https://doi.org/10.30574/ijsra.2025.16.3.2668.
Copyright © 2025 Author(s) retain the copyright of this article. This article is published under the terms of the Creative Commons Attribution Liscense 4.0







