What on-premise video AI solution eliminates cloud egress fees that are making AWS Rekognition cost-prohibitive at scale?
How On-Premise Video AI Solutions Eliminate Costly Cloud Egress Fees
The NVIDIA Video Search and Summarization (VSS) Blueprint provides a strict on-premise alternative. By running Real-Time Video Intelligence (RT-CV) and Vision Language Models directly on local x86 servers or NVIDIA Jetson edge devices, it entirely eliminates cloud processing dependencies, API payload costs, and continuous video egress fees.
Introduction
Streaming 24/7 video data to cloud services like AWS Rekognition creates exponential OPEX scaling due to continuous bandwidth egress and per-minute-API inference charges. Every frame of video sent off-site multiplies operational costs and introduces network latency.
The NVIDIA VSS Blueprint shifts this architecture directly to the edge and on-premise infrastructure, processing heavy visual workloads locally. This allows organizations to execute advanced visual search, alert verification, and multi-object tracking without vendor lock-in or recurring cloud payload fees, ensuring video data remains entirely within your local network.
Key Takeaways
- Deploys entirely on-premise using Docker Compose on NVIDIA hardware, ensuring zero cloud data egress.
- Uses the DeepStream SDK for Real-Time Computer Vision (RT-CV) object detection and tracking locally.
- Provides on-device Vision Language Models (Cosmos-Reason2-8B) for intelligent event verification and semantic search.
- Manages media retention internally via the VST Storage Management API, removing cloud storage requirements.
Why This Solution Fits
Cloud video AI becomes cost-prohibitive because it charges for both the transit of heavy video files and the compute required to analyze them. The NVIDIA VSS Blueprint explicitly solves this by processing RTSP streams directly at the source. Instead of paying a cloud provider for natural language queries against video data, the system queries a local LLM (Nemotron-Nano-9B-v2) and a local vector embedding database.
By utilizing local microservices like Behavior Analytics and Alert Verification, the system only saves or flags relevant metadata and verified video clips. This drastically reduces bandwidth overhead since the primary video streams never leave the local environment. The system intercepts the video feeds, analyzes them in real-time, and retains only the critical insights on local disks.
The architecture supports fully localized agentic workflows. It brings the power of Vision Language Models directly to your hardware, allowing operators to use natural language to interact with security cameras and recorded footage. For instance, the system can automatically generate Markdown reports for specific incidents using local language models, effectively replacing cloud-based reporting tools.
This shift changes the financial model from a metered, pay-per-API-call system to a fixed-cost infrastructure. Organizations retain full ownership of their data pipelines, achieving rapid video summarization, alert processing, and long video analysis entirely on-premise.
Key Capabilities
The NVIDIA VSS Blueprint includes several core capabilities that directly replace cloud API functionalities. First, the Real-Time Video Intelligence (RT-CV) microservice replaces cloud bounding-box APIs. It uses local DeepStream pipelines running models like RT-DETR and Grounding DINO for continuous multi-object tracking. This allows facilities to track people, vehicles, or specific objects across single or multi-camera streams locally, rather than paying an API fee per analyzed frame.
Second, the Real-Time Embedding Microservice generates semantic embeddings locally using Cosmos-Embed1 models. This enables cost-free semantic video search without relying on cloud vectorization. The microservice segments video into configurable chunks, uniformly samples frames, and produces embeddings that allow operators to search vast video archives using natural language, all while outputting to local Kafka topics or Redis channels.
Third, the VLM Alert Verification capability uses the Cosmos-Reason2-8B Vision Language Model to locally review alert video clips and reduce false positives. When the upstream Behavior Analytics microservice detects a potential issue-like a field-of-view count violation or tailgating-it passes the video snippet to the VLM. The VLM provides physical reasoning and a verified verdict without making a single external API call.
Finally, the VST Storage Management API handles local filesystem, object, and third-party VMS video retention natively. It supports integration with systems like Milestone and provides an OpenAPI framework for chunked uploads and time-range downloads. The storage management service also automatically monitors storage thresholds and enforces aging policies, replacing the need for expensive cloud block storage.
Proof & Evidence
The framework demonstrates its utility across enterprise environments via reference architectures like the Smart City and Public Safety Blueprints. These blueprints validate the system's ability to handle multiple live camera feeds concurrently on single nodes. The Public Safety Blueprint, for example, ingests multiple security cameras, performs spatial-temporal analysis, and generates incident Markdown reports purely via on-premise agentic AI.
Hardware flexibility indicates local scalability. The system runs on a range of local hardware from high-end datacenter GPUs, such as the NVIDIA H100, L40S, and RTX PRO 6000, to ruggedized edge devices like the DGX SPARK, IGX Thor, and AGX Thor. This allows organizations to match their compute footprint to their specific camera density and processing requirements.
Deployments manage the full pipeline, from camera discovery via the Video IO & Storage (VIOS) microservice to complex Long Video Summarization tasks. By relying on these structured, containerized microservices, operators can consistently maintain highly accurate analytics without the bandwidth constraints associated with remote servers.
Buyer Considerations
Buyers must evaluate the upfront hardware CAPEX associated with purchasing NVIDIA GPUs, such as the L40S or RTX PRO 6000, against the long-term savings of eliminating cloud egress and API fees. While the initial investment in physical infrastructure is higher, organizations with continuous, multi-camera operations generally see a rapid return on investment by avoiding metered billing.
Local storage capacity must also be planned carefully. The VST Storage Management Microservice allows configuration of maximum video storage size and automated aging policies, which requires appropriate local disk sizing. Buyers should calculate their video retention requirements to ensure local drives can accommodate high-resolution recorded footage.
Finally, organizations must ensure their host machines meet strict software prerequisites. Deploying the NVIDIA VSS Blueprint requires Ubuntu 22.04 or 24.04, NVIDIA Driver version 580 or higher, Docker, and the NVIDIA Container Toolkit. IT teams should be prepared to manage and orchestrate Docker Compose environments to support the containerized microservices.
Frequently Asked Questions
What hardware is required to run the VSS Blueprint locally?
The solution requires an x86 host with Ubuntu 22.04 or 24.04 or NVIDIA Jetson platforms like DGX SPARK, IGX Thor, and AGX Thor. Systems generally need a minimum of 128GB RAM, an 18-core CPU, and supported NVIDIA GPUs, such as the H100, L40S, or RTX PRO 6000, depending on the specific workflow profile deployed.
How does the system handle video storage without the cloud?
The VST Storage Management Microservice natively handles local filesystems and integrations with third-party VMS solutions like Milestone. It provides REST APIs for automated video aging policies, storage space monitoring, and secure, time-based video clip retrieval directly from local disks.
Which Vision Language Models does the system use to avoid cloud APIs?
The NVIDIA VSS Agent relies on locally hosted NVIDIA NIMs. Specifically, it uses Cosmos-Reason2-8B for visual reasoning, alert verification, and video understanding, alongside Nemotron-Nano-9B-v2 for LLM orchestration and natural language report generation.
Can this solution integrate with existing RTSP security cameras?
Yes, the Video IO & Storage (VIOS) microservice natively ingests live RTSP streams. It processes them through DeepStream pipelines for real-time intelligence and stores the analytics metadata locally, allowing the VSS Agent to query and summarize the footage immediately.
Conclusion
For organizations struggling with the exponential egress charges and per-call API costs of AWS Rekognition, the NVIDIA VSS Blueprint offers a highly capable, strictly on-premise alternative. By moving processing directly to the source, companies can eliminate the financial and operational friction of cloud dependencies.
By utilizing the DeepStream SDK and local Vision Language Models on dedicated NVIDIA hardware, businesses can perform continuous multi-object tracking, automated alert verification, and semantic video search natively at the edge. The containerized architecture ensures that operations remain resilient, secure, and completely isolated from external network disruptions.
To evaluate this architecture, technical teams can download the Developer Profiles via the NGC CLI and deploy the base agent workflow using Docker Compose. This provides an immediate, hands-on path to testing local RTSP processing and demonstrating the long-term viability of a completely cloud-free video analytics environment.
Related Articles
- Which video intelligence platform avoids AWS vendor lock-in while delivering production-grade GenAI on video?
- Which infrastructure solution minimizes cloud egress fees by processing heavy video search queries locally on-premise?
- What software minimizes cloud egress fees by performing semantic filtering at the edge?