AI video analytics security system with real-time object detection, facial recognition, and alerts

A security camera that just records is a recording device. A security camera that understands what it’s seeing — that flags a fight before the fists land, spots the forklift drifting into the pedestrian lane, recognises the licence plate of a vehicle that’s been circling the lot for an hour — is a security system. Closing that gap is what AI video analytics does, and in 2026 it’s no longer a moonshot. It’s a procurement line item that boards expect to see.

This guide is the playbook we wish someone had handed us when we shipped our first AI video security product nine years ago. It walks through what AI video analytics for security actually is in 2026, the eight use cases where it pays for itself, the seven-layer reference architecture we deploy in production, the seven smart-security platforms worth comparing, the compliance landscape (NDAA, EU AI Act, GDPR, BIPA), a real cost model for a 200-camera deployment, and the pitfalls that quietly kill these projects. It’s long. It’s practical. We wrote it for security directors, IT/OT leaders, and product owners who need to make a call this quarter.

Key takeaways

  • The global video analytics market is on track for roughly $6.8B in 2026, with a 22–30% CAGR pushing it past $49B by 2035.
  • Modern detectors (YOLOv9-E, YOLOv10) push 56% mAP on COCO and run 1.8× faster than RT-DETR — real-time analytics on a $200 Jetson are now table stakes.
  • Retail shrinkage hit $112.1B in 2025; AI-powered loss-prevention case studies show 30–83% reductions in theft incidents within twelve months.
  • EU AI Act prohibitions are in force since Feb 2025; high-risk obligations land August 2026. NDAA Section 889 already bans Hikvision/Dahua from federal contracts.
  • Edge inference cuts bandwidth 80–95% and is now the default for any deployment over 50 cameras — the cloud is for re-search, federated learning, and audit trails.

01. Why Fora Soft wrote this guide

We’ve been building AI video software since 2017 and shipping production video systems since 2005. That gives us an unusual vantage point on smart security: we’ve watched it go from forensic search (replay yesterday’s footage to find the suspect) to live decisioning (alert the on-shift guard before the suspect reaches the door). Most of what we know we learned shipping things that broke in production and then fixing them on a Saturday at 2 a.m.

A few of the products that informed this guide:

  • MindBox — an incident-detection platform processing 500,000+ vehicles a day across 80+ camera locations, with 99.5% facial recognition accuracy and sub-second alerts.
  • V.A.L.T. — a multi-room video-recording platform deployed in 770+ organisations with 2,500+ cameras, used for security, observation, and compliance recording.
  • Industrial PPE detection systems — hard-hat, vest, and exclusion-zone detectors deployed at construction and energy sites.
  • Retail loss-prevention pilots — sweethearting, ORC pattern recognition, and self-checkout monitoring.

Two notes on what makes this guide different from the marketing PDFs you’ll find elsewhere. First, we have skin in this game — everything you’ll read here we either operate ourselves or have built for paying customers. Second, our Agent Engineering practice means we typically deliver these projects 30–50% faster than a traditional shop — we use AI to write boilerplate, generate test fixtures, draft ONVIF integrations, and so on, which means our estimates often look surprisingly low. They’re not. They’re just current.

Skip the reading? We do free 30-minute scoping calls for security analytics projects — a CTO will walk through your camera fleet, your VMS, and your compliance constraints, and tell you what’s feasible. Book a call →

02. What AI video analytics for security actually is

AI video analytics is the layer that converts raw video frames into structured events — “person crossed line at 14:32:08, confidence 0.94” — so that downstream systems (a VMS, an access-control panel, a security operations centre, a phone app) can act on them. For security workloads it spans seven analytic primitives: object detection, multi-object tracking, person re-identification, action/behaviour classification, facial recognition, licence-plate recognition, and anomaly detection. Each is a model. Each model has a latency, an accuracy, a hardware budget, and a compliance footprint.

It is not: motion detection (that’s a 1990s technology that fires on swaying trees), generic ChatGPT-on-video (the latency and cost don’t add up for live monitoring), or “smart cameras” with proprietary firmware that locks you into one vendor’s ecosystem (you’ll regret it in year three).

A useful mental model is that AI security analytics adds three new layers above the camera: perception (what’s in the frame), understanding (what’s happening across frames), and action (what to do about it). Most off-the-shelf platforms ship perception well, ship understanding partially, and leave action almost entirely to you. That last layer is where most projects either succeed or quietly fail.

03. Market snapshot: where the smart-security spend is heading in 2026

The numbers are mostly converging now. Different analyst houses still publish different absolute totals (the global “video analytics” segment runs anywhere from $5B to $11B depending on whether facial recognition, ALPR, and traffic analytics are bundled in), but the shape of the market is consistent: a 22–30% compound annual growth rate through the back half of the decade, with the lion’s share of new spend going to AI-driven analytics rather than traditional VMS or storage.

A few data points worth holding onto:

  • Global video analytics market: roughly $6.83B in 2026, projected to exceed $49B by 2035.
  • Edge AI camera shipments: estimated to surpass 120 million units annually by 2027 as Jetson Orin and Hailo-8 drop below the $80 BOM threshold.
  • Retail shrinkage in the United States: $112.1 billion in 2025, with external theft and ORC accounting for roughly 36% of losses.
  • Average AI-pilot conversion to production: still only ~30% in physical security, mostly stalling on integration with the legacy VMS and on alert fatigue.
  • NDAA Section 889 enforcement: federal contracts can no longer use Hikvision, Dahua, Hytera, Huawei, or ZTE equipment — and the rule is propagating to state and enterprise procurement.

If you’re planning a 2026 budget, the practical takeaway is: hardware is cheaper than ever, models are commoditising, and the differentiator is integration. The cost has shifted from “buy the analytics” to “wire the analytics into the workflows people actually run all day.”

04. Eight use cases where AI security analytics earns its keep

We’ll spare you the “the possibilities are endless” framing. In our project intake, eight use cases account for roughly 90% of what gets shipped. They’re ordered by maturity: the higher up the list, the more likely you are to find off-the-shelf models that perform well out of the box.

1. Retail loss prevention & ORC detection

Self-checkout sweethearting, organised retail crime (ORC) pattern recognition, after-hours intrusion. AI looks for skip-scans (item moved past the scanner without a beep), bagging anomalies (item placed in bag without scanning), and known-offender re-entry. Reported reductions in shrinkage from production deployments: 30–83% within twelve months. See our deep dive on retail video analytics.

2. Public safety & smart cities

ALPR for stolen-vehicle alerts, crowd density estimation, weapon detection in public squares, fight detection in transit hubs. The legal envelope is tight (especially in the EU under the AI Act’s “real-time biometric identification in public spaces” clause), so most production systems run with consent overlays or are limited to investigative replay rather than live alerting.

3. Transportation & parking

Wrong-way detection on highways, tailgating in tolls, parked-vehicle classification, abandoned-luggage detection in airports. Ingest is usually via PTZ feeds and ANPR cameras at lane level. MindBox handles 500k+ vehicles a day across this segment.

4. Industrial PPE & safety compliance

Hard hat, high-vis vest, safety glasses, and harness detection on construction sites, refineries, and warehouses. Plus exclusion-zone monitoring (no person in the swing radius of an excavator) and forklift-pedestrian proximity alerts. We unpack this in our guide to hard-hat detection.

5. Healthcare campuses

Patient-fall detection, elopement monitoring (a wandering patient leaves a unit), aggression escalation in EDs. HIPAA forces strict on-prem inference and tight access logs — cloud-only solutions are usually a non-starter.

6. Schools & universities

Weapon detection at perimeters and entrances, lockdown automation, after-hours intrusion. The market here is highly sensitive to false positives — a single false weapon alert that triggers a SWAT response is a career-ending event for an administrator. Vendors typically gate alerts behind a human reviewer in a 24/7 SOC.

7. Government & critical infrastructure

Substation perimeters, water treatment plants, ports. NDAA-compliant cameras only (Axis, Hanwha, Bosch, i-PRO, Verkada). Most deployments are air-gapped from the internet and use one-way data diodes for reporting up to a SIEM.

8. Construction & site progress

Material-delivery counting, equipment idle-time analysis, after-hours theft, perimeter intrusion. Often deployed as a temporary mast-mounted camera with cellular backhaul; analytics run on a small Jetson box at the base of the mast.

Build vs buy decision

Stuck between a turn-key VMS and a custom platform? We’ve done both, dozens of times. A 30-minute call with our CTO will tell you which side of the line you’re actually on.

Book a 30-min scoping call →

05. Reference architecture: the seven-layer security analytics pipeline

Every smart-security system we’ve shipped looks roughly the same under the hood. The seven layers below are the ones we end up building or integrating in nearly every project. Skip a layer at your peril — the one you skip is usually the one that bites you in production.

Layer 1: Camera & ingest

ONVIF Profile S/T/G compliant cameras feeding RTSP (H.264/H.265) into a media server. For greenfield projects we recommend NDAA-compliant brands (Axis, Hanwha, Bosch, i-PRO) at minimum 4MP, 30fps, with WDR. Avoid embedded analytics on the camera itself unless the use case is binary (motion / no motion) — you want the model on a box you control.

Layer 2: Edge inference node

A small Jetson Orin (Nano/NX/AGX) or Intel + Hailo-8 box at the closet level, handling 8–32 camera streams. Runs YOLOv9/v10 for detection, ByteTrack/BoT-SORT for tracking, and a quantised face/ALPR model where needed. Outputs structured events over MQTT or gRPC to the server tier.

Layer 3: Server tier

NVIDIA Triton + TensorRT for any models too heavy for the edge (re-identification across cameras, complex activity recognition). This tier also runs the rules engine that combines events into actionable alerts — “person + loitering > 60s + after hours = alert.”

Layer 4: Data & index

PostgreSQL/TimescaleDB for events, S3-compatible object storage (MinIO or AWS S3) for clips, and a vector DB (Qdrant or Weaviate) for similarity search — “find every clip with a person wearing a red jacket between 6 and 8 p.m.”

Layer 5: VMS integration

Pushing detections back into Milestone XProtect, Genetec Security Center, Avigilon Control Center, or Hanwha Wisenet via their respective SDKs. This is where most projects underestimate the work — vendor SDKs are often poorly documented and require significant glue code.

Layer 6: Alert & workflow

A SOC dashboard, mobile app, and integrations with PagerDuty/Opsgenie, two-way radio dispatch, and access control panels (HID Origo, LenelS2, Genetec Synergis). The user interface here is what your customer actually sees every day — budget for it accordingly.

Layer 7: Audit & governance

Tamper-evident audit logs for every detection, every override, every clip access. RBAC and SSO at the operator level. Retention policies that map to your local privacy law — usually 14–90 days for raw video, longer for tagged events.

Where teams cut corners and regret it: Layer 6 (alerts) and Layer 7 (audit). Skipping audit logs makes a GDPR or CJIS audit a paperwork emergency; skipping alert workflow design means the SOC drowns in noise within two weeks.

06. Comparison matrix: seven smart-security platforms benchmarked

If you’re evaluating off-the-shelf, the practical shortlist in 2026 looks like this. Pricing is indicative — everyone discounts heavily on multi-year deals, so use these as anchor points rather than quotes.

Platform Best for Edge or cloud Indicative price Watch-out
VerkadaMid-market enterprise; SMB unified physical securityEdge-first, cloud command$500–3,000/cam + $199–1,799/cam/yrLocked-in hardware; export is hard
Avigilon (Motorola)Large enterprise; appearance searchHybrid (server + edge)Quote-based; ~$700–1,500/cam + ACC licensePer-camera license adds up fast
Genetec Security CenterGovernment, transport, large campusesOn-prem with cloud add-onsQuote-based; channel-onlySteep learning curve; integration heavy
RhombusMulti-site SMB; cloud-nativeEdge + cloud~$700–1,400/cam + $200–500/cam/yrLimited integrations vs. Verkada
Eagle Eye NetworksCloud VMS over existing camerasCloud-first via on-site bridge$15–50/cam/mo + bridge hardwareBandwidth-hungry; latency varies
Spot AIAdds AI to existing camera fleetOn-prem appliance + cloud~$50–100/cam/mo all-inNewer ecosystem; fewer system integrations
Custom (Fora Soft & similar)Anything that doesn’t fit a SKU; product platformsYou decide$80k–$400k MVP, then per-camera marginalNeed a real engineering partner

Pricing reflects publicly observed ranges as of Q1 2026 and routinely shifts with discounts, multi-year terms, and bundle deals. Treat the column as a directional anchor, not a quote.

Procurement tip: ask any vendor to commit, in writing, to the published per-camera price for years 2–5. Multi-year price escalators of 8–15% per year are common and rarely surfaced during the initial pitch.

07. Edge vs cloud: where to put the inference

For any deployment over about 50 cameras, edge inference is now the default. The reasons are unsexy and decisive: bandwidth, latency, and cost. We covered the trade-offs in detail in how to integrate analytics with your existing VMS, but here’s the short version.

Dimension Edge (Jetson / Hailo) Cloud (T4 / L4 / A10G)
Detection latency30–80 ms200–800 ms (round-trip)
Bandwidth per camera~10–50 kbps (events only)2–8 Mbps (full stream)
Failure mode if WAN diesKeeps detecting locallyGoes blind
CapEx per 8-camera node$1.5–3k~$0 upfront
OpEx per camera per month~$2–6 (power + mgmt)~$15–40 (compute + bandwidth)
Privacy postureFrames never leave siteFrames cross network boundary
Best forLive alerting, >50 cams, regulatedForensic re-search, model retraining, <20 cams

Our standing recommendation is edge for live, cloud for forensic and training. Compare with our broader piece on AI-powered video surveillance for the architectural framing.

08. The model layer: what to actually run on your cameras

Models drift, but in early 2026 the practical shortlist for production security workloads is small. Here’s what we deploy when:

  • Detection: YOLOv9-E (56% mAP on COCO) for accuracy-first deployments; YOLOv10-S (43.8% mAP, 1.8× faster than RT-DETR) for edge-first. RT-DETR for transformer-based use cases. We dig into this in our AI video surveillance overview.
  • Tracking: ByteTrack and BoT-SORT for general use; DeepSORT only when you need re-ID baked in.
  • Re-identification: OSNet or CLIP-ReID for cross-camera tracking. Vector DB (Qdrant/Weaviate) for the gallery.
  • Action recognition: SlowFast or VideoMAE for fight/fall/loitering. These are heavier — usually run on the server tier, not the edge.
  • Anomaly detection: See our deeper write-up on anomaly detection models. Memory-augmented autoencoders for unsupervised scenes.
  • Face recognition: ArcFace / AdaFace embeddings, FAISS or Qdrant index. Performance ceiling is usually photographic quality, not the model.
  • ALPR: A specialised pipeline (detector + plate-reader). Open-source: PaddleOCR + custom detector. Commercial: Plate Recognizer, Genetec AutoVu.

09. VMS integration: making analytics play nice with the existing stack

In greenfield projects you can pick your VMS. In nearly every brownfield project — which is most of them — you’re bolting AI onto whatever’s already there. The big four still cover most of the installed base: Milestone XProtect, Genetec Security Center, Avigilon Control Center, and Hanwha Wisenet WAVE. Each exposes an SDK or REST API for sending events back; each has quirks; each requires a paid integration partnership in some scenarios.

Common integration patterns we use: writing detected events as bookmarks (Milestone), as alarms (Genetec), as appearance-search vectors (Avigilon), or as overlay metadata (Hanwha). The choice changes how the operator interacts with the alert — bookmarks are great for forensic review, alarms force an acknowledgement, and overlays drop in the live wall view.

Budget rule of thumb: VMS integration is usually 20–35% of the project by hours. Underestimating this is the most common reason these projects slip.

10. The alert pipeline: turning detections into action

A detection is not an alert. A detection is a signal that, combined with context, a rules engine, and a human operator’s attention budget, may produce an alert. Get this layer wrong and your customer turns the system off within a month because the SOC is drowning in noise.

A few patterns we’ve found work in production:

  • Compose, don’t chain. A “person” detection alone is meaningless after hours; “person + restricted zone + after hours + dwell > 30s” is an alert. Use a small rules engine (Drools, Open Policy Agent, or homegrown).
  • Tier the responses. Critical → phone call to on-shift guard. Major → SOC dashboard alert + acknowledgement timer. Minor → logged for audit; reviewed in shift handover.
  • Make false-positive feedback one click. Operators must be able to mark a false positive in two seconds — that data is the gold for retraining.
  • Always show the clip. No alert ships without a 10-second clip preview. Operators do not trust headless alerts and they shouldn’t.

11. Compliance: NDAA, EU AI Act, GDPR, HIPAA, BIPA

Compliance has shifted from afterthought to procurement gate. Here’s the 2026 lay of the land — bookmark it and consult an actual lawyer before launch, but this’ll keep you out of obvious traps.

Regime Where it applies Practical impact
NDAA Section 889US federal contracts, propagating to state & enterpriseNo Hikvision, Dahua, Hytera, Huawei, ZTE; component-level matters
EU AI ActAnything sold or used in EUReal-time biometric ID in public = prohibited (since Feb 2025); high-risk obligations from Aug 2026
GDPREU residents’ data, anywhere in worldDPIA required for video analytics; lawful basis usually legitimate interest + signage
HIPAAUS healthcarePHI in video = covered; on-prem inference, BAA, audit logs
BIPAIllinois (and similar TX/WA laws)Written consent for biometric collection — class actions are real
CJISUS law-enforcement dataPersonnel screening, encryption, on-prem-friendly
SOC 2 Type IIEnterprise customersIncreasingly table-stakes; budget 6–9 months to first report

For European deployments, our companion piece on 2026 ethics & AI video surveillance goes deeper into how teams are operationalising the AI Act’s high-risk requirements.

12. Mini case: how MindBox runs 99.5% face recognition on 500k vehicles a day

MindBox is one of our flagship security analytics deployments and a useful case for what production scale actually looks like. Five things make it work:

  1. Edge-first ingest. Each gate or zone has a Jetson AGX Orin running detection + tracking + ALPR locally. Server tier only sees structured events.
  2. Quality-gated face capture. Faces are only embedded when a face-quality classifier scores them above 0.85 — this is what gets the system to 99.5% on the cases that matter.
  3. Watchlist tiers. Three tiers (BOLO, person-of-interest, banned) with different alerting policies and audit requirements.
  4. Sub-second alert latency. From face crossing the line to ringing the on-duty officer’s phone: under 800 ms p95.
  5. Operator feedback loop. Every false positive is a one-click action; the model is retrained weekly on the corrected data.

The principle is portable. We use the same pattern in retail loss-prevention (the “face” becomes a sweethearting event), in industrial (it becomes a PPE violation), and in healthcare (an elopement). The business logic changes; the architectural pattern doesn’t.

13. Cost model: pricing a 200-camera deployment end-to-end

A 200-camera, multi-site deployment is a useful sizing point because it’s where SaaS and custom converge in price — under that, SaaS usually wins; over that, custom catches up fast. Below is a representative all-in for year one, split between off-the-shelf and a custom build with our team.

Line item Off-the-shelf (Verkada-class) Custom (Fora Soft-class)
Cameras (200 @ NDAA-compliant)$300k (bundled)$160k (Axis/Hanwha)
Edge nodes (25 @ 8 cams each)included$60k (Jetson AGX Orin)
Server tier (2 GPU nodes)cloud (in subscription)$30k
Software / buildincluded$140–220k (with Agent Engineering)
Year-one subscription$120–240k~$15k cloud OpEx
Install & commissioning$60–120k$60–100k
Year-one total$480–660k$465–625k
Year 2–5 OpEx (annual)$120–240k recurring$30–60k (hosting + support)

The custom build looks similar in year one and pulls dramatically ahead from year two as you stop paying per-camera SaaS fees. The catch: you (or your partner) own operations, security patching, and model maintenance. That’s a real responsibility. If you’re not staffed for it, off-the-shelf is the right call.

Get a tailored estimate

We’ll size your project against this model in 30 minutes — camera count, sites, VMS, compliance regime — and send you a one-page quote within two business days.

Book your scoping call →

14. Decision framework: pick your AI security approach in five questions

When customers come to us undecided, we run them through five questions. Honest answers usually point at exactly one of three options: SaaS platform, hybrid (existing cameras + AI overlay), or custom build.

  1. How many cameras across how many sites? Under 50 cams or 1–2 sites → SaaS is almost always cheaper. Over 200 cams or 10+ sites → custom catches up fast.
  2. What’s your existing camera fleet? Mostly Hikvision/Dahua and you sell to government? You have an NDAA problem you need to solve first. Mostly Axis/Hanwha? You can layer AI on top with Spot AI / custom.
  3. What’s the highest-risk use case? Weapon detection at a school? You need a vendor with a 24/7 human-in-the-loop SOC. PPE on a construction site? Off-the-shelf or custom both work.
  4. What’s your compliance regime? EU + biometrics → AI Act risks; US healthcare → HIPAA on-prem; multi-state retail with biometrics → BIPA exposure. These constrain vendor and architecture choices hard.
  5. Do you sell this as a product, or operate it internally? If you’re building a product (e.g. a VSaaS platform for the parking industry), you almost always want custom — off-the-shelf locks you out of differentiation.

For more on the build-product path, see our deep dive on custom video surveillance solutions.

Rule of thumb for pilots: insist on a four-week pilot on your own cameras, not a vendor demo reel. Any vendor unwilling to do this on live data has something to hide about how their models behave outside of scripted conditions.

15. Pitfalls to avoid — the six mistakes we see most often

  1. Treating the camera as the product. The camera is the worst place to spend on AI. Spend on the SOC dashboard, the alert logic, and the operator UX. Replace the camera every 5–7 years; replace the analytics layer every 18 months.
  2. Skipping the alert-tier design. The single fastest way to kill a deployment is to push every detection to the SOC dashboard. Tier the alerts before you ship.
  3. Using consumer-grade cameras. No, the Wyze cam in the storage closet is not a security camera. ONVIF compliance, WDR, and a maintained firmware roadmap are non-negotiable.
  4. No retraining loop. Models drift — new lighting, new uniforms, new vehicle types. If you don’t have a feedback loop from the SOC back into the training set, your accuracy quietly degrades over 6–12 months.
  5. Forgetting the audit log. When the regulator (or the lawsuit) shows up, the question is not “did the system detect it” but “can you prove who looked at the clip and when.” Build the audit trail from day one.
  6. Underestimating VMS integration hours. Budget for 20–35% of total project hours just for VMS work. Vendor SDKs are rarely as documented as their datasheets suggest.

16. KPIs: what to measure and the targets that matter

A short list of metrics we track on every deployment, with the targets that customers tend to settle on:

  • True positive rate (recall) per use case: >85% for safety-critical (weapons, falls); >75% for loss-prevention.
  • False positive rate per camera per day: <3 for SOC-monitored alerts.
  • End-to-end alert latency p95: <1.5 s for live alerts; <5 s acceptable for forensic.
  • System uptime: 99.9% for the alert pipeline; 99.5% for the analytics workers.
  • Time-to-clip: from alert ring to operator-watching-clip — <3 s.
  • Operator acknowledgement time: p50 < 30 s; p95 < 2 min.
  • Model drift: <5 percentage-point drop in mAP over 6 months.

17. Privacy by design: building trust with people on camera

Privacy-by-design is no longer a checkbox — it’s a procurement question and an investor question. Operationally, it means:

  • Default to embeddings, not images. When you ship to a vector DB or send to an alert, send a hash or embedding rather than the raw face crop wherever possible.
  • Pixelation pipelines for review. Operators looking at a clip should see pixelated bystanders by default; un-pixelating is a logged action.
  • Tight retention defaults. 14 days for raw video, 90 days for tagged events, 7 years for audit logs is a defensible baseline. Allow site-level override.
  • Subject-access requests as a UI. If you’re GDPR-bound, build an internal tool to honour SARs in <72 hours. Doing it ad hoc burns your DPO out fast.
  • Public signage that links to a policy. Belt-and-braces but expected.

18. When NOT to deploy AI security analytics yet

A short list of when our honest answer to a prospect is “wait six months, do these other things first.”

  • You don’t have a SOC or a clear escalation path. Detections without action are just noise.
  • Your existing camera footage is unusable (poor angles, low resolution, blocked lenses). Fix the optics before the AI.
  • You haven’t mapped the legal regime. Especially in the EU, getting a DPIA wrong is more expensive than the entire project.
  • Your SOC operators are at their limit on alert volume already. AI without alert design just multiplies the noise.
  • The use case can be solved by a $50 sensor (a door magnetic contact, a microwave motion sensor). AI is overkill there.

19. Hardening: what to wire in from day one

Smart-security platforms are themselves attractive targets — an attacker that owns the cameras owns the surveillance picture. A short list of what we put in on day one:

  • Mutual TLS between cameras, edge nodes, and server tier. No plaintext RTSP on the LAN.
  • Per-device certs (not shared keys) so revoking a stolen camera doesn’t mean re-keying the fleet.
  • SSO + RBAC for operators, with hardware MFA for admin actions.
  • Network segmentation: camera VLAN, analytics VLAN, ops VLAN. Firewalled.
  • Tamper-evident logs (append-only, hash-chained) so an attacker can’t cover their tracks.
  • Quarterly red-team against the SOC dashboard and the alert pipeline. Especially the “mute alert” button.
  • Camera firmware patching as a tracked operational task — not “we’ll get to it.”

20. What’s next: the three 2026–2027 shifts to plan for

  1. VLM-augmented review. Vision-language models (Gemini, GPT-4o-class) summarising hours of footage in plain language. Not for live alerting yet (latency, hallucinations) but a step-change for forensic review and shift handover.
  2. On-device fine-tuning. Federated learning across edge nodes so each site improves its own models without shipping pixels to the cloud. Keeps GDPR/HIPAA happy.
  3. Camera commoditisation hitting analytics SaaS. When a $200 NDAA-compliant Hanwha bundled with a $300 edge node delivers what a $2k smart camera did in 2024, the per-camera SaaS pricing model gets squeezed. Expect consolidation.

For the broader trend lines on Android-side surveillance and edge AI, see our Android video surveillance AI trends in 2026.

21. FAQ

How accurate is AI video analytics in 2026?

For object detection on COCO-class targets, top open models (YOLOv9-E, RT-DETR-X) reach 56% mAP. In real deployments, scene-tuned recall on safety-critical events lands around 90–95% with false-positive rates under 3 per camera per day — with the right alert design.

Can I add AI analytics to my existing camera fleet?

Yes — if your cameras are ONVIF-compliant and at least 2MP at 15fps, you can layer analytics via an on-prem appliance (Spot AI, Camio, or a custom Jetson tier). NDAA-restricted brands are a separate gating question for federal customers.

Is facial recognition legal in my jurisdiction?

Mixed. The EU AI Act prohibits real-time biometric identification in public spaces with narrow exceptions; the US has no federal ban but state laws (Illinois BIPA, Texas, Washington) require informed consent. We strongly recommend a written legal opinion before deployment.

Edge or cloud — which should I pick?

Edge for live alerting, cloud for forensic search and model training. Above 50 cameras the bandwidth and latency math makes edge-first the default. Below 20 cameras pure cloud is often simpler and cheaper.

How long does a typical deployment take?

For a 200-camera multi-site deployment: 12–20 weeks for an off-the-shelf rollout, 16–24 weeks for a custom build with our Agent-Engineering practice (vs 24–36 for a traditional shop). Pilot to first signal: typically 3–5 weeks.

What does NDAA Section 889 actually ban?

The use, sale, or integration of equipment from Hikvision, Dahua, Hytera, Huawei, and ZTE in federal contracts — including at the component level. State and enterprise procurement is increasingly mirroring it.

Can AI prevent shoplifting in real time?

It can detect — sweethearting, skip-scans, ORC patterns — in real time, and it can alert a loss-prevention officer’s phone in under a second. Whether it “prevents” depends on your store’s response posture and local law on intervention.

What’s the difference between video surveillance and video analytics?

Surveillance is the recording layer (cameras + VMS + storage). Analytics is the intelligence layer that interprets the footage — detecting objects, classifying behaviours, generating events. Modern systems combine both.

Integration guide

How to integrate AI video analytics with your existing surveillance stack

VMS hooks, SDK quirks, and the integration patterns that work in production.

Models

Anomaly detection models for video surveillance

Memory autoencoders, normality models, and what works on real CCTV.

Compliance

AI video surveillance & ethics in 2026

EU AI Act high-risk obligations, BIPA, and how to keep your DPO happy.

Retail

Retail video analytics: from shrinkage to conversion

Sweethearting detection, queue analytics, and the playbook our retail customers run.

To sum up

AI video analytics for security in 2026 is no longer a research project — it’s a procurement category. The hardware is cheap, the open models are excellent, and the regulatory landscape has matured to the point where you can plan around it. The work that remains is integration, alert design, and operational discipline. That’s where projects actually succeed or quietly fail.

If you’re evaluating an AI security analytics project — greenfield or layered onto an existing fleet — we’d love to help you scope it. We’ve been building this stuff since 2017 and we have the bruises to prove it.

Ready to scope your project?

Talk to our CTO — 30 minutes, no slides, real answers.

Book a call →
  • Technologies