Future of AIAI & Technology

Own Your Edge Control your AI

A lot of people today manage brick-and-mortar locations such as retail stores, grocery stores, or factory floors, and want to deploy AI solutions. However,ย it’sย hard to manage it, toย monitorย itย and toย maintainย control. In this article,ย I’llย talk about how to manage your Edge AI with relative ease, and how toย maintainย control over it.ย 

Why Edge AI?ย 

First up, edge AI lets you reduce latency dramatically by bringing computation closer to the data source. As they say, “Bandwidth is cheap, latency is expensive”.ย 

However, inย caseย of Vision AI, bandwidthย isnโ€™tย cheap either.ย Letโ€™sย run some quick numbers. For a typical fisheye camera, you can expect anywhere between 20 to 30 megabits of compressed video. For a medium-sized store, you would need about 8 cameras, totaling about 200 megabits bandwidth requirement. Which means you need a fiber internet connection. And fiber coverage in the US is sorelyย lackingย and whereย itโ€™sย available, it is prohibitively expensive.ย 

Running on edge lets you control your privacy and security posture. Your data, your customerโ€™s data, does not leave your premises.ย 

Finally, it lets you keep AI systems in your control. Weย donโ€™tย want a Skynet situation, do we?ย 

Pressure of AI adoptionย 

According to a recentย KPMG study, investor pressure for AI adoption jumped from 68% to 90% in just the last quarter.ย 

Over half of S&P 500 companies mentioned โ€œAI investmentsโ€ in their latest earnings calls.ย 

This surge reflects not just elevated chatterโ€”but intense pressure from investors, analysts, and board members. Everyone and their grandmothersย areย demanding an โ€œAI strategyโ€.ย 

The conclusion isย pretty clear; to keep your โ€œedgeโ€, you need to adopt AI.ย 

State of AI Adoptionย 

Let’sย talk about how the industry is adopting AI right now. Here are some reports from industry leaders.ย 

“95% of AI pilot programs fail to deliver P&L impact.” – MIT NANDA, July 2025ย 

“Orgs abandon 46% of AIย PoCsย prior to production.” – S&P Global Market Intelligence 2025 Generative AI Outlookย 

“According to Writer, March 2025, 42% of execs said AI adoption was tearing their company apart.” – Writer, March 2025ย 

Consider a typicalย new technologyย adoption curve below. All the signalsย indicateย thatย weโ€™reย inย this downward slope beforeย the productivityย stabilizes.ย 

The Pilot Problemย 

Lot of retailers are facing what I will call โ€œThe Pilot Problemโ€.ย Letโ€™sย sayย youโ€™reย running some AI pilots.โ€ฏย 

Vendor pilots in this space often arrive asย a black box.ย 

They have custom hardware, custom software, and custom telemetry.โ€ฏย 

For IT teams, it meansย theyโ€™reย now dealing with heterogeneous deployments: GPUs in camera or in a cooler screen or in displays or POS systems. You also haveย smartsย cameras now that lets you run some inferenceย on-device. On the one hand,ย itโ€™sย interesting tech. Butย itโ€™sย still anย additionalย complication, right? How do youย deploy toย it? How do youย monitorย it?ย 

You have limited or no visibilityย intoย these vendor systems.ย There’sย an uncomfortable security exposure: multiple authentication implementations and PII concerns throughout.ย 

IT teamsย forcedย to support, patch, and secure each pilot as ifย itโ€™sย production-grade. What looks like innovation to the business side often looks like chaos to IT.ย 

How?ย 

So,ย weโ€™veย establishedย you need Edge AI. Now,ย letโ€™sย talk about how you would manage it with ease. Ifย you’reย a retailer, this knowledge would let you judge how mature your vendors really are.ย 

AI Appsย 

Letโ€™sย talk about that first piece.ย Are AI applicationsย really special?ย Short answer: no, not really.ย 

AI applications areย basically justย software running on semi-specialized hardware.ย 

Weย can and shouldย use open and standard terms to describe AI apps. What are the inputs and outputs? What filesystem access does it need? What are the network requirements? What sort ofย computeย load does it generate? What kind of monitoring does it require? And so on. Any question you would ask for an API app, you should ask of an AI app.ย 

AI apps must be containerized. When Standard started, like a lot of young startups,ย deployย strategy was not mature. All software dependencies were installed manually. And deploys involvedย sshโ€™ingย onto the production server and running โ€œgit pullโ€.ย 

In conclusion,ย donโ€™tย pedestal your AIย – you want to version it,ย monitorย it, and move on.ย 

Orchestrationย 

So,ย you’veย builtย your AI apps, but you need a tool that can orchestrate how and when to run those applications. The best tool for orchestration of edge applications isย Kubernetes.ย ย 

You might ask: thereโ€™s no auto-scaling when youโ€™re on-prem, so why do you want the complications of Kubernetes? Itโ€™s a fair question. Kubernetes used to be hard. Thatโ€™s not really the case anymore. Kubernetes is much easier now, even on-prem. There are even lightweight Kubernetes flavors specialized for deploying on low-power edge devices. You can even run Kubernetes on a Raspberry-Piย at this point.ย 

Kubernetes lets you run containerized applications with ease and lets the system recover automatically. If a server fails, your app is moved to another one with very minimal downtime without any manual intervention. If youโ€™reย already using Kubernetes in the cloud, the uniformity of infra across cloud and edge would be super valuable.ย 

You can also easily deploy supporting actors, like maybe you need a message bus or a database or a cache.ย Itโ€™s allย really easy.ย 

Finally, my favorite part is that it lets me treat physical servers as disposable.ย Youย donโ€™tย want to get attached to the metal. Iย installย vanilla OS, my favorite Kubernetes distribution, andย thatโ€™sย it. Kubernetes handles everything else – including GPU drivers. If a server fails, I replace it with another one without skipping a beat.ย 

Application Specificationย 

Application specification encapsulates all the infra requirements for an app.ย Whatโ€™sย the stable version for each environment? How much memory does it need? What permissions does it need? And so on. And my favorite tool for managing app specs andย deploysย isย ArgoCD.ย 

It’sย declarative – Argo usesย what’sย called theย gitOpsย framework, more about that below. It automaticallyย rollย out your application changes and you can manage multiple environments from a single instance.ย 

There’s support for various forms of manifests – Helm,ย Kustomize, plainย YAML files.ย 

Finally,ย it’sย freeย and open source with a beautiful UI and UX.ย It’sย truly a joy to use.ย 

gitOpsย 

By far, my favoriteย deployย framework isย gitOps.ย It’sย the practice of using git repositories asย sourceย of truth for your infrastructure, including application specifications.โ€ฏgitOpsย lets you apply software development lifecycle best practices to your manifests. You can do code reviews, auditย yourย deploys, version them, and trivially roll back.ย 

Hereโ€™sย a workflow for how to adoptย gitOpsย 

ย 

Steps:ย 

  1. A developer commits application specification changes to aย specsย repo.ย 
  2. A watcherย componentย rendersย Kubernetes manifests and commits them to theย manifestsย repoย 
  3. ArgoCDย polls the manifests repo for changes and applies the changed manifests to theย appropriate clusters.ย 

All of this takes about a minute or two, soย yourย deploys areย really quick.ย 

Manage Compute (Cluster)ย 

So,ย weโ€™veย establishedย that you should use a standard application specification format and Kubernetes in combination withย gitOps.โ€ฏHowever, sinceย weโ€™reย running on edge, there are potentially thousands of independentย computeย clusters to manage.ย Soย we need some way to manage all these clusters.ย 

There are several solutions in this space. One of my favorite is Google Distributed Cloud (GDC). Itโ€™s based on Kubernetes, so you don’t have to change anything about application packaging or orchestration. There’s an integrated hardware and software offering which lets you offload some of the physical aspects of maintaining machines in the wild. GDC has first-class support for cutting-edge AI, including running some LLMs on edge!ย 

There’s also Rancher from SUSE which is an open source software based on Kubernetes. The installation is really straightforward, it’s really stable and battle tested, which extended (paid) support options.ย 

Talos takes an interesting approach of building a new Kubernetes optimized Linux distro. With a solid foundation, it adds first class emphasis on Kubernetes on bare metal. It’s fully managed using an API, so need to ssh onto a production server, ever! The disk is immutable, which increases resiliency and reliability of the system. And it comes with related offerings like Omni which let you control all your clusters from a single dashboard.ย 

If none of these fit the bill, you could also develop a “home grown” Docker Compose based solution.ย 

Monitoringย 

Now thatย we’veย deployed all this software, you obviously need some way to monitorย it’sย health. My recommendation is to use a standard stack ofย OpenTelemetry,ย Grafanaย andย Prometheusย and youย can’tย go wrong.ย 

All applications must expose a way toย monitorย its health using either metrics or HTTP APIs which can beย scraped. Theย applicationย health must encapsulate all details about what makes the application healthy. Is it able to reach its database? Is it able to keep up with the incoming throughput? Is it running into a crash loop? In addition, you want to use custom metrics to expose detailed health status.ย 

Conclusion: Your Edge AI Journey Starts Nowย 

The pressure to adopt AI is real, yet as we’ve seen, 95% of AI pilots fail to deliver P&L impact. The difference between success and failure? Control and standardization. Edge AI isn’t just about latency reduction or bandwidth savings, it’s about maintaining sovereignty over your AI infrastructure. By treating AI applications as what they truly areโ€”containerized software on specialized hardwareโ€”you can escape the pilot purgatory that traps so many organizations.ย 

The path forward is clear:ย Containerize everything, Embrace Kubernetes,ย Adoptย gitOps,ย Monitorย relentlessly, and Demand transparency.ย The retail landscape is littered with abandoned AI pilots and unfulfilled promises. But armed with the right approachโ€”treating AI as disciplined software engineering rather than magicโ€”you can join the 5% who deliverย real businessย value.ย 

Remember: The goalย isn’tย to haveย AI. The goal is to own your edge, literally and figuratively. Start with one properly architected deployment. Prove the model. Then scale with confidence.ย 

Your customers are waiting. Your competitors are moving. And now you have the blueprint to do Edge AI right.ย 

ย 

Author

Related Articles

Back to top button