Skip to content
Background1b_small_60percent_opacity

My Personal AI : I am Aible!

I have been in the AI business for a while. Studied AI at Stanford in 1995 when the last generation of faculty that went through an ‘AI Winter’ were still teaching. Built an AI startup that was acquired by Salesforce. Wrote a book called “AI Is a Waste of Money” based on a thousand AI projects I had done at that time. Even co-created and co-taught the first AI course in the Harvard Business School MBA program. But until today I couldn't reconcile three key fatal flaws of the world’s current approach to AI.

1. Why are we obsessing over complex generalized agents that are expensive to build and govern, instead of agents that improve our own lives?

We work with several Fortune 500s and the common pattern is that the scope of an agent is constantly expanded until it is no longer good at any one thing. At the same time, users constantly come up with ideas that can save them one hour a day. Over and over again, we have seen them build such an agent that is ‘good enough’ for what they want in a matter of minutes. For an employee with a fully loaded annual cost of $100,000, a single such agent that saves them 1 hour a day delivers a $10,000 impact. But $10,000 seems too small a payback to pursue building an agent? Could we maybe generalize this agent so 1,000 people can use it? So that we can have a $10 Million impact? But now the agent has to be good enough to meet the superset of the needs of the 1,000 users, it has to be governed appropriately, the operating costs have to be managed, etc. and the cost of creating and managing the agent just grew by more than 1000X. So, your ROI as a percent of costs actually dropped. Then why are we doing this?


2. Why are we sacrificing our individuality and privacy to gain an all-powerful AI Overlord?

One of my formative experiences was watching the 1984 Apple ad celebrating the power of the individual against a dehumanizing AI overlord. How did we go from that powerful vision to where the world has almost accepted that eventually we will have to sacrifice all of our data to the altar of Artificial General Intelligence (AGI)? Clearly, we want the benefits of powerful AI AND our individuality and privacy. Enterprises cannot give their data to a public AI. And thus, AGI efforts will essentially remain consumer solutions, because they are only learning from data in the public domain (or at least accessible from public sources). Why then are we treating this as a tradeoff? We have already proven that small models that are post-trained for specific tasks, data, terminologies, processes perform much better than 100 times larger generalized models. Many independent analysts have shown that large models that perform wonderfully on public benchmarks perform very poorly on enterprise data that they have never seen. Then why are we not building thousands and millions of specialized agents that do specific tasks much better than a potential AGI, and in the aggregate can already deliver much of what we want from AGI?

3. Why are we building platform-specific AI when we know the greatest power of AI will come from spanning silos?

An order to cash AI that can look at an order, check against contract terms, check inventory availability, estimate payment risk, optimize delivery options, then auto-generates the order response is far more powerful than agents doing each individual step. An agent that runs at the telco edge to detect issues, checks with a centralized agent to see if similar patterns have been encountered elsewhere, analyzes prior repair logs in a third system to formulate a response and automatically applies a fix via a fourth system is far more powerful than an agent built into the service ticket system alone. Enterprises always span platforms, different clouds / on prem / edge, different CRM / ERP / Finance, different geographies, etc. then why are we building point solutions when the value will come from AI systems that will span them all?


So, what changed today that helped us finally solve these three fundamental flaws? Aible announced the personal AI on the personal supercomputer. 



All of Aible’s capabilities can now run on a single small desktop server in the form of the NVIDIA DGX Spark. Aible can run the agents superfast by running the user experience, the agent coordination, agent tools including vector databases, and the language models, all on the same silicon with shared memory and communicate synchronously between them without writing anything to disk. This is faster than current alternatives because agents on the cloud have to communicate asynchronously, writing data back and forth between different servers running the models and tools, and coordination logic.  It can do all this completely air-gapped to ensure complete privacy. And it does so at a fixed annual cost per user so there is no surprise language model expenses. The device and all software is included in the annual cost. How does this change things?

 

1. The individual can now build agents that improve their own lives because:

  • There is no security risk from the agent – because it runs solely on their own personal supercomputer

  • There is no marginal cost from the agent – because it runs solely on their own personal supercomputer

  • They can make the agent uniquely their own for their own ‘job to be done’ – because again it is their own personal agent running on their own personal supercomputer

But can users really build their own agents? Well, we have already proven that they can. For example, at the State of Nebraska hackathon, 36 users with no prior training, and most with no AI expertise, built 222 agents in 90 minutes. Users can also share their agents or agent templates with other users or publish them to internal or public marketplaces. Also with a push of a button, the agents can be published at enterprise scale on all three major clouds or in private cloud solutions from HPE & Dell.

2. There is no compromise between accuracy and privacy because: 

  • There is no privacy risk from the agent – because the data remains solely on their own personal supercomputer

  • Accuracy is improved – because multiple specialized post-trained variants of the models (different LORA layers but the same base model) can run efficiently on the same personal supercomputer

  • Complexity is reduced and thus the overall solution is more efficient – because a given user uses a small subset of use cases, agentic tools, and model variants, these can be kept efficiently in the 128 GB of built-in fast memory and thus run more optimally

But can users really post-train models themselves? Can these post-trained models actually outperform larger models? We have already proven they can. Business users can easily provide focused feedback on the reasoning steps of the models. Aible completely automates the post-training process from data collection to post training to validation to model management. We even showed that a small model like an 8 Billion parameter Llama variant can outperform 100 times larger proprietary models with 22 minutes of post-training. Note that we are already running 120 Billion parameter models on the NVIDIA DGX Spark and believe it will efficiently hold 405 Billion parameter models as well. So, the open source ‘small’ models here are much closer in size to the proprietary large language models that are only 2 to 10 times larger at this point.

3. Agents can run on any platform while coordinating agents across multiple platforms because:

  • The exact same code that is running on the NVIDIA DGX Spark also runs on all three major clouds, on HPE and Dell Private Clouds, on most NVIDIA processors, at the telco edge as proven in the NVIDIA Aerial Lab and by Fujitsu/1Finity. Lighter workloads can even run on Intel Xeon processors.

  • The Aible agents can process data where it is in Snowflake, Big Query, Databricks,  Redshift, etc. and only move necessary metadata to the personal supercomputer where the agents are running

  • Aible uses open standards like Model Context Protocol (MCP) and open source capabilities like the NVIDIA Agent Toolkit that enables it to talk to a wide variety of agents that are being built into various platforms. Aible agents can talk to and coordinate across both Aible agents and standards compliant external agents.

But can a startup really run across all these platforms? Our partners have already publicly validated that we can. See NVIDIA, Intel, HPE, Google, Snowflake, AWS, Microsoft,  etc. and several other announcements are imminent. In fact, we spent seven years building the comprehensive solution that could be truly platform and model agnostic. The integration with other types of agents are assured by the open standards and tooling that we have adopted.

But, there is no better proof than the reactions of real users. At the Gartner Symposium event in October 2025, we asked users to build their own agents using Aible on the NVIDIA DGX Spark. Here are some videos of those interactions. The excitement is palpable.

Now where do we go from here? The DGX 1 was released 9 years before the DGX Spark. The DGX Spark is 1000X more powerful and 8X smaller. In the last couple of years, open source models have gone from 7 Billion parameters to over 405 Billion parameters. The personal supercomputer will only get smaller and the personal AI will only get smarter. In the not so distant future, each of us will have our own personal AI on our own personal supercomputers that keep our data private, understand our needs across use cases perfectly, and efficiently interact with the necessary external systems to get the job done. That is the empowering future Aible is building. Thus, our enduring motto since 2018: I am Aible!