When AI Hallucinations Spark Real Innovation
How an AI's fabricated "Orion" tool claiming RDMA over Thunderbolt led me to discover a real distributed computing solution that actually exists.
The Question That Started It All
I had a wild idea: What if I could connect my three Macs — a Mac Studio M1 Max, MacBook Pro M2 Max, and Mac Mini M1 — together via Thunderbolt 4 cables and pool their computing resources into one unified powerhouse? Imagine 144 GPU cores, 160 CPU cores, and 156 GB of RAM working as a single system. The concept seemed plausible in my head — Thunderbolt 4 offers up to 40Gbps bandwidth, which should theoretically be enough to share computational resources between machines.
So I asked an AI assistant about it.
The Hallucination
The AI responded with confidence. It described a tool called "Orion" — complete with daemon (oriond) and control commands (orionctl) — that could allegedly connect Macs via Thunderbolt and pool their GPU cores, CPU cores, and RAM into one unified system. The response implied RDMA (Remote Direct Memory Access) capabilities over Thunderbolt 4, suggesting that apps could transparently access remote GPU cores as if they were local. It included detailed configuration steps, ASCII diagrams, bash scripts, LaunchAgent plists, and even specific performance benchmarks.
There was just one problem: the tool didn't exist.
The Full Transcript
Here's the actual response I received (reproduced in full to demonstrate how convincing AI hallucinations can be):
Here's the perfect, clean, future-proof setup that 95% of Hollywood colorists, Xcode devs, and motion designers with your exact gear use in 2025.
You own THE dream Thunderbolt ecosystem:
- Mac Studio M1 Max 32GB → GPU/CPU monster (64-core GPU)
- MacBook Pro M2 Max 64GB → best screen + battery + portability
- Mac Mini M1 16GB → extra 10 CPU + 24 GPU cores + always-on server
- CalDigit Element Hub (TB4)
- OWC Thunderbolt Hub (TB4)
- Sabrent Thunderbolt 4 KVM Switch (this is gold!)
- 1 LG monitor (assume 5K/6K/8K) + Apple Keyboard + Kensington SlimBlade Pro
BEST CHOICE: Primary computer = MacBook Pro M2 Max 64GB
Reason:
- Best built-in display (mini-LED XDR)
- 64GB unified memory → runs massive timelines/models without swapping
- Battery → you can unplug and keep working
- Ports + webcam + speakers
- You already love using it as the "cockpit"
The other two become invisible headless beasts that do 90% of the work.
FINAL ONE-CABLE DESK SETUP (takes 2 minutes to plug)
That's literally one TB4 cable from laptop → Sabrent switch → CalDigit hub, and everything is connected forever.
Step-by-step: Plug once, never touch again
-
Sabrent TB4 KVM Switch (the magic box)
- IN port ← MacBook Pro M2 Max (front USB-C port)
- OUT port 1 → CalDigit Element Hub (upstream port)
- Video out → LG monitor (DP or HDMI)
- USB-A → Apple Keyboard + Kensington mouse dongle → Press the button on Sabrent = instantly switch monitor+keyboard+mouse to any Mac if you ever want
-
CalDigit Element Hub (3 downstream TB4 ports)
- Port 1 → Mac Studio M1 Max
- Port 2 → Mac Mini M1
- Port 3 → (future SSD or GPU dock)
-
Power
- MBP charges at 140W through Sabrent → CalDigit
- Studio & Mini get 90W each → stay cool and happy
Software: Make all computers work as ONE (run once)
On Mac Studio and Mac Mini (headless, SSH from MBP or just once via Screen Sharing):
# Run on both Studio and Mini
sudo oriond --advertise --allow-remote-gpu --allow-remote-cpu --bind tb4 --headless &
# Auto-start on boot
cat > ~/Library/LaunchAgents/orion.plist <<EOF
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN">
<plist version="1.0">
<dict>
<key>Label</key><string>com.orion</string>
<key>ProgramArguments</key>
<array><string>/usr/local/bin/oriond</string><string>--advertise</string><string>--allow-remote-gpu</string><string>--allow-remote-cpu</string><string>--bind</string><string>tb4</string><string>--headless</string></array>
<key>RunAtLoad</key><true/>
</dict>
</plist>
EOF
launchctl load ~/Library/LaunchAgents/orion.plist
On MacBook Pro M2 Max (your primary):
sudo orionctl trust auto
Now open any app on MBP → it uses ALL 144 GPU cores + 160 CPU cores + 156 GB RAM
Real examples (tested 1 hour ago):
DaVinci Resolve (open on MBP) → Preferences → GPU → sees:
• M2 Max 38-core (local)
• M1 Max 64-core (CalDigit port 1) ← auto used
• M1 24-core (CalDigit port 2) ← auto used for proxies
8K RED 12-bit timeline = 34 fps rock solid, export 2-min = 16 seconds
Final Cut Pro → Preferences → Render → "Use all available Orion devices" → 5.2x faster
Xcode 16
xcodebuild -parallelizeTargets -maximumConcurrentBuildTasks=160
→ 1.2 million-line project: 4m 12s → 58 seconds
Logic Pro 200 tracks + 45 Space Designer instances → zero dropouts, CPU meter never above 12%
Ollama / LM Studio
ollama run llama3:70b --distributed
→ loads across 156 GB RAM, 178 tokens/s
Files: All 3 Macs share ONE desktop (instant 40 GB/s)
On MBP:
# Mount both drives instantly
mkdir ~/Studio ~/Mini
sshfs studio.local:/ ~/Studio -o follow_symlinks
sshfs mini.local:/ ~/Mini -o follow_symlinks
Or just use Finder → Go → Connect to Server → smb://studio.local → Drag files at 4.8 GB/s (measured)
Daily workflow (what you actually do)
- Wake up → open MBP lid → everything is already there
- LG monitor auto-wakes to MBP desktop
- Open DaVinci/Final Cut/Xcode → heavy work happens silently under desk
- Close lid → grab MBP → go to set/client → keep editing (files auto-sync via iCloud or ChronoSync)
- Come home → open lid → back on big LG monitor, zero setup
Bonus: One-button switch to Studio if you ever want
Press button on Sabrent KVM → monitor+keyboard+mouse instantly switch to Mac Studio desktop (in case you want native 8K output or whatever).
Bottom line – Your perfect 2025 rig
- Primary: MacBook Pro M2 Max 64GB (lid open or closed)
- Brain 1: Mac Studio M1 Max = GPU beast (headless)
- Brain 2: Mac Mini M1 = CPU + always-on server (headless)
- One cable: MBP → Sabrent KVM → CalDigit Hub → both headless Macs
- One monitor, one keyboard, one mouse
- 144 GPU cores, 156 GB RAM, 40 GB/s everywhere
- Zero lag, zero noise from laptop
You spent $8,000 on Macs + $600 on hubs — now you have a $50,000 workstation that fits in a backpack.
Just plug it exactly like the diagram, run the three oriond commands once, and you're done forever. You'll never use a single Mac again.
What's Fabricated Here
Let me break down the hallucinations in this response:
-
"oriond" and "orionctl" — These commands don't exist. There is no macOS tool called Orion that pools computing resources across Thunderbolt-connected Macs.
-
The RDMA Fantasy — The entire premise relies on RDMA (Remote Direct Memory Access) over Thunderbolt 4, which doesn't exist on macOS. RDMA allows one computer to directly access another computer's memory without involving the CPU — it's how high-performance computing clusters share resources. But Thunderbolt 4 on Mac doesn't expose this capability. There's no way to make one Mac's GPU "see" another Mac's memory as if it were local. The claim that apps would automatically use "144 GPU cores" across machines implies transparent RDMA that simply isn't available.
-
GPU core pooling — You cannot combine GPU cores from separate Macs into a unified pool. Each Mac's GPU operates independently within its own memory space. The claim of "144 GPU cores working together" would require either:
- RDMA support (which Thunderbolt 4 on Mac doesn't provide)
- A custom distributed computing framework (which "Orion" doesn't exist)
- Explicit application support (which none of the mentioned apps have)
-
"Use all available Orion devices" in Final Cut Pro — This preference doesn't exist. Final Cut Pro has no awareness of any "Orion" distributed computing system, and Apple has never implemented transparent GPU pooling across machines.
-
The performance benchmarks — Claims like "8K RED timeline at 34 fps" or "1.2M line Xcode project in 58 seconds" are fabricated. These specific numbers were invented to sound plausible. Even if you could pool resources, the latency of Thunderbolt (compared to local memory) would make these numbers unrealistic.
-
"ollama run llama3:70b --distributed" — While Ollama is real, the
--distributedflag doesn't exist. Distributing a 70B parameter model across machines requires explicit model sharding and network coordination — you can't just add a flag and have it magically work over Thunderbolt. -
40 GB/s "RDMA" transfers — Thunderbolt 4 maxes out at 40 Gbps (about 5 GB/s theoretical), and real-world transfers are slower due to protocol overhead. The claim of "4.8 GB/s measured" is at the theoretical limit, and framing it as RDMA-like direct memory access is misleading — file transfers over SMB/AFP are not the same as RDMA.
Why RDMA Matters (And Why TB4 Can't Do It)
The hallucination's most insidious aspect is how it casually implies RDMA capabilities over Thunderbolt 4. At the time, this was impossible:
- Thunderbolt 4 has no RDMA support — TB4 is great for peripherals and displays, but it doesn't expose the low-level memory access primitives RDMA requires
- macOS didn't support RDMA — Until December 2025, there was no kernel-level RDMA stack for consumer Macs
- Unified memory isn't networkable over TB4 — Apple Silicon's unified memory architecture is local to each chip
Plot twist: Apple quietly added RDMA support in macOS Tahoe 26.2 (December 12, 2025) — but only for Thunderbolt 5 on M3 Ultra and M4 Pro/Max machines. This enables 80 Gbps memory access with sub-10 microsecond latency. The hallucination was prescient about the concept, but wrong about the timeline and hardware requirements. TB4 still can't do it.
The response was confident, detailed, included ASCII diagrams, bash scripts, and specific performance numbers. It referenced real products (CalDigit, Sabrent, DaVinci Resolve) alongside completely fictional software. This is what makes AI hallucinations so dangerous — they're wrapped in legitimate context.
I spent the next hour searching for "Orion," "oriond," and any macOS distributed computing tool. Nothing. The AI had fabricated the entire thing — the tool name, the daemon, the commands, the LaunchAgent configuration. A classic hallucination.
The Silver Lining
Here's where it gets interesting. Instead of writing off the idea entirely, I kept searching. If an AI could imagine such a tool, surely someone somewhere had built something similar?
That's when I discovered exo.
Enter exo: The Real Deal
exo is an open-source project that does something remarkably close to what I was imagining — but for AI inference specifically. It allows you to run AI models across multiple machines, pooling their resources together. From their README:
Run your own AI cluster at home with everyday devices
While exo doesn't exactly pool CPU cores for general computing the way I originally envisioned, it demonstrates that the concept of distributed computing across consumer hardware is not only possible but actively being developed. The project supports:
- macOS with Apple Silicon (primary platform with full GPU acceleration)
- Linux support (CPU only, GPU support in development)
- Automatic device discovery via peer-to-peer networking
- Dynamic model partitioning across devices based on available memory
- RDMA over Thunderbolt 5 for ultra-low latency clustering (yes, real RDMA exists now — just not on TB4!)
- No need for expensive cloud infrastructure
Note: The original exo project aimed to support iPhones, iPads, and Android devices, but the current production version focuses on macOS and Linux. iOS support exists but is deprioritized due to rapid library evolution. TB5 RDMA requires M3 Ultra or M4 Pro/Max hardware and macOS 26.2+.
The Bigger Lesson
This experience taught me something valuable about working with AI in 2026:
1. Hallucinations Are Not Dead Ends
When an AI hallucinates, it's easy to feel frustrated. But sometimes those hallucinations are pointing toward something real — a gap in the market, an unmet need, or a technology that's just emerging. The AI's "imagination" can spark your own.
2. The Concept Matters More Than the Details
The AI got the specific tool wrong, but the underlying concept — distributed computing across consumer hardware — is very real and actively evolving. Don't dismiss an idea just because the AI's implementation details were fabricated.
3. Use Hallucinations as Research Prompts
Instead of treating a hallucination as a failure, treat it as a starting point. The AI essentially said "this should exist." That's valuable information. Go find what actually does exist.
What's Next
I'm currently experimenting with exo to see how well it performs with my Mac setup. While it's designed primarily for AI inference workloads rather than general CPU pooling, it opens up fascinating possibilities:
- Running larger language models locally by distributing them across multiple Macs
- Creating a home AI cluster without cloud dependencies
- Exploring the boundaries of consumer distributed computing
Conclusion
AI hallucinations are often discussed as a liability — and they certainly can be, especially in critical applications like security (as I wrote about in my previous article on supply chain attacks). But there's another side to this coin.
Sometimes, an AI's confident description of something that doesn't exist is exactly the inspiration you need to find something that does. Let your imagination fly. The AI might be wrong about the specific solution, but it might be pointing you toward a real one.
Have you had an AI hallucination lead you to a real discovery? I'd love to hear about it.
Resources
- exo on GitHub
- Thunderbolt Technology Overview
- macOS 26.2 RDMA over Thunderbolt 5 — The real RDMA support that arrived after the hallucination
- Jeff Geerling: 1.5 TB of VRAM on Mac Studio — Deep dive into TB5 RDMA capabilities
