<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Local Inference on Thede Technologies</title><link>https://thedetech.com/tags/local-inference/</link><description>Recent content in Local Inference on Thede Technologies</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Thu, 23 Apr 2026 22:42:53 -0500</lastBuildDate><atom:link href="https://thedetech.com/tags/local-inference/index.xml" rel="self" type="application/rss+xml"/><item><title>A general-purpose MoE multimodal beat every dedicated vision model on my father's handwriting</title><link>https://thedetech.com/blog/2026-04-23-moe-beats-dedicated-vision/</link><pubDate>Thu, 23 Apr 2026 00:00:00 +0000</pubDate><guid>https://thedetech.com/blog/2026-04-23-moe-beats-dedicated-vision/</guid><description>&lt;p>For context: I&amp;rsquo;ve been transcribing a multi-generational archive of handwritten family letters on my own hardware. The first two posts covered &lt;a href="https://thedetech.com/blog/2026-04-23-family-archives-ai-can-read/">why I&amp;rsquo;m doing this at all&lt;/a>
 and &lt;a href="https://thedetech.com/blog/2026-04-23-local-models-aging-mac/">how to set it up on your own machine&lt;/a>
. This post is the surprising finding — the one I didn&amp;rsquo;t expect going in.&lt;/p>
&lt;p>I assumed the right tool for a vision task was a vision model. That&amp;rsquo;s the obvious reach. If you&amp;rsquo;re reading handwriting, you reach for something labeled &amp;ldquo;VL.&amp;rdquo; If you can find one fine-tuned for OCR and handwriting specifically, even better.&lt;/p></description></item><item><title>Running real models locally on a Mac Studio that isn't new anymore</title><link>https://thedetech.com/blog/2026-04-23-local-models-aging-mac/</link><pubDate>Thu, 23 Apr 2026 00:00:00 +0000</pubDate><guid>https://thedetech.com/blog/2026-04-23-local-models-aging-mac/</guid><description>&lt;p>I have a multi-generational archive of handwritten family letters sitting in my house, and I wanted to read it without sending the private content to a cloud provider. The &lt;a href="https://thedetech.com/blog/2026-04-23-family-archives-ai-can-read/">first post in this series&lt;/a>
 is the &lt;em>why&lt;/em>. This one is the &lt;em>how&lt;/em> — on the specific hardware I already own, with the specific software I landed on.&lt;/p>
&lt;p>My Mac Studio is almost four years old now, but it has 64 GB of RAM and it has an M1 Max chip. It has proven itself to be quite capable of running large language models locally for personal benefit. It helps with maintaining privacy, and it also is much less expensive than leveraging any cloud provider.&lt;/p></description></item><item><title>What family archives are for, now that the AI can read them</title><link>https://thedetech.com/blog/2026-04-23-family-archives-ai-can-read/</link><pubDate>Thu, 23 Apr 2026 00:00:00 +0000</pubDate><guid>https://thedetech.com/blog/2026-04-23-family-archives-ai-can-read/</guid><description>&lt;p>I have three collections of letters sitting in my house.&lt;/p>
&lt;p>The oldest is a stack from my great-grandparents, written between 1910 and 1913. The second is from my grandparents, written during World War II. The third is from my father — letters he wrote from the moment I was born until the moment his mother died.&lt;/p>
&lt;p>There are other things too. Memoirs. Loose documents. A family record that shows up in more places than I can keep track of. But those three collections are the weight of it. Three generations, each one writing to the next, each one leaving behind something I&amp;rsquo;ve never fully read.&lt;/p></description></item></channel></rss>