• 0 Posts
  • 8 Comments
Joined 3 years ago
cake
Cake day: June 14th, 2023

help-circle

  • And to put the pets dot com thing in context a bit, you have to remember that pets dot com was basically offering to ship pet food and pet furniture (notoriously voluminous and heavy objects) to people at a time when courier and direct-to-consumer freight services were still slow, insanely expensive and unreliable, the contemporary joke was that Fedex parcels would arrive looking like they had been run over (or even worse if they were marked fragile). This was before Amazon was doing free 2-day shipping and all the consequences that had in transforming the logistics market.

    So you have to realize that the world was a different place back then, and expectations were different, and the realities of doing business online were different, so when pets dot com said they were going to make a business out of selling notoriously voluminous and heavy objects online with no brick-and-mortar locations to customers who would then have to wait for it and be prepared for disruption in the shipping process while their animals starved, reasonable people thought they were completely fucking unhinged. But business people thought it was fucking genius.

    And it turns out they were in fact completely unhinged, and it was not genius, their failure ended up being emblematic for the complete insanity and detachment from reality that was going on while the “dot com” bubble was inflating, and represented the failure of the false idea that if you slapped a “dot com” on any particular industry you would capture and revolutionize that industry automatically, which is what pets dot com thought they were going to do. At the time everyone thought brick and mortar was so expensive and online was so cheap they were all going to be Netflix and completely kill their respective industry’s Blockbuster overnight. They were very wrong. Turns out you can’t just go ahead and start selling pet food online and replace all pet food stores and nobody had really thought that through, they were so overconfident that success was assured because they were doing it online and online is obviously better than brick-and-mortar in every way, right? Right?

    It is a very similar kind of insanity and a very similar breathless fascination with AI’s imagined potential to replace every job and be used in every business that is propelling the AI bubble. They will inevitably find out they are wrong too but they can stay irrational about it for a very long time before reality catches up with them and we don’t know exactly how harmful it is going to be when it finally does, other than “probably a lot”.



  • Part of what big tech has done is to divide us from one another and “curate” our information spaces to make it feel like we’re the only ones experiencing these feelings, like we are the only ones who are actually as desperate for change as we are, when the reality is that I think everyone is actually on pretty close to the same page for a lot of the same reasons. Believe it or not, we do all inhabit the same reality, we have just been made to feel that that reality is itself fictional. It does not serve big tech or big media or big government’s interests for us to know exactly how much we have in common, because they don’t want us to find a common purpose.


  • Absolutely. There are tons of open-licenced, open-weight (the equivalent of open-source for AI models) models capable of what is called “tool usage”. The key thing to understand is that they’re never quite perfect, and they don’t all “use tools” quite as effectively or in the same way as each other. This is common to LLMs and it is critical to understand that at the end of the day they are just text generators, they do not “use tools” themselves. They create specific structured text that triggers some other software, typically called a harness but could also be called a client or frontend, to call those tools on your system. Openclaw is an example of such a harness (and not a great or particularly safe one in my opinion but if you want to be a lunatic and give an AI model free reign it seems to be the best choice) You can use commercial harnesses too by configuring or tricking them into connecting to a local model instead of their commercial one, although I don’t recommend this for a variety of reasons if you really want to use claude code itself people have done it but I don’t find it works very well since all its prompts and tool calling is optimized for Claude models. Besides OpenClaw, Other popular harnesses for local models include OpenCode (as close as you’re going to get to claude for local models) or Cursor, even Ollama has their own CLI harness now. Personally I use OpenCode a lot but I am starting to lean towards pi-mono (it’s just called pi but that’s ungoogleable) it is very minimal and modular, making it intentionally easy to customize with plugins and skills you can automatically install to make it exactly as safe or capable or visual as you wish it to be.

    As a minor diversion we should also discuss what a “tool” is, in this context there are some common basic tools that some or most tool-use models will have or understand some variation of, out of the box. Things like editing files, running command-line tools, opening documents, searching the web, are common built-in skills that pretty much any model advertising itself capable of “tool use” or “tool calling” will support, although some agents will be able to use these skills more capably and effectively than others. Just like some people know the Linux commandline fluently and can completely operate their system with it, while others only know basic commands like ls or cat and need a GUI or guidance for anything more complex, AI models are similar, some (and the latest models in particular) are incredibly capable with even just their basic built-in tools. However they’re not limited by what’s built in, as like I said, they can accept guidance on what to use and how to use it. You can guide them explicitly if you happen to be fluent in their tools, but there are kind of two competing models for how to give them that guidance automatically. These are MCP (model context protocol) which is a separate server they can access that provides structured listings of different kinds of tools they can learn to use and how they work, basically allowing them to connect to a huge variety of APIs in almost any software or service. Some harnesses have an MCP built-in. The other approach is called “skills” and seems to be (to me) a more sensible and flexible approach to giving the AI model enough understanding to become more capable and expand the tools it can use. Again, providing skills is usually something handled by the harness you’re using.

    To make this a little less abstract you can put it in perspective of Claude: Anthropic provides several different Claude models like Haiku, Sonnet, and Opus. These are the text-generation models and they have been trained to produce a particular tool usage format, but Opus tends to have more built-in capability than something like Haiku for example. Regardless of which model you choose though (and you can switch at any time) you’ll be using a harness, typically “claude code” which is typically the CLI tool most people use to interact with Claude in an agentic, tool calling capacity.

    On the open and local side of the landscape, we don’t have anything quite as fast or capable as Claude code unfortunately, but we can do surprisingly okay considering we’re running small local models on consumer hardware, not massive data center farms being enticingly given away or rented for pennies on the dollar of what they’re actually costing these companies on the hopes of successful marketshare-capture and vendor-lock-in leading to future profits.

    Here are some pretty capable tool-use models I would recommend (most should be available for download through ollama and other sources like huggingface)

    • gemma4 (the latest and greatest hotness, MIT licensed using TurboQuant to deliver pretty incredible capability, performance and results even with limited VRAM)
    • qwen3.5 (from Alibaba, a consistent and traditional leader in open models so far with good capability and modest performance)
    • qwen3-coder-next (a pretty huge coding-focused model you might struggle to run unless you have a very beefy system and GPU)
    • glm4.7-flash (a modestly capable and reasonably fast option)
    • devstral-small-2 (an older, not-so-small variant of mistral, the French open-weight AI model if you’re looking for a non-Chinese, non-US based model which are few and far between)


  • Hot take: Manipulative and mentally destructive social media algorithms are the reason your sleep is disrupted. It’s what is on the screens that is the problem, not what color it is.

    But of course, the tech companies would rather have you blame the color of the screen than their own products. I’m sure they loved adding those color-shifting features to their next products too. not only do they avoid the blame, they get to sell you the “solution”.


  • This is the kind of nuanced usage of AI I like to see. Some would argue it’s not ideal to use any AI at all, and I agree, but we don’t live in an ideal world and I think this is realistically fine. AI writes better tests and docs than the ones I never write. Sure, maybe they’re not great objectively speaking, but they’re not worse than nothing. It’s better at keeping them up to date than I am too. Which is also probably not great, but strictly better than me.