Current LLM tech stack

Published on November 29, 2024
Current LLM tech stack

because people ask, not because I recommend

November 2024:

For coding, I use Cody in my local IDE VSCode and whatever is on offer in cloud IDEs (recently CoPilot and ReplitAI).

For getting my code to talk with LLMs, I use Simon Willison's llm.

For running models in the cloud, I use Replicate. I've used Ollama locally – but honestly I haven't touched a local model recent months. I use DrawThings to explore image-making locally (but while generative, that's not using an LLM).

For general stuff (including explaining code and learning new stuff), I currently use msty, which allows chat-like interfaces with several AIs via their API. This lets me set up the context for a chat with system prompts, and handily means that I can pay-as-I-go with AIs, which is rather cheaper for me than a monthly subscription.

I'm currently using Claude-3.5-Sonnet, GPT 4o and GPT 4o-mini via their API