Posts tagged: llm

All posts with the tag "llm"

87 posts latest post 2026-04-19
Publishing rhythm
Apr 2026 | 13 posts

Really interesting way to generate a rules file for agentic workflows based on your current repo. John uses gitingest here, looks like a fantastic tool, but probably not useful for most private repos. I’m sure you can replicate the same thing in a private repo wtih a small amount of effort the few times you need to do it. gitingest looks like a great way to pull in some extra context for some open source dependencies that you have though.

Replace hub with ingest in any github and get a prompt friendly codebase ready to feed into any llm. It combines the entire codebase, based on a gitignore style glob that you pass in, into a single TXT file.

I recently updated ollama, and it now installs a systemd service that I was not expecting. Seems like a great option, but I hadn’t expeted this and I was able to kill it previously. It was using up gpu, and I do other things on my machine with a gpu. I tried pkill, kill, and everything, it was still coming back.

No matter what it comes back

# stop it systemctl stop ollama.service # disable it if you want systemctl disable ollama.service # confirm its status systemctl status ollama.service

You can confirm this with the following command.

# checking running processes ps aux | grep ollama pgrep ollama # checking gpu processes gpustat --show-cmd --show-pid

Next time you want to start you can do it as before with ollama serve.

This is pretty sick, I wanted this early on when I was making lockhart. I wanted to do the git hook thing but could not figure it out and did not know that prepare-commit-msg was a hook that I could use.

Git Hooked Then I remembered! Git hooks! Lol. Why would I have that in my brain - who knows!

I asked claude again, and they whipped up a simple script that would act as a hook that triggers with the prepare-commit-msg event.

This is awesome, cuz if you want to add a git message, you can skip the hook. But if you are lazy, you exclude the message and it will call the LLM.

...

This really makes me want to try Dolphin Mixtral with ollama now. It looks very impressive from this video. The ability to keep adding features before becoming confused is though with a lot of these llms.

Being chat based, this is not a co pilot replacement. I was really hoping for an in line co pilot like tool that I can run locally. I have not used co pilot yet, but I have had great luck with codeium.

This is a pretty sweet interface into llms. I used it a bit with my son tonight while he was asking me for datapack ideas.

❯ mods -f 'I am trying to have fun on my minecraft server and am creating a minecraft datapack send me some load.mcfuncions that will make it fun'

You can continue the conversation with a -C

❯ mods -C -f 'I like where you are going with number 4, can you make it so that it runs when a player opens a door'

You can pass it some data

ollama is the easiest to get going local llm tool that I have tried, and seems to be crazy fast. It feels faster than chat gpt, which has not been the experience I have had previously with running llm’s on my hardware.

Kinda mindblown that this is even possible. This is so far outside of my current thinking that i didn’t even think of an elegant way to implement semantic search accross images and text at the same time. I know it happens at Google, but I envision that as still text search accross tags and meta data about the image.

Based on the number of responses CLIP is the thing that does this.