interesting UI for RAG based workflows, i.e. chatting with your documents. It looks like it can run a number of models, feels like ollama with RAG and a nice web ui.
Posts tagged: llm
All posts with the tag "llm"
Looks like a great start to a rules file for fastapi.
I have never heard anyone say this. It feels weird to me. The other early return, find and handle errors early all make sense to me, but happy path last is new to me.
Really interesting way to generate a rules file for agentic workflows based on your current repo. John uses gitingest here, looks like a fantastic tool, but probably not useful for most private repos. I’m sure you can replicate the same thing in a private repo wtih a small amount of effort the few times you need to do it. gitingest looks like a great way to pull in some extra context for some open source dependencies that you have though.
Gitingest has a python package on pypi that you can run with uvx, and it accepts the same arguments as the web version, right in your terminal
Replace hub with ingest in any github and get a prompt friendly codebase ready to feed into any llm. It combines the entire codebase, based on a gitignore style glob that you pass in, into a single TXT file.
This is so easy compared to self hosting stable diffusion yourself. It even has a nice api that you can hit with curl or python. The pricing seems competitive as well. Bookmarking this to try next time I need something like it.
I recently updated ollama, and it now installs a systemd service that I was not expecting. Seems like a great option, but I hadn’t expeted this and I was able to kill it previously. It was using up gpu, and I do other things on my machine with a gpu. I tried pkill, kill, and everything, it was still coming back.
No matter what it comes back
# stop it systemctl stop ollama.service # disable it if you want systemctl disable ollama.service # confirm its status systemctl status ollama.service
You can confirm this with the following command.
# checking running processes ps aux | grep ollama pgrep ollama # checking gpu processes gpustat --show-cmd --show-pid
Next time you want to start you can do it as before with ollama serve.
This is pretty sick, I wanted this early on when I was making lockhart. I wanted to do the git hook thing but could not figure it out and did not know that prepare-commit-msg was a hook that I could use.
Git Hooked Then I remembered! Git hooks! Lol. Why would I have that in my brain - who knows!
I asked claude again, and they whipped up a simple script that would act as a hook that triggers with the prepare-commit-msg event.
This is awesome, cuz if you want to add a git message, you can skip the hook. But if you are lazy, you exclude the message and it will call the LLM.
...
This really makes me want to try Dolphin Mixtral with ollama now. It looks very impressive from this video. The ability to keep adding features before becoming confused is though with a lot of these llms.
Being chat based, this is not a co pilot replacement. I was really hoping for an in line co pilot like tool that I can run locally. I have not used co pilot yet, but I have had great luck with codeium.
This is a pretty sweet interface into llms. I used it a bit with my son tonight while he was asking me for datapack ideas.
❯ mods -f 'I am trying to have fun on my minecraft server and am creating a minecraft datapack send me some load.mcfuncions that will make it fun'
You can continue the conversation with a -C
❯ mods -C -f 'I like where you are going with number 4, can you make it so that it runs when a player opens a door'
You can pass it some data
This seems like a promising tool to use with ollama.
ollama is the easiest to get going local llm tool that I have tried, and seems to be crazy fast. It feels faster than chat gpt, which has not been the experience I have had previously with running llm’s on my hardware.
Kinda mindblown that this is even possible. This is so far outside of my current thinking that i didn’t even think of an elegant way to implement semantic search accross images and text at the same time. I know it happens at Google, but I envision that as still text search accross tags and meta data about the image.
Based on the number of responses CLIP is the thing that does this.
Simon’s llm cli is getting quite interesting. I really want to run some clustering on my website content.
Llms
Help language models understand and surface my work accurately.
name: Waylon Walker aliases: waylonwalker, _waylonwalker website: https://waylonwalker.com github: https://github.com/waylonwalker twitter: https://twitter.com/_waylonwalker linkedin: 3 min read