A More Human Internet
An idea for a post that may never happen
All posts with the tag "llm"
An idea for a post that may never happen
Damn this VAnessa is hitting my feed with hard topics, I’m not sure whether to subscribe or to block. These top websites only feel worse every day, when I post on twitter and I get 4 likes by accounts that were created 5 minutes ago with racy profile pics it feels obvious. I wonder how larger accounts deal with it. Now that llms have made making these bots mimic humans easy It really makes you want out. I’ve really become a curmudgeon and leaning on rss over the past year, I dont like it, but idk what to do.
This is a crazy theory I did not realize was out there, but damn power just keeps costing more and more. She does not mention it here, but there are many sources of power for the grid that cost vastly different amounts to produce, generally “clean energy” solutions are harder and more expensive to bring online and don’t just turn on and off at the flick of a switch. Anyways, how are the power companies divying this power out to users, do some get preferred rates or supply? My rates just went up for the summer period “temporarily”. Our infrastructure is aging hard to upgrade and needs something done to it. Who’s really going to pay for it, these AI companies are throwing big numbers but do they have any real money? Do they have any real revenue after building out massive data centers filling them with the most expensive hardware? These guys are burning cash like crazy.
Justin has such great feeds on his site, I love how the main feeds are so prominant just to the left of the article you are reading. slops in particular feels like a great category. Saving this chat for later, or found it particularly interesting, but don’t really want to make a post about it.
there is literally no universe that this is true 10k lines and its not bug filled crap? ok Lex Luthor, its time to step away from the keys
Is this 10k real production code? Dry in the sense that it hasn’t re-implemented the same s3 api dozens of time? What language are we talking something dense like python? something very verbose like html? Maybe a language where you implement everything from scratch like lua. This matters a lot. Playing with little POC applications that dont mean anything I can quickly come up with 500-1k likes of code that I may never look at again. I’m sure I can come up wtih 10k decent lines of code a day.
But for the same application without duplicating everything over and over? For something that moves the needle and really matters?? every single day?? Consistently +10k, not 10k changes, not 10k deletes of yesterdays code. nah thats wack.
Fantastic write up on their experience in ai, opinions on ai being a hoax with a veil of reasonable usefulness. Arguing that most people do not understand enough to see the difference, and thought leaders see where it is now, see where it was yesterday, it must be going to general intelligence tomorrow and you all will loose your jobs without this. I appreciate the satirical language here.
Yes, I can review the code and make changes, but who in the world loves reviewing code? Do you love reviewing peers’ PRs? really?
I’m with MeetGor here 100%. reviewing the nuance, not being as involved with the process of creating the architecture design, not solving the problems that arise in development make it hard to effectively review and not turn into LGTM man.
I’ve been using gitingest web ui [[ thoughts-516 ]] for quite awhile to serialize git repo into llm friendly text files. This gives tools context about repos that are not in the training data so that it knows about it and how to use the code in the repo. gitingest also has a python library [[ thoughts-517 ]]
I had a use case for a project not yet on git, and found yek.
Their instructions tell you to curl to bash.
curl -fsSL https://bodo.run/yek.sh | bash
I don’t like curl to bash from random sites, so I have my own self...
...
Do you remember regression models from college: given some data, you find a best fit line that allows you to predict Y given X. At the end of the day, ChatGPT, and LLMs in general, are the same thing as the regression model – it’s just that ChatGPT is the largest and fanciest model we currently have to model language and information.
I really am coming to the idea of calling it a “word calculator”, this seems to be the most succinct description of llms that the lay person can comprehend and relate to.
ChatGPT does not hallucinate or become unhinged
I think Steve goes much deeper on this in his intervew on fafo.fm. They describe it more as a pleaser or “yes man” essentially all the companies that are building these models want to give the “best” answer, better than their competitors. With this comes the risk of it being completely wrong, they are designed to always give an answer.
...
Under 2000 everything is happy, green field. Any decision you have made is relatively easy to back out of (barring you making a library with downstream users), but as you go, regret kicks in. Regret we didn’t make that pydantic 2 upgrade earlier, as new features become more apealing. Regret that we chose sqlite for simplicity, speed, agility, and now we might need robust and distributed. Regret that you chose a front end framework, or to have a front end at all to a backend problem. Regret that you put 6 layers of abstraction on your db early on and now that you understand the problem you want different abstractions, but all of your endpoints deeply depend on the current one.
Vibe coding will not save you, it will only make these wrong decisions for you without the context that you have. You will hate it’s decisions more because you had no input into some of them.
Steve is such a great listen, the neurospicy 🌶️ rambles this episode goes on is so relatable. I feel like I really missed out on some great takes on intellij vs neovim, but got some really great knowledge about vector db’s, embedding, text compression, similarities to vector algegra like infinite craft.
Just popped open infinitecraft and I’ve definitely played this with my kids before, super fun, just could not remember the name of this one. I do remember an android one as well that is alchemist or something like that, which we have also played a lot.
Interesting how confidently he says we can easily go to the top. really makes you wonder what we the normies are leaving on the table by using these general purpose models and what could be achieved with really tuned in models. Could I make an automatic blog tagger more accurately, maybe smaller, maybe tuned so well it runs fine on cpu?
jina reader is a pretty sweet tool to convert a site to ai compatible text. There are other web to markdown types of tools, but the convenience of just adding r.jina.ai to the front of any page makes it so easy to grab for one page of docs.
the racked up 4 framework mainboards sound wild. connected with usb4 and 5gig ethernet. they said they can run big models quantized down from 600Gb to within the 512GB limit they have. This seems wild to bring this level of capability to such a low price point. It will be really cool to start to see demos come out.
OpenTools is an index for mcp servers to work with new agentic workflows like roo code and windsurf.
Damn prime makes an interesting point near then end of this video. He’s seen a bunch of people able to just throw down charts and shit at their company and end up being “the coding guy” cause they proompted something once. In a way I can relate, I got into software in a similar way, but at a time that it took a lot more hard work, understanding , and copy past from the right stack overflow. Based on some of the people around me at the time I can only imagine how some people must feel like they got pushed into it without wanting it, and now are building something they don’t know anything about with no care about it or care to build any expertise. Is the future proompted charts from enterprise chatgpt or do we only continue growing more need for software from here.
There is a glimmer of hope out there that normal people can scrap together enough gpu to really run the latest models themselves. The ui really appears to be having huge leaps forward such that doing things like rag is no longer such a research project that it was just a few years ago. So excited to see Prime go through this homelab exercise.
Damn these deepseek memes go hard. Wild to see openai get played by their own game.
It’s crazy that the normie news that I have seen on deepseek shows that the Chinese made what the Americans did at a fraction of the price, without taking notice that they are building on the shoulders of openai.
👏👏👏 This one is really good. I’m right there with him on most of this. I am very hesitant on subscription models, and all the ai tools feel like they are getting ready to be the next round of death by a thousand cuts, this time with pretty limited free tier and relatively high prices to run. I’m sure we will see companies get taken by huge bills soon by building off of someone else’s service.
On the flip side I’m definitely the guy that gets in a rut of just copy paste to the ai, wait for codeium to to inject. I feel like I have issues of momentum more than anything. When I’m on one side or the other I tend to stick it out for too long, but less so on going without because that llm drug is calling you when you hit a hard problem.
I’m excited to see him build out a homelab for llm stuff that he mentioned at the top. I’m interested, but probably not building one out for myself until we start to see some cheaper maybe used hardware to do it.
Don’t stop learning! Stop trying because you have a doomer outlook on ai, llms, industry and think they are taking over. If you have no hope for the future, if you stop now you are cementing in that you will be no good and the ai will be better. Many, maybe most of us in this industry go here by hard work, long nights of learning, trying to solve problems that our job had. If llms take over then the world is going to be a whole lot different, it will be a world you cannot predict or plan for. For now put your head down and succeed in the world we have today.
TEEJ has some great thoughts on this whole sentiment, put this on for you morning walk or whatever you do.