What llm did you use, and how long ago was it? Claude sonnet usually writes pretty good python for smaller scripts (a few hundred lines)
What llm did you use, and how long ago was it? Claude sonnet usually writes pretty good python for smaller scripts (a few hundred lines)
Any moment now…
The printing press, of course
Last two are clearly “US dollars” and “printing press”
I still use http a lot for internal stuff running in my own network. There’s no spying there… I hope … And ssl for local network only services is a total pita.
So I really hope browsers won’t adapt https only
But even if you use GoMommy extra super duper triple snake oil security checked ssl cert, if I trick LetsEncrypt to sign a key for that domain I still have a valid cert for your site.
They want to force people to be hetero, Christian, and either white male and upper class, or anything else and subservient.
So for them, other groups forcing/ brainwashing people to be gay/trans, atheist and so on makes perfect sense.
I doubt the disk will bottleneck at 40mb/s when doing sequential write. Torrent downloads are usually heavy random writes, which is the worst you can do to a HDD.
I was trying to find an article I read about a year ago, about an experiment where AI was assisting a doctor. Where it suggested questions and possible diagnosis for the doctor to look into.
IIRC the result was both faster and more accurate diagnosis. Too bad I can’t find it again now :(
You’re not great taking medical advice from a doctor either, seeing how often they’re wrong.
Well this one needs a new job, maybe he can work school security. He’s available m-f after all
Llama3 8b can be run at 6gb vram, and it’s fairly competent. Gemma has a 9b I think, which would also be worth looking into.
That’s super green!
“braid made us money. We like money. Braid stopped giving us money. We want more money”
Isn’t there a Geneva convention against inflicting such horror on an enemy?
Better background backups
Rework background backups to be more reliable
Hilarious for a system which main point / feature is photo backup
I worked on one where the columns were datanasename_tablename_column
They said it makes things “less confusing”
Reasonable smart… that works preferably be a 70b model, but maybe phi3-14b or llama3 8b could work. They’re rather impressive for their size.
For just the model, if one of the small ones work, you probably need 6+ gb VRAM. If 70b you need roughly 40gb.
And then for the context. Most models are optimized for around 4k to 8k tokens. One word is roughly 3-4 tokens. The VRAM needed for the context varies a bit, but is not trivial. For 4k I’d say right half a gig to a gig of VRAM.
As you go higher context size the VRAM requirement for that start to eclipse the model VRAM cost, and you will need specialized models to handle that big context without going off the rails.
So no, you’re not loading all the notes directly, and you won’t have a smart model.
For your hardware and use case… try phi3-mini with a RAG system as a start.
This is a very simple one, but someone lower down apparently had issue with a script like this:
https://i.imgur.com/wD9XXYt.png
I tested the code, it works. If I was gonna change anything, probably move matplotlib import to after else so it’s only imported when needed to display the image.
I have a lot more complex generations in my history, but all of them have personal or business details, and have much more back and forth. But try it yourself, claude have a free tier. Just try to be clear in the prompt what you want. It might surprise you.