I prefer waterfox, OpenAI can keep its Chat chippy tea browser.
Here’s the thing: I’m not willing to pay for AI. I liked Siri when its “cookie monster” joke about dividing zero by zero wasn’t considered offensive and before it had to Google everything or ask ChatGPT for everything. Now I just don’t care about it at all. And that’s Siri — I’m intentionally on the platform with the crappiest, deadest, most useless AI because I really don’t want AI in my life. And it’s great.
As long as I can use Firefox on the Mac and not worry about AI — Firefox did add some chatbot thing, but it was very easy to disable — I’m just going to keep doing that.
My only worry will be, at some point, the Net might get to where you need AI. Hopefully by then they will have figured out a way to make it free. I hope I can just ride that wave. If not, who knows. I worry for younger users, though many of them seem to be embracing the changes, kind of like how we embraced Web 2.0 before social media went to shit (and that was before fascists started taking over/spinning up their own).
Net might get to where you need AI
I hate to say it, but we’re basically there, and AI doesn’t help a ton. If the net is slop and trash, there’s not a lot it can do.
Hopefully by then they will have figured out a way to make it free.
Fortunately self hosting is 100% taking off. Getting a (free) local agent to sift through the net’s sludge will be about as easy as tweaking Firefox before long.
You can already do it. I already do it (and am happy to ramble about how when asked), but it’s more of an enthusiast/tinkerer thing now.
Local is also slower and… less robust in capability. But it’s getting there. I run local AI and I’m really impressed with gains in both. It’s just still a big gap.
We’re headed in a good direction here, but I’m afraid local may be gated by ability to afford expensive hardware.
Not anymore.
I can run GLM 4.6 on a Ryzen/single RTX 3090 desktop at 7 tokens/s, and it blows lesser API models away. I can run 14-49Bs (or GLM Air) in more utilitarian cases that do just fine.
And I can reach for free/dirt cheap APIs called locally when needed.
But again, it’s all ‘special interest tinkerer’ tier. You can’t do that with
ollama run, you have to mess with exotic libraries and tweaked setups and RAG chains to squeeze out that kind of performance. But all that getting simplified is inevitable.I’ll look into it. OAI’s 30B model is the most I can run in my MacBook and it’s decent. I don’t think I can even run that on my desktop with a 3060 GPU. I have access to GLM 4.6 through a service but that’s the ~350B parameter model and I’m pretty sure that’s not what you’re running at home.
It’s pretty reasonable in capability. I want to play around with setting up RAG pipelines for specific domain knowledge, but I’m just getting started.
I have access to GLM 4.6 through a service but that’s the ~350B parameter model and I’m pretty sure that’s not what you’re running at home.
It is. I’m running this model, with hybrid CPU+GPU inference, specifically: https://huggingface.co/Downtown-Case/GLM-4.6-128GB-RAM-IK-GGUF
You can likely run GLM Air on your 3060 desktop if you have 48GB+ RAM, or a smaller MoE easily. Heck. I’ll make a quant just for you, if you want.
Depending on the use case, I’d recommend ERNIE 4.5 21B (or 28B for vision) on your Macbook, or a Qwen 30B variant. Look for DWQ MLX quants, specifically: https://huggingface.co/models?sort=modified&search=dwq
I’m going to upgrade my ram shortly because I found a bad stick and I’m down to 16GB currently. I’ll see if I can swing that order this weekend.
To what?
64G would be good, as that’s enough to fit GLM Air. There are some good 2x64GB 6000Mhz kits for 128GB as well.
deleted by creator
OpenAI: “Chrome is a monopoly that must be stopped!”
Also OpenAI: releases a spyware browser based on Chromium
deleted by creator
Are you talking about chrome?
deleted by creator
deleted by creator
One of the examples of its brilliance is how he found a site he visited yesterday by typing in a question for the AI rather than pressing CTRL-H. I guess it is designed for people who type “Google” into the search bar of their browser and then double click on the link.
Hell no!
Why is ChatGPT and proplexity now jumping into the browser bandwagon.
Because there are morons out there who haven’t figured out OpenAI would do or say anything to keep the scam going.
Gotta keep the investor money flowing.
“Trust me bro, I know x didn’t work but y will, give me a few more billions bro, AI will make us rich bro”
Don’t know why this article is getting downvoted, it’s a solid article that feels relatively balanced.
I say this as someone who doesn’t trust AI conmen (while recognizing the enormous potential of LLMs and ML systems such AI upscaling of older SD/VHS videos), especially American AI conmen.
Because AI conmen do not deserve air time. This is a puff piece, not news (I assume, I’m not going to read it).
Anything even tangentially AI, anything that even sounds like AI gets mass downvotes on Lemmy.
Don’t get me wrong, Sam Altman is an even bigger con artist than Musk. But outside the self hosting niches, Lemmy skews towards whatever the opposite extreme of ‘tech bro’ is.







