• 1 Post
  • 13 Comments
Joined 1 year ago
cake
Cake day: June 14th, 2023

help-circle

  • For LLMs it entirely depends on what size models you want to use and how fast you want it to run. Since there’s diminishing returns to increasing model sizes, i.e. a 14B model isn’t twice as good as a 7B model, the best bang for the buck will be achieved with the smallest model you think has acceptable quality. And if you think generation speeds of around 1 token/second are acceptable, you’ll probably get more value for money using partial offloading.

    If your answer is “I don’t know what models I want to run” then a second-hand RTX3090 is probably your best bet. If you want to run larger models, building a rig with multiple (used) RTX3090 is probably still the cheapest way to do it.













  • It’s not an easy question, but I don’t think it’s about just choosing to stand idly by while they’re dying. Finding and recovering the sub would be an incredibly difficult and expensive operation. It might not apply to you specifically, but if someone thinks that the government should try to save these people, regardless of the cost, I think it raises the question why we’re letting other people die from preventable causes. Perhaps you disagree with current politics and think the government should do everything to save both rich and poor alike, but IMO if a multi-million $ rescue operation had been launched, it would’ve been a reminder that we, as a society, are letting other people die.