Alternate account for @[email protected]

  • 194 Posts
  • 406 Comments
Joined 2 years ago
cake
Cake day: July 3rd, 2023

help-circle
  • AMD’s compatibility has gotten better the past year, and ones that aren’t compatible usually have workarounds. But yeah, Nvidia is probably better if LLMs are important for you.

    The 4060 Ti 16 GB version… that sounds good. About $500?

    More like ~$800-900 unless you’re okay with buying used. The market is pretty darn bad, and it’s gotten SO much worse due to the tariff scare. Like I said, you’re better off waiting a few months if you can.


  • Not 100% sure about the chassis and compatibility with the rest of your rig so hopefully someone follows up on me, but in general I would try getting the AMD 9070 XT if you can find it near its real price ($600). That might be a bit difficult since it just released, so I would wait until it starts being more available if you can.

    If not that, I’d recommend the 4060 Ti (16GB). Note that there’s an 8GB version too, do NOT buy that one. Nvidia is generally better when it comes to running AI, but the new AMD gpus are pretty good too.









  • in developing our reasoning models, we’ve optimized somewhat less for math and computer science competition problems, and instead shifted focus towards real-world tasks that better reflect how businesses actually use LLMs.

    I was just about to say how useless these benchmarks are. Plenty of LLMs claim to be better than Claude and GPT4, but in real world use they’ve always been more reliable. Claude especially. Good to hear they’re not just chasing scores.





  • simple@lemm.eetoOpen Source@lemmy.mlProton's biased article on Deepseek
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    1
    ·
    edit-2
    1 month ago

    I understand it well. It’s still relevant to mention that you can run the distilled models on consumer hardware if you really care about privacy. 8GB+ VRAM isn’t crazy, especially if you have a ton of unified memory on macbooks or some Windows laptops releasing this year that have 64+GB unified memory. There are also websites re-hosting various versions of Deepseek like Huggingface hosting the 32B model which is good enough for most people.

    Instead, the article is written like there is literally no way to use Deepseek privately, which is literally wrong.


  • simple@lemm.eetoOpen Source@lemmy.mlProton's biased article on Deepseek
    link
    fedilink
    English
    arrow-up
    123
    arrow-down
    4
    ·
    1 month ago

    DeepSeek is open source, meaning you can modify code(new window) on your own app to create an independent — and more secure — version. This has led some to hope that a more privacy-friendly version of DeepSeek could be developed. However, using DeepSeek in its current form — as it exists today, hosted in China — comes with serious risks for anyone concerned about their most sensitive, private information.

    Any model trained or operated on DeepSeek’s servers is still subject to Chinese data laws, meaning that the Chinese government can demand access at any time.

    What??? Whoever wrote this sounds like he has 0 understanding of how it works. There is no “more privacy-friendly version” that could be developed, the models are already out and you can run the entire model 100% locally. That’s as privacy-friendly as it gets.

    “Any model trained or operated on DeepSeek’s servers are still subject to Chinese data laws”

    Operated, yes. Trained, no. The model is MIT licensed, China has nothing on you when you run it yourself. I expect better from a company whose whole business is on privacy.






  • There’s the samsung tab S9 but it’s expensive and not exactly what you’re looking for. It’s 11", has a high resolution 120Hz AMOLED screen, and I can vouch that it’s really premium. You can put on a matte screen protector if you hate reflections.

    The problem is that a 90Hz+ oled screen is almost exclusive to high-end tablets, so it’s best if you find one used or refurbished if you’re just going to use it for reading.