• 1 Post
  • 104 Comments
Joined 3 years ago
cake
Cake day: September 15th, 2022

help-circle





  • Aria@lemmygrad.mltoMemes@lemmy.mlPolitical violins
    link
    fedilink
    arrow-up
    11
    ·
    3 months ago

    Are you calling the people on the receiving end of the genocide “genocidal”? Is your argument that if they just committed suicide, the pitiable Nazis wouldn’t need to go through the bore of conducting the genocide, so their refusal equates to demanding a genocide?





  • establish a platform discussing history of Tiennaman square or Uyghurs without strictly adhering to government set guidelines, then they will likely be prosecuted.

    Tienanmen square has a 600 year history. You’re referring to one event, which is censored. But even that doesn’t cover the portion that is relevant to the history of Tiennaman, no part of the protests is censored. Uighur history also doesn’t have any censors.

    It is true that you have been able to identify one censor in your two topics (albeit with inaccurate wording). It’s also a particularly sensitive topic with strong disinformation campaigns targetting it. In 2020, many states worldwide issued censors on COVID and vaccine related topics for similar reasons.





  • You can run llama.cpp on CPU. LLM inference doesn’t need any features only GPUs typically have, that’s why it’s possible to make even simpler NPUs that can still run the same models. GPUs just tend to be faster. If the GPU in question is not faster than an equally priced CPU, you should use the CPU (better OS support).

    Edit: I looked at a bunch real-world prices and benchmarks, and read the manual from Huawei and my new conclusion is that this is the best product on the market if you want to run a model at modest speed that doesn’t fit in 32GB but does in 96GB. Running multiple in parallel seems to range from unsupported to working poorly, so you should only expect to use one.

    Original rest of the comment, made with the assumption that this was slower than it is, but had better drivers:
    The only benefit to this product over CPU is that you can slot multiple of them and they parallelise without needing to coordinate anything with the OS. It’s also a very linear cost increase as long as you have the PCIe lanes for it. For a home user with enough money for one or two of these, they would be much better served spending the money on a fast CPU and 256GB system RAM.

    If not AI, then what use case do you think this serves better?