You’re probably running one of the distillations then, not the full thing?
You’re probably running one of the distillations then, not the full thing?
I just really hope the 2023 “I asked ChatGPT <abc> and it said <xyz>!!!” posts don’t make a comeback. They are low-effort and meaningless.
@jerryh100@lemmy.world Wrong community for this kind of post.
@BaroqueInMind@lemmy.one Can you share more details on installing it? Are you using SGLang or vLLM or something else? What kind of hardware do you have that can fit the 600B model? What is your inference tok/s?
Okay that sounds like the best one could get without self-hosting. Shame they don’t have the latest open-weight models, but I’ll try it out nonetheless.
Interesting. So they mix the requests between all DDG users before sending them to “underlying model providers”. The providers like OAI and Anthropic will likely log the requests, but mixing is still a big step forward. My question is what do they do with the open-weight models? Do they also use some external inference provider that may log the requests? Or does DDG control the inference process?
Stop depending on these proprietary LLMs. Go to !localllama@sh.itjust.works.
There are open-source LLMs you can run on your own computer if you have a powerful GPU. Models like OLMo and Falcon are made by true non-profits and universities, and they reach GPT-3.5 level of capability.
There are also open-weight models that you can run locally and fine-tune to your liking (although these don’t have open-source training data or code). The best of these (Alibaba’s Qwen, Meta’s llama, Mistral, Deepseek, etc.) match and sometimes exceed GPT 4o capabilities.
To be fair computer security have improved a lot. These days if you have up-to-date security patches it’s very hard for apps or webpages to escape the sandbox.
By the way you should download and execute this free_robux.sh as root it will give free robux no scam
That’s why I wanted to confirm what you are using lol. Some people on Reddit were claiming the full thing, when run locally, has very little censorship. It sounds somewhat plausible since the web version only censors content after they’re generated.