- cross-posted to:
- [email protected]
- cross-posted to:
- [email protected]
Are the people who work at OpenAI smoking crack?
“Over the last year and a half there have been a lot of questions around what might happen if influence operations use generative AI,” Ben Nimmo, principal investigator on OpenAI’s Intelligence and Investigations team, told members of the media in a press briefing
Here’s a clue, look around you.
ChatGPT isn’t the only fish in the sea and state actors using a public service like it deserve to be caught. Running your own system privately, without scrutiny, without censorship, without constraints is so trivial that teenagers are doing this on their laptops, so much so that you can
docker pull
your way into any number of LLM images.Seriously, this is so many levels of absurd that it’s beyond comprehension…
Having tried many different models on my machine and being a long-time GPT-4 user, I can say the self-hosted models are far more impressive in sheer power for their size. However, the good ones still require a GPU that most people nor teenagers can’t afford.
Nonetheless, GPT-4 remains the most powerful and useful model, and it’s not even a competition. Even Google’s Gemini doesn’t compare, in my experience.
The potential for misuse increases alongside usefulness and power. I wouldn’t use Ollama or GPT-3.5 for my professional work because they’re just not reliable enough. However, GPT-4, despite also having its useless moments, is almost essential.
The same holds true for scammers and malicious actors. GPT-4’s voice will technically allow live, fluent conversations through a phone using a dynamic voice. That’s the holy grail for scamcallers. OpenAI is right to want to eliminate as much abuse of their system as possible before releasing such a thing.
There is an argument to be made for not releasing such dangerous tools, but the counter is that someone malicious will inevitably release it someday. It’s better to be prepared and understand these systems before that happens. At least i think thats what OpenAi believes, i am not sure what to think. How could i known they Arent malicious?
Saying you wouldn’t use ‘ollama or gpt3.5’ is such a… I want to say uneducated statement? These are not two like terms
You’re aware that ollama isn’t an LLM? You’re aware there are LLMs available via ollama that exceed gpt4s capabilities? I mean, you’re right that you need an array of expensive gpus to run them effectively, but… Just comparing ollama to gpt-3.5 is like comparing an NCAA basketball star to the Harlem globe trotters. It’s ridiculous at its face. A player compared to a team, for starters.
Correct, i kept it simple on purpose and could probably have worded it better.
It was a meant as a broader statement including “both publicly available free to download models like those based on the ollama architectures as well as free to acces proprietary llm’s like gpt3.5”
I personally tried variations of the vicuna, wizardLM and a few other models (mostly 30B, bigger was to slow) which are all based on ollama’s architecture but i consider those individual names to be less known.
Neither of these impressed me all that much. But of course this is a really fast changing industry. Looking at the hf leaderboard i don’t see any of the models i tried. Last time i checked was January.
I may also have an experience bias as i have become much more effective using gpt4 as a tool compared to when i just started to use it. This influences what I expect and how i write prompts for other models.
I’d be happy to try some new models that have since archived new levels. I am huge supporter for self-hosting digital tools and frankly i cant wait to stop funding ClosedAi
Llama3-70b is probably the most general purpose capable open source
There are a bunch of contenders for specific purposes, like coding and stuff, though. I wanna say Mistral has a brand new enormous one that you’d need like 4 4090s to run smoothly.
Ur missing the point the goal is yo ban anyone except the big companies.
I’m still baffled at how good Ollama is on working on paltry hardware like ARM and small VMs. Give it GPUs and it’s amazing.
The next step should be to encrypt information at Transit and rest to as to purchase GPU power from the cloud but maintaining client-side encryption throughout. That’ll bring even more power to the masses: imagine giving Ollama a Cloud endpoint to remote GPUs which it can compute on without the consumer purchasing any hardware.
ARM is not paltry, it’s in small/portable devices because it’s efficient, not weak.
deleted by creator
Tell that to groq.
I guess the ones they stopped just weren’t covert enough.
When you stare into the AI, the AI stares back at you.