He should build a sub to go to the bottom of the ocean.
And take Elon Musk with him
they’ll never do that, but maybe a spaceship shaped like a giant penis could do the trick
Give it exactly 420.69 lb of rocket fuel and fire it into the sun
If they hesitate, we just tell them there are neither poor people, nor taxes in space 😜
And skimp on materials, build it out of carbon fiber instead of titanium or steel.
Sure, we could point to thousands of years of really smart people trying and utterly failing to build mathematical models for innovation and thought, but it also does make a certain amount of sense that, if you pile up enough transistors and wish really hard, that your investment will Frosty the Snowman itself into being your friend, right…?
I wonder if he really thinks that AGI is just AI with more parameters and gpus thrown into the mix.
Zuckerberg: Why are my pupils vertical slits? Why am I always cold? Why do people find me so repellent?
AI: Sir, I can answer all three with one response, but you won’t like it.
Iirc the h100s are $30k per gpu at this time.
Retail price of $10,500,000,000. That’s nuts.
I’m sure he’d get a hefty discount on 350k of them
Still less than Musk paid for Twitter. It’s totally reasonable from a biggest-billionaire-toy entry.
Building a metaverse that people want to actually engage with was too hard, so he’s decided to scale back his ambitions and tackle something less difficult: AGI.
He just want some virtual friends to hang out with in the metaverse, since humans weren’t interested.
Is Zuckerberg an idiot? Or does he have an actual plan with this?
Seems to me it’s completely useless like Metaverse.
If the LLM is so stupid it can’t figure out the sides of an equal sign can be reversed as simple as in 2+2=4 <=> 4=2+2. He will never achieve general intelligence by just throwing more compute power at it.
As powerful as LLM is, it’s still astoundingly stupid when it hits its limitations.Humans are astoundigly stupid when they hit their limitations.
The difference is that we can go beyond that limitation. Even self-coding AI will either solve a problem, or compound its own inefficiencies before asking an operator to help out.
Your post sounds almost as dense as:
“everything that can be invented has been invented.” - Duell 1899.
I don’t know much, but from what I know, we still haven’t reach a point of diminishing returns, so more power = more better.
There is a lot of theoretical work on this problem, but I’m in the camp that isn’t convinced large language models are the path towards general intelligence.
Throw 10x the computing power on it and it might learn that a maths equation is reversible, because it will probably have seen enough examples of that. But it won’t learn what an equation represents, and therefore won’t extrapolate situations that can be solved by equations.
You can already ask ChatGPT to model a real life scenario with a simple math equation. There is at least a rough model of how basic math can be used to solve problems.