AGI is software with human like intelligence that can self-improve and the last two years is very measurable progress. We went from essentially nothing to multimodal models that can run on consumer hardware. OpenAIs new model, if they are to believed, is apparently much better at reasoning and can do long term research. I’ve also seen a few papers in the wild talking about self teaching methods and framework.
To be clear, I don’t know which will come first. It’s hard to know if the next leap is just a step or if there’s a giant chasm laying infront. I do know that it’s a lot easier to prototype with AI then fusion and there’s a lot more people working on it, both behind closed doors and publically on the internet. Fusion doesn’t have this advantage.
Your statement is basically a shot in the dark imo.
AGI is software with human like intelligence that can self-improve and the last two years is very measurable progress. We went from essentially nothing to multimodal models that can run on consumer hardware. OpenAIs new model, if they are to believed, is apparently much better at reasoning and can do long term research. I’ve also seen a few papers in the wild talking about self teaching methods and framework.
To be clear, I don’t know which will come first. It’s hard to know if the next leap is just a step or if there’s a giant chasm laying infront. I do know that it’s a lot easier to prototype with AI then fusion and there’s a lot more people working on it, both behind closed doors and publically on the internet. Fusion doesn’t have this advantage.
Your statement is basically a shot in the dark imo.