7/11/25 AI thread
| genuinely altruistic poaster | 07/11/25 | | scholarship | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | scholarship | 07/11/25 | | .,,,.,.,.,.,.,,,,..,,..,.,.,., | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | cock of michael obama | 07/11/25 | | .,,,.,.,.,.,.,,,,..,,..,.,.,., | 07/11/25 | | .,.,.,.,.,.,.,.,...,,..,.,., | 07/11/25 | | Business school fucking ROCKS!!! | 07/11/25 | | cock of michael obama | 07/11/25 | | '"'"'"''" | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | '"'"'"''" | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | ai addict | 07/11/25 | | scholarship | 07/11/25 | | Live Caged and Cry | 07/11/25 | | .,,,.,.,.,.,.,,,,..,,..,.,.,., | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | .,,,.,.,.,.,.,,,,..,,..,.,.,., | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | Wang Hernandez | 07/11/25 | | genuinely altruistic poaster | 07/11/25 | | derek pajeeter | 07/11/25 | | """'"'"""'' | 07/11/25 | | Live Caged and Cry | 07/12/25 | | Dave Prole | 07/12/25 | | ,.,....,...,,,..,..,.,..,.,.,.,. | 07/12/25 | | genuinely altruistic poaster | 07/12/25 |
Poast new message in this thread
Date: July 11th, 2025 1:36 PM Author: genuinely altruistic poaster
scholarship wants me to do these daily again
https://x.com/elder_plinius/status/1943171871400194231
system prompt for new grok 4. it has the same line about being allowed to say politically incorrect things, which supports my suspicion that the will stancil-raping version of grok the other day had more tweaks than just this
it appears to now be the strongest AI currently available. they spent a huge amount of compute resources on post-pretraining RL compared to all the other models. this is probably why it's performing so well on reasoning and problem-solving benchmark tests, because RL training helps a lot with this
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49093385) |
Date: July 11th, 2025 2:54 PM
Author: .,,,.,.,.,.,.,,,,..,,..,.,.,.,
i think i am going to stick to Opus, o3 and 2.5 pro. the differences look pretty marginal and i have zero faith xAI won't use my data for training.
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49093634) |
 |
Date: July 11th, 2025 3:01 PM
Author: .,,,.,.,.,.,.,,,,..,,..,.,.,.,
Google does if you use aistudio but i buy API credits. Same with Anthropic. i highly doubt they are lying about not training on API users if you opt out. i have less faith in OpenAI but i'm also not too concerned.
an Elon company? no way am i trusting them unless the model is way better.
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49093657) |
Date: July 11th, 2025 3:02 PM
Author: .,.,.,.,.,.,.,.,...,,..,.,., ( )
I trained a local llm on XO and it called me a fag
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49093659) |
Date: July 11th, 2025 5:44 PM Author: genuinely altruistic poaster
https://x.com/keyonV/status/1943730495264584079
https://arxiv.org/pdf/2507.06952
paper that demonstrates that while LLMs excel at predictive tasks that fall within their training data, they can't generalize that predictive ability into a complete and accurate world model to make correct predictions on tasks that weren't within their training
lecun is right imo. statistical inferences do not lead to the ability to make generalized inferences
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49094177) |
 |
Date: July 11th, 2025 6:48 PM
Author: .,,,.,.,.,.,.,,,,..,,..,.,.,.,
this is why i think the reinforcement learning approach of trying to paper over these problems with more data is wrongheaded. they are getting insufficient generalization from 30 trillion token datasets so they think they just need to use RL and chain of thought to make 300 trillion token datasets or whatever and their problems will be solved. the architectures and training methods should be fixed first. there's a lot of research showing inadequate generalization on toy tasks even with lots of compute.
https://arxiv.org/abs/2207.02098
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49094428) |
 |
Date: July 11th, 2025 7:05 PM Author: genuinely altruistic poaster
yeah, i've become convinced that LLMs just fundamentally cannot out-of-distribution generalize. and i don't see how RL could ever solve it. all you're doing is training it on additional specific tasks that it....still can't generalize out of
i think companies are doing RL now as more of a marketing gimmick than anything else. people are saying that the latest grok 4 is only performing so well on benchmarks because its RL training had overlap with the benchmark tests. this is what i mean when i say that they're just gaming the benchmark tests
i think the next few years might end up being the big base models trained with RL into a bunch of different specialized sub-models that are used in larger multi-agent architectures in order to be more useful in practice in different specific contexts
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49094487) |
 |
Date: July 11th, 2025 7:18 PM
Author: .,,,.,.,.,.,.,,,,..,,..,.,.,.,
right. i view RL as the way to get strong superhuman agents once you have the right base learning model. they should be able to use the existing data to train robust human level agents, but that's apparently not happening because the learning algorithm is inadequate.
that paper i linked seems to imply there are already models that consistently generalize better than transformers, so it's not clear we see companies flailing about for years using RL gimmicks with transformers. memory augmented transformers could be a near-term replacement for current models. if models don't improve substantially over the next year, there will be a strong motivation to try novel approaches like this.
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49094525) |
 |
Date: July 11th, 2025 7:34 PM Author: genuinely altruistic poaster
https://itcanthink.substack.com/p/what-are-robot-world-models
really good short article explaining how people think that we'll be able to use generative AI video to train robots to have functional world models
the notion of this actually working seems crazy to me but these guys are all very smart so it must be somewhat viable or someone would be calling it out as BS. it would cut robot training time and costs IMMENSELY and make them a lot more commercially viable and speed up robotics development by years
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49094551) |
Date: July 12th, 2025 1:33 AM Author: Live Caged and Cry
OpenAI got cucked again. They tried to acquire Windsurf but didn’t want Microsoft to have access to its IP. Now the acquisition is off and Google has hired their CEO.
https://x.com/ns123abc/status/1943806065524507007?s=46&t=YKr-jZOYUHE15Tew69wt4w
The Microsoft partnership was key to getting them off the ground, but now it’s an adversarial relationship.
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49095182) |
 |
Date: July 12th, 2025 3:08 AM
Author: ,.,....,...,,,..,..,.,..,.,.,.,.
Microsoft should put them out of business. The decision to partner with them rather than hiring away their engineers and scaling themselves was strange.
(http://www.autoadmit.com/thread.php?thread_id=5749144&forum_id=2#49095224) |
|
|