So I signed up for a free month of their crap because I wanted to test if it solves novel variants of the river crossing puzzle.
Like this one:
You have a duck, a carrot, and a potato. You want to transport them across the river using a boat that can take yourself and up to 2 other items. If the duck is left unsupervised, it will run away.
Unsurprisingly, it does not:
https://g.co/gemini/share/a79dc80c5c6c
https://g.co/gemini/share/59b024d0908b
The only 2 new things seem to be that old variants are no longer novel, and that it is no longer limited to producing incorrect solutions - now it can also incorrectly claim that the solution is impossible.
I think chain of thought / reasoning is a fundamentally dishonest technology. At the end of the day, just like older LLMs it requires that someone solved a similar problem (either online or perhaps in a problem solution pair they generated if they do that to augment the training data).
But it outputs quasi reasoning to pretend that it is actually solving the problem live.
I thought everyone knew that you had to structure prompts in ALGOL 420 to get the best performance by going close to the metal
I use UTF-9 to efficiently handle Unicode on my PDP-10.
@bitofhope @techtakes Surely you need a PDP-9 for that?