Do you have a dumb question that you're kind of embarrassed to ask in the main thread? Is there something you're just not sure about?
This is your opportunity to ask questions. No question too simple or too silly.
Culture war topics are accepted, and proposals for a better intro post are appreciated.
Jump in the discussion.
No email address required.
Notes -
I struggle to think of any AI architecture that works the way you envision, using fractional ratios of reward to available room for reward instead of plain absolute magnitude of reward. I could be wrong, but I still doubt that's ever done.
It's impossible to answer that without digging into the exact specifications of the AI in question, and what tie-breaker mechanism it has to adjudicate between options when all of them have the same (zero) reward. Maybe it picks the first option, maybe it chooses randomly.
However, I am under the impression that in the majority of cases, a reward maximizing agent will simply try to minimize the risk of losing its accrued reward if it's maxed out, which will likely result in large scale behavior indistinguishable from attempting to increase the reward itself (turning the universe into computronium).
Why could you not measure the fitness? Even if we can't evaluate each decision chain in chess, we know how many there are, so a reward that increases linearly for each tree solved should work.
How does it follow that it's a fractional ratio? The only relevant fact is whether the maximum value has been reached. How could it even compare the absolute magnitude, if it can't store a larger number?
I agree with this, but based on my knowledge of speculative ways to survive until the end of the Universe, few involve turning it into computronium. Presumably, AI would still factor in risk.
I mean that, in practice, it could never be realised, for the reasons you mentioned- as in, achievement beyond a certain value would be impossible, since you can't strongly solve chess within current physical limits.
More options
Context Copy link
More options
Context Copy link