How can you (in general) conclude something by examining the source code of an agent, without potentially implicating the Halting Problem?
Nothing stops the Halting problem being solved in particular instances. I can prove that some agent halts, and so can it. See FairBot in Robust Cooperation in the Prisoner’s Dilemma.
In this case, we have (by assumption) an output of the program, so we just look at the cases where the program gives that output.
How can you (in general) conclude something by examining the source code of an agent, without potentially implicating the Halting Problem?
Nothing stops the Halting problem being solved in particular instances. I can prove that some agent halts, and so can it. See FairBot in Robust Cooperation in the Prisoner’s Dilemma.
In this case, we have (by assumption) an output of the program, so we just look at the cases where the program gives that output.