Yeah, I think the main problem with active learning is that H is either hard to split evenly (with individual x points) or is not very general. If we somehow created a nice H that is both sufficiently general and easy to split, then we might get exchanges like this (but starting with more than 4 hypotheses, of course):
(AI's current hypotheses: cancer is good, cancer is bad, QALYs are good, QALYs are bad)
AI: Is removing a tumor good?
Programmer: Yes.
(AI's current hypotheses: cancer is bad, QALYs are good)
AI: Is killing everyone good?
Programmer: No.
(AI's only hypothesis: QALYs are good)
Obviously this works very badly if H does not contain the hypothesis we really want! Luckily, as long as H does contain the correct hypothesis, and we don’t accidentally falsify it, then the system will either determine the correct hypothesis or fail gracefully by reporting that it is uncertain.
I think you might be able to use concept learning to extract humans’ native ontology (of the type studied in the ontological crisis paper) and values expressed in this ontology. The next step is to make a more rational version of this ontology (e.g. by mapping it to the AI’s ontology), which does not look like a concept learning problem.
Yeah, I think the main problem with active learning is that H is either hard to split evenly (with individual x points) or is not very general. If we somehow created a nice H that is both sufficiently general and easy to split, then we might get exchanges like this (but starting with more than 4 hypotheses, of course):
Obviously this works very badly if H does not contain the hypothesis we really want! Luckily, as long as H does contain the correct hypothesis, and we don’t accidentally falsify it, then the system will either determine the correct hypothesis or fail gracefully by reporting that it is uncertain.
“killing everyone” seems a very high level and ambiguous concept.
Certainly. This is why any use of concept learning gets into ontology identification issues.
Can concept learning help effectively at that level?
I think you might be able to use concept learning to extract humans’ native ontology (of the type studied in the ontological crisis paper) and values expressed in this ontology. The next step is to make a more rational version of this ontology (e.g. by mapping it to the AI’s ontology), which does not look like a concept learning problem.