The thing about training an AI is that it is complex and unpredictable. This isn't the same as programming software, it is self learning and that adds levels of complexity. With AI you give it a problem to solve, you give it parameters and after that you are probably out of the loop.
For example you will tell your drone to kill a target. If this is time sensitive or it is being jammed then calling home for final approval might not be an option. In a simular situation a human may have to use their own initiative, and to be honest, if it is going to be effective, you will often have to give your drone that same capability. This is where things get complex because you are essentually asking your machine to make a judgement call. Will it risk killing civillians to achieve its mission, will it ignore calls to abort a mission if it has doubts about the authenticity of that order and so on.
In the case of a simulation I would say a big part of it is actually testing to see how an AI would react in various scenarios. To be honest a problem solving AI deciding that taking out a communication tower to prevent a ground controller from interfering with a priority mission does have a certain logic about it.