Task coordination for autonomous mobile service robots typically involves a substantial amount of background knowledge and explicit action sequences to acquire the relevant information nowadays. We strive for a system which, given a task, is capable of reasoning about task-relevant knowledge to automatically determine whether that knowledge is sufficient. If missing or uncertain, the robot shall decide autonomously on the actions to gain or improve that knowledge. In this paper we present our baseline system implementing the foundations for these capabilities. The robot has to analyze a tabletop scene and increase its object type confidence. It plans motions to observe the scene from multiple perspectives, combines the acquired data, and performs a recognition step on the merged input.