Video Question Answering (Video QA) is a task to answer a text-format question based on the understanding of linguistic semantics, visual information, and also linguisticvisual alignment in the video. In Video QA, an object detector pre-trained with large-scale datasets, such as Faster RCNN, has been widely used to extract visual representations from video frames. However, it is not always able to precisely detect the objects needed to answer the question because
of the domain gaps between the datasets for training the object detector and those for Video QA. In this paper, we
propose a text-guided object detector (TGOD), which takes text question-answer pairs and video frames as inputs, detects the objects relevant to the given text, and thus provides intuitive visualization and interpretable results. Our experiments using the STAGE framework on the TVQA+ dataset show the effectiveness of our proposed detector. It achieves a 2.02 points improvement in accuracy of QA, 12.13 points improvement in object detection (mAP50), 1.1 points improvement in temporal location, and 2.52 points improvement in ASA over the STAGE original detector.