Search Results for author: Taiki Miyanishi

Found 6 papers, 4 papers with code

Map-based Modular Approach for Zero-shot Embodied Question Answering

no code implementations26 May 2024 Koya Sakamoto, Daichi Azuma, Taiki Miyanishi, Shuhei Kurita, Motoaki Kawanabe

We conduct comprehensive experiments on virtual environments (MP3D-EQA) and two real-world house environments and demonstrate that our method can perform EQA even in the real world.

Embodied Question Answering Navigate +1

JDocQA: Japanese Document Question Answering Dataset for Generative Language Models

1 code implementation28 Mar 2024 Eri Onami, Shuhei Kurita, Taiki Miyanishi, Taro Watanabe

Document question answering is a task of question answering on given documents such as reports, slides, pamphlets, and websites, and it is a truly demanding task as paper and electronic forms of documents are so common in our society.

Hallucination Question Answering +1

Vision Language Model-based Caption Evaluation Method Leveraging Visual Context Extraction

no code implementations28 Feb 2024 Koki Maeda, Shuhei Kurita, Taiki Miyanishi, Naoaki Okazaki

Given the accelerating progress of vision and language modeling, accurate evaluation of machine-generated image captions remains critical.

Image Captioning Language Modelling

Cross3DVG: Cross-Dataset 3D Visual Grounding on Different RGB-D Scans

1 code implementation23 May 2023 Taiki Miyanishi, Daichi Azuma, Shuhei Kurita, Motoki Kawanabe

We present a novel task for cross-dataset visual grounding in 3D scenes (Cross3DVG), which overcomes limitations of existing 3D visual grounding models, specifically their restricted 3D resources and consequent tendencies of overfitting a specific 3D dataset.

3D Reconstruction 3D visual grounding

Cannot find the paper you are looking for? You can Submit a new open access paper.