フォロー
Andrew Shin
Andrew Shin
Sony Corporation
確認したメール アドレス: mi.t.u-tokyo.ac.jp - ホームページ
タイトル
引用先
引用先
Dualnet: Domain-invariant network for visual question answering
K Saito, A Shin, Y Ushiku, T Harada
2017 IEEE International Conference on Multimedia and Expo (ICME), 829-834, 2017
612017
Beyond caption to narrative: Video captioning with multiple sentences
A Shin, K Ohnishi, T Harada
2016 IEEE International Conference on Image Processing (ICIP), 3364-3368, 2016
302016
Image Captioning with Sentiment Terms via Weakly-Supervised Sentiment Dataset.
A Shin, Y Ushiku, T Harada
BMVC, 2016
162016
Melody generation for pop music via word representation of musical properties
A Shin, L Crestel, H Kato, K Saito, K Ohnishi, M Yamaguchi, M Nakawaki, ...
arXiv preprint arXiv:1710.11549, 2017
132017
The color of the cat is gray: 1 million full-sentences visual question answering (fsvqa)
A Shin, Y Ushiku, T Harada
arXiv preprint arXiv:1609.06657, 2016
82016
Dense image representation with spatial pyramid vlad coding of cnn for locally robust captioning
A Shin, M Yamaguchi, K Ohnishi, T Harada
arXiv preprint arXiv:1603.09046, 2016
72016
Customized image narrative generation via interactive visual question generation and answering
A Shin, Y Ushiku, T Harada
Proceedings of the IEEE Conference on Computer Vision and Pattern …, 2018
62018
True-negative label selection for large-scale multi-label learning
A Kanehira, A Shin, T Harada
2016 23rd International Conference on Pattern Recognition (ICPR), 3673-3678, 2016
42016
Context-Dependent Automatic Response Generation Using Statistical Machine Translation Techniques
A Shin, R Sasano, T Hiroya, M Okumura
NAACL-HLT 2015, 1345-1350, 2015
42015
Perspectives and prospects on transformer architecture for cross-modal tasks with language and vision
A Shin, M Ishii, T Narihira
International Journal of Computer Vision, 1-20, 2022
32022
Neural Network Libraries: A Deep Learning Framework Designed from Engineers' Perspectives
T Narihira, J Alonsogarcia, F Cardinaux, A Hayakawa, M Ishii, K Iwaki, ...
arXiv preprint arXiv:2102.06725, 2021
22021
Transformer-Exclusive Cross-Modal Representation for Vision and Language
A Shin, T Narihira
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 …, 2021
2021
Reference-Based Video Colorization with Spatiotemporal Correspondence
N Akimoto, A Hayakawa, A Shin, T Narihira
arXiv preprint arXiv:2011.12528, 2020
2020
現在システムで処理を実行できません。しばらくしてからもう一度お試しください。
論文 1–13