Follow
Tatsuya Hiraoka
Tatsuya Hiraoka
Mohamed bin Zayed University of Artificial Intelligence
Verified email at mbzuai.ac.ae - Homepage
Title
Cited by
Cited by
Year
Stochastic tokenization with a language model for neural text classification
T Hiraoka, H Shindo, Y Matsumoto
Proceedings of the 57th Annual Meeting of the Association for Computational …, 2019
292019
Named entity recognition and relation extraction using enhanced table filling by contextualized representations
Y Ma, T Hiraoka, N Okazaki
Journal of Natural Language Processing 29 (1), 187-223, 2022
282022
Joint entity and relation extraction based on table labeling using convolutional neural networks
Y Ma, T Hiraoka, N Okazaki
Proceedings of the sixth workshop on structured prediction for NLP, 11-21, 2022
172022
Joint optimization of tokenization and downstream model
T Hiraoka, S Takase, K Uchiumi, A Keyaki, N Okazaki
arXiv preprint arXiv:2105.12410, 2021
172021
Optimizing word segmentation for downstream task
T Hiraoka, S Takase, K Uchiumi, A Keyaki, N Okazaki
Findings of the Association for Computational Linguistics: EMNLP 2020, 1341-1351, 2020
162020
MaxMatch-Dropout: subword regularization for wordpiece
T Hiraoka
arXiv preprint arXiv:2209.04126, 2022
52022
Recurrent neural hidden Markov model for high-order transition
T Hiraoka, S Takase, K Uchiumi, A Keyaki, N Okazaki
Transactions on Asian and Low-Resource Language Information Processing 21 (2 …, 2021
52021
Word-level perturbation considering word length and compositional subwords
T Hiraoka, S Takase, K Uchiumi, A Keyaki, N Okazaki
Findings of the Association for Computational Linguistics: ACL 2022, 3268-3275, 2022
32022
テキストベクトルの重みづけを用いた タスクに対する単語分割の最適化
平岡達也, 高瀬翔, 内海慶, 欅惇志, 岡崎直観
自然言語処理 28 (2), 479-507, 2021
32021
RNN により高次の依存を考慮したニューラル隠れマルコフモデル
平岡達也, 高瀬翔, 内海慶, 欅惇志, 岡崎直観
言語処理学会第 26 回年次大会発表論文集 3, 16-19, 2020
32020
Single Model Ensemble for Subword Regularized Models in Low-Resource Machine Translation
S Takase, T Hiraoka, N Okazaki
arXiv preprint arXiv:2203.13528, 2022
22022
単語分割と後段モデルの損失値を用いた同時最適化
平岡達也, 高瀬翔, 内海慶, 欅惇志, 岡崎直観
自然言語処理 29 (1), 112-143, 2022
22022
Tokenization Tractability for Human and Machine Learning Model: An Annotation Study.
T Hiraoka, T Iwakura
arXiv preprint arXiv:2304.10813, 2023
12023
単語分割の最適化に関する研究は雑談と偶然の出会いに育まれた
平岡達也
自然言語処理 29 (2), 688-693, 2022
12022
LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs
A Aizawa, E Aramaki, B Chen, F Cheng, H Deguchi, R Enomoto, K Fujii, ...
arXiv preprint arXiv:2407.03963, 2024
2024
An Analysis of BPE Vocabulary Trimming in Neural Machine Translation
M Cognetta, T Hiraoka, N Okazaki, R Sennrich, Y Pinter
arXiv preprint arXiv:2404.00397, 2024
2024
Constructing Multilingual Visual-Text Datasets Revealing Visual Multilingual Ability of Vision Language Models
J Atuhurra, I Ali, T Hiraoka, H Kamigaito, T Iwakura, T Watanabe
arXiv preprint arXiv:2406.15359, 2024
2024
Knowledge of Pretrained Language Models on Surface Information of Tokens
T Hiraoka, N Okazaki
arXiv preprint arXiv:2402.09808, 2024
2024
Vocabulary Replacement in SentencePiece for Domain Adaptation
T Kajiura, S Takano, T Hiraoka, K Kuramitsu
Proceedings of the 37th Pacific Asia Conference on Language, Information and …, 2023
2023
Tokenization Preference for Human and Machine Learning Model: An Annotation Study
T Hiraoka, T Iwakura
arXiv preprint arXiv:2304.10813, 2023
2023
The system can't perform the operation now. Try again later.
Articles 1–20