What's New

Prof. Kurohashi was selected as ACL Fellows 2025.

We will present the following paper at AAAI 2026 (2026/1).

  • Yang Liu, Masahiro Kaneko, Chenhui Chu:
    On the Alignment of Large Language Models with Global Human Opinion

The following presentations and roles are scheduled for AACL 2025 (2025/12)

  • Professor Kurohashi will deliver the keynote address.
  • Shimizu, a doctoral student, will chair the Student Research Workshop.
  • We will present the following paper.
    • Kunal Kingkar Das, Manoj Balaji Jagadeeshan, Nallani Chakravartula Sahith, Jivnesh Sandhan, and Pawan Goyal: Still Not There: Can LLMs Outperform Smaller Task-Specific Seq2Seq Models on the Poetry-to-Prose Conversion Task?
    • Wan Jou She, Lis Pereira, Fei Cheng, Sakiko Yahata, Panote Siriaraya, Eiji Aramaki. EmplifAI: a Fine-grained Dataset for Japanese Empathetic Medical Dialogues in 28 Emotion Labels

We will present the following papers at EMNLP 2025 (2025/11).

  • Zhengdong Yang, Zhen Wan, Sheng Li, Chao-Han Huck Yang, Chenhui Chu:
    CoVoGER: A Multilingual Multitask Benchmark for Speech-to-text Generative Error Correction with Large Language Models
  • Yoshiki Takenami, Yin Jou Huang, Yugo Murawaki, Chenhui Chu:
    How Does Cognitive Bias Affect Large Language Models? A Case Study on the Anchoring Effect in Price Negotiation Simulations (Findings)
  • Yang Liu, Chenhui Chu:
    Do LLMs Align Human Values Regarding Social Biases? Judging and Explaining Social Biases with LLMs (Findings)
  • Jivnesh Sandhan, Fei Cheng, Tushar Sandhan, and Yugo Murawaki:
    From Disney-World to Reality: A Context-Dependent Testbed for Personality Assessment of Large Language Models (Findings)
  • Ruiyi Yan and Yugo Murawaki:
    Addressing Tokenization Inconsistency in Steganography and Watermarking Based on Large Language Models.
  • Yin Jou Huang, Rafik Hadfi:
    Beyond Self-Reports: Multi-Observer Agents for Personality Assessment in Large Language Models (Findings)
  • Sakiko Yahata, Zhen Wan, Fei Cheng, Sadao Kurohashi, Hisahiko Sato, Ryozo Nagai: Causal Tree Extraction from Medical Case Reports:
    A Novel Task for Experts-like Text Comprehension
  • Kazuma Kobayashi, Zhen Wan, Fei Cheng, Yuma Tsuta, Xin Zhao, Junfeng Jiang, Jiahao Huang, Zhiyi Huang, Yusuke Oda, Rio Yokota, Yuki Arase, Daisuke Kawahara, Akiko Aizawa, Sadao Kurohashi:
    Leveraging High-Resource English Corpora for Cross-lingual Domain Adaptation in Low-Resource Japanese Medicine via Continued Pre-training (Findings)

Research Overview

Language is the most reliable medium of human intellectual activities. Our objective is to establish the technology and academic discipline for handling and understanding language, in a manner that is as close as possible to that of humans, using computers. These include syntactic language analysis, semantic analysis, context analysis, text comprehension, text generation and dictionary systems to develop various application systems for machine translation and information retrieval.

Search Engine Infrastructure based on Deep Natural Language Processing

TSUBAKI.png

The essential purpose of information retrieval is not to retrieve just a relevant document but to acquire the information or knowledge in the document. We have been developing a next-generation infrastructure of information retrieval on the basis of the following techniques of deep natural language processing: precise processing based not on words but on predicate-argument structures, identifying the variety of linguistic expressions and providing a bird's-eye view of search results via clustering and interaction.

Machine Translation

EBMT.png

To bring automatic translation by computers to the level of human translation, we have been studying next-generation methodology of machine translation on the basis of text understanding and a large collection of translation examples. We have already accomplished practical translation on the domain of travel conversation, and constructed a translation-aid system that can be used by experts of patent translation.

Fundamental Studies on Text Understanding

To make computers understand language, it is essential to give computers world knowledge. This was a very hard problem ten years ago, but it has become possible to acquire knowledge from a massive amount of text in virtue of the drastic progress of computing power and network. We have successfully acquired linguistic patterns of predicate-argument structures from automatic parses of 7 billion Japanese sentences crawled from the Web using grid computing machines. By utilizing such knowledge, we study text understanding, i.e., recognizing the relationships between words and phrases in text.

Policy Regarding Acceptance of Students from Outside

Prof. Kurohashi does not supervise new students. Assoc. Prof. Murawaki or Assoc. Prof. Chu will be responsible for supervising their research.

Master course

PhD course

Access


Attach file: filedaimonzi_20241127.jpg 818 download [Information] filetrip_20230913.jpg 673 download [Information] filemanshuin_20211117.jpg 405 download [Information] filecovid-19_banner.en.png 249 download [Information] fileTSUBAKI.png 1628 download [Information] fileEBMT.png 1571 download [Information]

Front page   Edit Freeze Diff Backup Attach Copy Rename Reload   New List of pages Search Recent changes   Help   RSS of recent changes