Chiori Hori

Chiori Hori
  • Biography

    Chiori has been a member of MERL's research team since 2015. Her work is focused on spoken dialog and audio visual scene-aware dialog technologies toward human-robot communications. She's on the editorial board of "Computer Speech and Language" and is a technical committee member of "Speech and Language Processing Group" of IEEE Signal Processing Society. Prior to joining MERL, Chiori spent 8 years at Japan's National Institute of Information and Communication Technology (NICT), where she held the position of Research Manager of the Spoken Language Communication Laboratory. She also spent time researching at Carnegie Mellon and the NTT Communication Science Laboratories, prior to NICT.

  • Recent News & Events

    •  NEWS    MERL congratulates Prof. Alex Waibel on receiving 2023 IEEE James L. Flanagan Speech and Audio Processing Award
      Date: August 22, 2022
      MERL Contacts: Chiori Hori; Jonathan Le Roux; Anthony Vetro
      Research Areas: Artificial Intelligence, Machine Learning, Speech & Audio
      Brief
      • IEEE has announced that the recipient of the 2023 IEEE James L. Flanagan Speech and Audio Processing Award will be Prof. Alex Waibel (CMU/Karlsruhe Institute of Technology), “For pioneering contributions to spoken language translation and supporting technologies.” Mitsubishi Electric Research Laboratories (MERL), which has become the new sponsor of this prestigious award in 2022, extends our warmest congratulations to Prof. Waibel.

        MERL Senior Principal Research Scientist Dr. Chiori Hori, who worked with Dr. Waibel at Carnegie Mellon University and collaborated with him as part of national projects on speech summarization and translation, comments on his invaluable contributions to the field: “He has contributed not only to the invention of groundbreaking technology in speech and spoken language processing but also to the promotion of an abundance of research projects through international research consortiums by linking American, European, and Asian research communities. Many of his former laboratory members and collaborators are now leading R&D in the AI field.”

        The IEEE Board of Directors established the IEEE James L. Flanagan Speech and Audio Processing Award in 2002 for outstanding contributions to the advancement of speech and/or audio signal processing. This award has recognized the contributions of some of the most renowned pioneers and leaders in their respective fields. MERL is proud to support the recognition of outstanding contributions to the field of speech and audio processing through its sponsorship of this award.
    •  
    •  NEWS    MERL presenting 8 papers at ICASSP 2022
      Date: May 22, 2022 - May 27, 2022
      Where: Singapore
      MERL Contacts: Anoop Cherian; Chiori Hori; Toshiaki Koike-Akino; Jonathan Le Roux; Tim K. Marks; Philip V. Orlik; Kuan-Chuan Peng; Pu (Perry) Wang; Gordon Wichern
      Research Areas: Artificial Intelligence, Computer Vision, Signal Processing, Speech & Audio
      Brief
      • MERL researchers are presenting 8 papers at the IEEE International Conference on Acoustics, Speech & Signal Processing (ICASSP), which is being held in Singapore from May 22-27, 2022. A week of virtual presentations also took place earlier this month.

        Topics to be presented include recent advances in speech recognition, audio processing, scene understanding, computational sensing, and classification.

        ICASSP is the flagship conference of the IEEE Signal Processing Society, and the world's largest and most comprehensive technical conference focused on the research advances and latest technological development in signal and information processing. The event attracts more than 2000 participants each year.
    •  

    See All News & Events for Chiori
  • Research Highlights

  • Internships with Chiori

    • SA1936: Multimodal scene-understanding for Robot Dialog or Indoor Monitoring

      We are looking for a graduate student interested in helping advance the field of multi-modal scene understanding, with a focus on scene understanding using natural language for robot dialog or indoor monitoring. The intern will collaborate with MERL researchers to derive and implement new models and optimization methods, conduct experiments, and prepare results for publication. The ideal candidate would be a senior Ph.D. student with experience in deep learning for audio-visual, signal, and natural language processing. The expected duration of the internship is 3-6 months, and start date is flexible.

    See All Internships at MERL
  • MERL Publications

    •  Hori, C., Hori, T., Le Roux, J., "Low-Latency Streaming Scene-aware Interaction Using Audio-Visual Transformers", Interspeech, DOI: 10.21437/​Interspeech.2022-10891, September 2022, pp. 4511-4515.
      BibTeX TR2022-116 PDF
      • @inproceedings{Hori2022sep,
      • author = {Hori, Chiori and Hori, Takaaki and Le Roux, Jonathan},
      • title = {Low-Latency Streaming Scene-aware Interaction Using Audio-Visual Transformers},
      • booktitle = {Interspeech},
      • year = 2022,
      • pages = {4511--4515},
      • month = sep,
      • doi = {10.21437/Interspeech.2022-10891},
      • url = {https://www.merl.com/publications/TR2022-116}
      • }
    •  Shah, A.P., Geng, S., Gao, P., Cherian, A., Hori, T., Marks, T.K., Le Roux, J., Hori, C., "Audio-Visual Scene-Aware Dialog and Reasoning Using Audio-Visual Transformers with Joint Student-Teacher Learning", IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), April 2022, pp. 7732-7736.
      BibTeX TR2022-019 PDF
      • @inproceedings{Shah2022apr,
      • author = {Shah, Ankit Parag and Geng, Shijie and Gao, Peng and Cherian, Anoop and Hori, Takaaki and Marks, Tim K. and Le Roux, Jonathan and Hori, Chiori},
      • title = {Audio-Visual Scene-Aware Dialog and Reasoning Using Audio-Visual Transformers with Joint Student-Teacher Learning},
      • booktitle = {IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP)},
      • year = 2022,
      • pages = {7732--7736},
      • month = apr,
      • publisher = {IEEE},
      • issn = {1520-6149},
      • isbn = {978-1-6654-0540-9},
      • url = {https://www.merl.com/publications/TR2022-019}
      • }
    •  Hori, C., Vetro, A., "At Last, A Self-Driving Car That Can Explain Itself", IEEE Spectrum, March 2022.
      BibTeX IEEE Spectrum
      • @article{Hori2022mar,
      • author = {Hori, Chiori and Vetro, Anthony},
      • title = {At Last, A Self-Driving Car That Can Explain Itself},
      • journal = {IEEE Spectrum},
      • year = 2022,
      • month = mar,
      • url = {https://spectrum.ieee.org/at-last-a-self-driving-car-that-can-explain-itself}
      • }
    •  Shah, A.P., Hori, T., Le Roux, J., Hori, C., DSTC10-AVSD Submission System with Reasoning using Audio-Visual Transformers with Joint Student-Teacher Learning, February 2022.
      BibTeX TR2022-025 PDF
      • @book{Shah2022feb,
      • author = {Shah, Ankit Parag and Hori, Takaaki and Le Roux, Jonathan and Hori, Chiori},
      • title = {DSTC10-AVSD Submission System with Reasoning using Audio-Visual Transformers with Joint Student-Teacher Learning},
      • year = 2022,
      • month = feb,
      • url = {https://www.merl.com/publications/TR2022-025}
      • }
    •  Hori, C., Shah, A.P., Geng, S., Gao, P., Cherian, A., Hori, T., Le Roux, J., Marks, T.K., "Overview of Audio Visual Scene-Aware Dialog with Reasoning Track for Natural Language Generation in DSTC10", The 10th Dialog System Technology Challenge Workshop at AAAI, February 2022.
      BibTeX TR2022-016 PDF
      • @inproceedings{Hori2022feb,
      • author = {Hori, Chiori and Shah, Ankit Parag and Geng, Shijie and Gao, Peng and Cherian, Anoop and Hori, Takaaki and Le Roux, Jonathan and Marks, Tim K.},
      • title = {Overview of Audio Visual Scene-Aware Dialog with Reasoning Track for Natural Language Generation in DSTC10},
      • booktitle = {The 10th Dialog System Technology Challenge Workshop at AAAI},
      • year = 2022,
      • month = feb,
      • url = {https://www.merl.com/publications/TR2022-016}
      • }
    See All Publications for Chiori
  • Videos

  • MERL Issued Patents

    • Title: "Low-latency Captioning System"
      Inventors: Hori, Chiori; Hori, Takaaki; Cherian, Anoop; Marks, Tim; Le Roux, Jonathan
      Patent No.: 11,445,267
      Issue Date: Sep 13, 2022
    • Title: "System and Method for a Dialogue Response Generation System"
      Inventors: Hori, Chiori; Cherian, Anoop; Marks, Tim; Hori, Takaaki
      Patent No.: 11,264,009
      Issue Date: Mar 1, 2022
    • Title: "Scene-Aware Video Dialog"
      Inventors: Geng, Shijie; Gao, Peng; Cherian, Anoop; Hori, Chiori; Le Roux, Jonathan
      Patent No.: 11,210,523
      Issue Date: Dec 28, 2021
    • Title: "Method and System for Multi-Label Classification"
      Inventors: Hori, Takaaki; Hori, Chiori; Watanabe, Shinji; Hershey, John R.; Harsham, Bret A.; Le Roux, Jonathan
      Patent No.: 11,086,918
      Issue Date: Aug 10, 2021
    • Title: "Position Estimation Under Multipath Transmission"
      Inventors: Kim, Kyeong-Jin; Orlik, Philip V.; Hori, Chiori
      Patent No.: 11,079,495
      Issue Date: Aug 3, 2021
    • Title: "Method and System for Multi-Modal Fusion Model"
      Inventors: Hori, Chiori; Hori, Takaaki; Hershey, John R.; Marks, Tim
      Patent No.: 10,417,498
      Issue Date: Sep 17, 2019
    • Title: "Method and System for Training Language Models to Reduce Recognition Errors"
      Inventors: Hori, Takaaki; Hori, Chiori; Watanabe, Shinji; Hershey, John R.
      Patent No.: 10,176,799
      Issue Date: Jan 8, 2019
    • Title: "Method and System for Role Dependent Context Sensitive Spoken and Textual Language Understanding with Neural Networks"
      Inventors: Hori, Chiori; Hori, Takaaki; Watanabe, Shinji; Hershey, John R.
      Patent No.: 9,842,106
      Issue Date: Dec 12, 2017
    See All Patents for MERL