item3a

最新情報 | ごあいさつ | 開催日程 | 構成員 | 連絡先 | English

2024 年

12 月
11 月
10 月
9 月
8 月
7 月
6 月
5 月
4 月

3 月
2 月
1 月

2023 年

12 月
11 月
10 月
9 月
8 月
7 月
6 月
5 月
4 月

3 月
2 月
1 月

2022 年

12 月
11 月
10 月
9 月
8 月
7 月
6 月
5 月
4 月

3 月
2 月
1 月

2021 年

12 月
11 月
10 月
9 月
8 月
7 月
6 月
5 月
4 月

3 月
2 月
1 月

2020 年

12 月
11 月
10 月
9 月
8 月
7 月
6 月
5 月
4 月

2020 年 3 月以前

2021 年 5


The 52nd Perceptual Frontier Seminar: Non-invasive exploration of the brain with visual, tactile, and auditory stimuli

Date and time: Monday, 10 May 2021, 16:40-18:10
Venue: Talks are online. After the talks, we will get together in Kazuo's Office, Room 709 on the 7th Floor of Building 3, Ohashi Campus, Kyushu University, Fukuoka, Japan <http://www.design.kyushu-u.ac.jp/kyushu-u/english/access>
Language: English
Organizer: Kazuo UEDA (Kyushu Univ./ReCAPS/Research and Development Center for Five-Sense Devices)

Program

16:40-17:10 Modality specificity of amodal completion
Hiroshige TAKEICHI*, Keito TANIGUCHI**, Hiroaki SHIGEMASU**
*RIKEN and ReCAPS, Kyushu University, **Kochi University of Technology

In order to elucidate the concept of amodal completion, which refers somewhat paradoxically to visual perception of occluded therefore invisible parts of objects, a virtual-reality experiment was conducted to examine the role of haptic information in amodal completion. Fifteen participants performed a visual word recognition task on partially-missing five-letter words that did or did not accompany a visually and/or haptically presented occluder as the stimuli. Word recognition was facilitated by a visual cue to depth separation both forwards and backwards by binocular stereopsis but not by a haptic cue, which suggests that amodal completion can be based on retinotopic segmentation alone but not by three-dimensional visual or haptic depth perception and is a modality-specific rather than an a-modal processing. (This presentation is based on the FY2020 graduation thesis of the second author.)

17:10-17:30 A short introduction to gaze metrics as early detection biomarkers in Schizophrenia
Alexandra WOLF*, Kazuo UEDA**, and Yoji HIRANO***
*JSPS International Research Fellow (ReCAPS, Kyushu University)
**Dept. Human Science/ReCAPS/Research and Development Center for Five-Sense Devices, Kyushu University
***Department of Neuropsychiatry, Kyushu University

Eye‐trackers noninvasively measure one's unconscious eye behavior towards presented stimuli. With modern eye‐tracking instrumentation, scientists aim to reflect the viewer's attention and consequently disclose a part of the viewer's thought processes. Recorded and analyzed gaze parameters help in answering questions about which elements attract one's eyes, in what order, and how often. Despite an increasing focus on eye behavior within cognitive neuroscience and an active import of eye‐tracking paradigms in other fields, vision science lags behind in clinical research. However, a promising trend has started to emerge: the study of visual information processing by re‐examining paradigms with modern eye‐tracking devices.

17:30-18:00 Checkerboard speech
Kazuo UEDA*, Riina KAWAKAMI**, Hiroshige TAKEICHI***
*Dept. Human Science/ReCAPS/Research and Development Center for Five-Sense Devices, Kyushu University
**Dept. Acoustic Design, Kyushu University
***RIKEN and ReCAPS, Kyushu University

"Checkerboard speech" is a kind of degraded speech discarding 50% of original speech, to study spectrotemporal characteristics of speech perception. Here we show that 20-band checkerboard speech maintained nearly 100% intelligibility irrespective of segment duration in the range from 20 to 320 ms, whereas 2- and 4-band checkerboard speech showed a trough of 35% to 40% intelligibility between the segment durations of 80 and 160 ms (n = 2 and n = 20), and that mosaicked checkerboard speech stimuli showed less than 10% intelligibility except for the stimuli with the finest resolution (20 frequency bands and 20-ms segment duration). The results suggest close connections with the modulation power spectrums of the stimuli, a spectrotemporal interaction in speech perception, and perceptual cue integration based on temporal fine structure.

After the talks, we will get together in Kazuo's office, Room 709 on the 7th Floor of Building 3.

最新情報 | ごあいさつ | 開催日程 | 構成員 | 連絡先 | English

Last updated:
Copyright (c) 2013-2024 Research Center for Applied Perceptual Science, Kyushu University. All rights reserved.