Cybernetics and Reality Engineering

Manipulate reality to augment human capabilities

Research Staff

  • Professor Kiyoshi KIYOKAWA

    Professor
    Kiyoshi KIYOKAWA

  • Associate Professor Hideaki UCHIYAMA

    Associate Professor
    Hideaki UCHIYAMA

  • Affiliate Associate Professor Nobuchika SAKATA

    Affiliate Associate Professor
    Nobuchika SAKATA

  • Assistant Professor Naoya ISOYAMA

    Affiliate Associate Professor
    Naoya ISOYAMA

  • Assistant Professor Monica PERUSQUÍA-HERNÁNDEZ

    Assistant Professor
    Monica PERUSQUÍA-HERNÁNDEZ

  • Assistant Professor Yutaro HIRAO

    Assistant Professor
    Yutaro HIRAO

  • Affiliate Assistant Professor Felix DOLLACK

    Affiliate Assistant Professor
    Felix DOLLACK

Research Areas

Fig.1 Research fields

Fig.1 Research fields

Fig.2: Super wide field-of-view occlusion-capable optical see-through HMD

Fig.2: Super wide field-of-view occlusion-capable optical see-through HMD

Fig.3: Gustatory manipulation by using GAN-based real-time food-to-food translation

Fig.3: Gustatory manipulation by using GAN-based real-time food-to-food translation

Fig.4: Controlling interpersonal distance

Fig.4: Controlling interpersonal distance

Fig.5: A variety of head mounted displays

Fig.5: A variety of head mounted displays

Humans have acquired new capabilities by inventing various tools long before computers came up and mastering them as if they were part of the body. In this laboratory, we conduct research to create "tools of the future" by making full use of virtual reality (VR), augmented reality (AR), mixed reality (MR), human and environmental sensing, sensory representation, wearable computing, context awareness, machine learning, biological information processing and other technologies (Fig. 1). We aim to live more conveniently, more comfortably, or more securely by offering "personalized reality" which empathizes each person. Through such information systems, we would like to contribute to the realization of an inclusive society where all people can maximize their abilities and help each other.

Sensing: Measuring people and the environment

We are studying various sensing technologies that assess human and environmental
conditions using computer vision, pattern recognition, machine learning, etc.

  • Estimation of user’s physiological and psychological state from gaze and body behavior
  • HMD calibration and gaze tracking using corneal reflection images

Display: Manipulating perception

We are studying technologies, such as virtual reality and augmented reality, to freely manipulate and modulate various sensations such as vision and auditory, their effects, and their display hardware.

  • Super wide field of view occlusion-capable see-through HMD (Fig. 2)
  • Gustatory manipulation by GAN-based food-to-food translation (Fig. 3)
  • Tendon vibration to increase vision-induced kinesthetic illusions in VR
  • A non-grounded and encountered-type haptic display using a drone

Interaction: Creating and using tools

We combine sensing and technologies to study new ways of interaction between human and human, and human and the environment.

  • Controlling interpersonal distance by a depth sensor and a video see-through HMD (Fig. 4)
  • Investigation on priming effects of visual information on wearable displays
  • AR pet recognizing people and the environment and having own emotions

Research Equipment

  • A variety of head mounted displays (Fig. 5)

Research Grants, Collaborations, Social Services etc. (2019)

  • MEXT Grants-in-Aid (Kakenhi) (A x 2, B x 3, C x 2), SCOPE, JASSO,etc.
  • Collaboration (TIS, CyberWalker, etc.)
  • Steering / Organizing Committee members of IEEE VR, ISMAR, APMAR, etc.