Cybernetics and Reality Engineering

Manipulate reality to augment human capabilities

Research Staff

  • Prof. Kiyoshi Kiyokawa

    Kiyoshi Kiyokawa

  • Assoc.Prof. Tomokazu Sato

    Affiliate Prof.
    Tomokazu Sato

  • Assoc.Prof. Nobuchika Sakata

    Nobuchika Sakata

  • Affiliate Assoc.Prof. Norihiko Kawai

    Affiliate Assoc.Prof.
    Norihiko Kawai

  • Assist.Prof. Naoya Isoyama

    Naoya Isoyama

E-mail { kiyo, tomoka-s, sakata, n-yuta, norihi-k, isoyama }[at]

Research Areas

Humans have acquired new capabilities by inventing various tools long before computers came up and mastering them as if they were part of the body. In this laboratory, we conduct research to create "tools of the future" by making full use of virtual reality (VR), augmented reality (AR), mixed reality (MR), human and environmental sensing, sensory representation, wearable computing, context awareness, machine learning, biological information processing and other technologies (Fig. 1). We aim to live more conveniently, more comfortably, or more securely by offering "personalized reality" which empathizes each person. Through such information systems, we would like to contribute to the realization of an inclusive society where all people can maximize their abilities and help each other.

Sensing: Measuring people and the environment

We are studying various sensing technologies that assess human and environmental
conditions using computer vision, pattern recognition, machine learning, etc.

Estimation of user’s physiological and psychological state from gaze and body behavior

HMD calibration and gaze tracking using corneal reflection images

Display: Manipulating perception

We are studying technologies, such as virtual reality and augmented reality, to freely manipulate and modulate various sensations such as vision and auditory, their effects, and their display hardware.

Super wide field of view occlusion-capable see-through HMD (Fig. 2)

Gustatory manipulation by GAN-based food-to-food translation (Fig. 3)

Tendon vibration to increase vision-induced kinesthetic illusions in VR

A non-grounded and encountered-type haptic display using a drone

Interaction: Creating and using tools

We combine sensing and technologies to study new ways of interaction between human and human, and human and the environment.

Controlling interpersonal distance by a depth sensor and a video see-through HMD (Fig. 4)

Investigation on priming effects of visual information on wearable displays

AR pet recognizing people and the environment and having own emotions

Research Equipment

A variety of head mounted displays (Fig. 5)

Research Grants, Collaborations, Social Services etc. (2019)

MEXT Grants-in-Aid (Kakenhi) (A x 2, B x 3, C x 2), SCOPE, JASSO, etc.

Collaboration (TIS, CyberWalker, etc.)

Steering / Organizing Committee members of IEEE VR, ISMAR, APMAR, etc.


Fig.1 Research fields

Fig.2: Super wide field-of-view occlusion-capable optical see-through HMD

Fig.2: Super wide field-of-view occlusion-capable optical see-through HMD

Fig.3: Gustatory manipulation by using GAN-based real-time food-to-food translation

Fig.3: Gustatory manipulation by using GAN-based real-time food-to-food translation

Fig.4: Controlling interpersonal distance

Fig.4: Controlling interpersonal distance

Fig.5: A variety of head mounted displays

Fig.5: A variety of head mounted displays