{"id":122,"date":"2023-08-09T15:40:49","date_gmt":"2023-08-09T06:40:49","guid":{"rendered":"https:\/\/www.teikyo.jp\/visualsciencelab\/?page_id=122"},"modified":"2024-09-19T08:15:17","modified_gmt":"2024-09-18T23:15:17","slug":"paper","status":"publish","type":"page","link":"https:\/\/www.teikyo.jp\/visualsciencelab\/work\/paper\/","title":{"rendered":"\u8ad6\u6587\u30fb\u7dcf\u8aac"},"content":{"rendered":"\n<h2 class=\"wp-block-heading has-text-align-center is-style-section_ttl\"><span class=\"swl-marker mark_blue\">\u8ad6\u6587\u30fb\u7dcf\u8aac<\/span><\/h2>\n\n\n\n<div class=\"swell-block-accordion is-style-main\">\n<details class=\"swell-block-accordion__item is-opened\" open data-swl-acc=\"wrapper\"><summary class=\"swell-block-accordion__title\" data-swl-acc=\"header\"><span class=\"swell-block-accordion__label\"><span class=\"swl-fz u-fz-l\"><strong>2024<\/strong><\/span><\/span><span class=\"swell-block-accordion__icon c-switchIconBtn\" data-swl-acc=\"icon\" aria-hidden=\"true\" data-opened=\"true\"><i class=\"__icon--closed icon-caret-down\"><\/i><i class=\"__icon--opened icon-caret-up\"><\/i><\/span><\/summary><div class=\"swell-block-accordion__body\" data-swl-acc=\"body\">\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">1\uff0e\u7279\u96c6\uff1aAI\u6280\u8853\u306e\u6700\u65b0\u52d5\u5411 \u30a2\u30a4\u30c8\u30e9\u30c3\u30ad\u30f3\u30b0\u3068\u4eba\u5de5\u77e5\u80fd<\/span><\/strong><br>\u5e83\u7530 \u96c5\u548c.<br><strong>\u8996\u899a\u306e\u79d1\u5b66, 44(4), 89-93, 2024\u5e742\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">2\uff0e<\/span><\/strong><span class=\"swl-inline-color has-swl-main-color\"><strong>\u4e73\u5150\u773c\u632f\u304b\u3089\u773c\u632f\u963b\u6b62\u75c7\u5019\u7fa4\u306b\u4ee5\u964d\u3057\u305f\u3068\u601d\u308f\u308c\u308b\u4e00\u4f8b<\/strong><\/span><br>Megumi Fukushima, Masakazu Hirota, Takafumi Yukimori, Akio Hayashi, Yoko Hirohara, Makoto Saika, Kumiko Matsuoka.<br><strong>\u773c\u79d1\u81e8\u5e8a\u7d00\u8981,&nbsp;17(3),&nbsp;237-239, 2024\u5e743\u6708<\/strong><\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">3\uff0e<\/span><\/strong><span class=\"swl-inline-color has-swl-main-color\"><strong>Evaluation of objective and subjective binocular ocular refraction with looking in type<\/strong><\/span><br>Megumi Fukushima, Masakazu Hirota, Takafumi Yukimori, Akio Hayashi, Yoko Hirohara, Makoto Saika, Kumiko Matsuoka.<br><strong>BMC Ophthalmology,&nbsp;24,&nbsp;170, 2024\u5e744\u6708<\/strong><\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">4\uff0e<\/span><\/strong><span class=\"swl-inline-color has-swl-main-color\"><strong>Ocular Accommodative and Pupillary Responses During Fixation on Augmented Reality With a Maxwellian Display<\/strong><\/span><br>Masakazu Hirota, Kakeru Sasaki, Kanako Kato, Ryota Nakagomi, Ryusei Takigawa, Chinatsu Kageyama, Seiji Morino, Makoto Suzuki, Toshifumi Mihashi, Atsushi Mizota, Takao Hayashi.<br><strong>Investigative Ophthalmology &amp; Visual Science,\u00a065(11),\u00a030, 2024\u5e749\u6708<\/strong><\/p>\n\n\n\n\n<\/div><\/details>\n\n\n\n<details class=\"swell-block-accordion__item\" data-swl-acc=\"wrapper\"><summary class=\"swell-block-accordion__title\" data-swl-acc=\"header\"><span class=\"swell-block-accordion__label\"><span class=\"swl-fz u-fz-l\"><strong>2023<\/strong><\/span><\/span><span class=\"swell-block-accordion__icon c-switchIconBtn\" data-swl-acc=\"icon\" aria-hidden=\"true\" data-opened=\"false\"><i class=\"__icon--closed icon-caret-down\"><\/i><i class=\"__icon--opened icon-caret-up\"><\/i><\/span><\/summary><div class=\"swell-block-accordion__body\" data-swl-acc=\"body\">\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">1\uff0e\u9060\u8996\u6027\u4e0d\u540c\u8996\u5f31\u8996\u306b\u304a\u3051\u308b OCT angiography \u3092\u7528\u3044\u305f\u4e2d\u5fc3\u7aa9\u7121\u8840\u7ba1\u9818\u57df\u306e\u8a55\u4fa1<\/span><\/strong><br>\u6e21\u90e8 \u7dad, \u81fc\u4e95 \u5343\u60e0, \u5e83\u7530 \u96c5\u548c, \u6797 \u5b5d\u96c4.<br><strong>\u773c\u79d1\u81e8\u5e8a\u7d00\u8981, 16(3): 174-178, 2023\u5e743\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">2\uff0e\u4e0a\u659c\u7b4b\u9ebb\u75fa\u306b\u304a\u3051\u308b\u4e0a\u4e0b\u878d\u50cf\u57df\u306e\u691c\u8a0e<\/span><\/strong><br>\u4f50\u3005\u6728 \u68a2, \u4e2d\u5ddd \u771f\u7d00, \u4f50\u3005\u6728 \u7fd4, \u6c34\u91ce \u5609\u4fe1, \u6797 \u5b5d\u96c4.<br><strong>\u773c\u79d1\u81e8\u5e8a\u7d00\u8981, 16(4): 265-268, 2023\u5e744\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">3\uff0eEvaluation of Aniseikonia in Patients with Successfully Treated Anisometropic Amblyopia Using Spatial Aniseikonia Test<\/span><\/strong><br>Ryusei Takigawa, Kakeru Sasaki, Masakazu Hirota, Maki Nakagawa, Kozue Sasaki, Toshifumi Mihashi, Yoshinobu Mizuno, Atsushi Mizota, Kumiko Matsuoka.<br><strong>Journal of Clinical Medicine, 12(11): 3766, 2023\u5e745\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">4\uff0eOcular refraction changes during post-rotatory nystagmus<\/span><\/strong><br>Masakazu Hirota, Ryusei Takigawa, Chinatsu Kageyama, Kanako Kato, Ryota Nakagomi, Kakeru Sasaki, Takao Hayashi.<br><strong>ISA Transactions 2022, 168-172, 2023\u5e746\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">5\uff0eLight-Field Visual System for the Remote Robot Operation Interface<\/span><\/strong><br>Tetsuro Morimoto, Masakazu Hirota, Kakeru Sasaki, Kanako Kato, Ryusei Takigawa, Shigenobu Yoneyama, Takao Hayashi, Takashi Fujikado, Satoru Tokuhisa.<br><strong>IEEE\/RSJ International Conference on Intelligent Robots and Systems 2023, 2023\u5e7410\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">6\uff0e\u30c9\u30e9\u30a4\u30a2\u30a4\u8a3a\u7642\u306e\u65b0\u6642\u4ee3\uff1a\u30c9\u30e9\u30a4\u30a2\u30a4\u3068\u773c\u75b2\u52b4<\/span><\/strong><br>\u5e83\u7530 \u96c5\u548c\uff0c\u6e21\u8fba \u771f\u751f.<br><strong><strong>Monthly Book OCULISTA,&nbsp;128,&nbsp;29-34, 2023\u5e7411\u6708<\/strong><\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">7\uff0eEffect of binocular vision during target shooting in archery<\/span><\/strong><br>Masakazu Hirota, Tatsuhiro Hanai, Takeshi Morimoto.<br><strong>PLos One 2022, 18(11), e0294985<\/strong>, <strong>2023\u5e7411\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">8\uff0e\u30a2\u30c8\u30d4\u30fc\u898b\u3048\u308b\u5316\u30a2\u30d7\u30ea\uff0d\u30a2\u30c8\u30d4\u30e8\u3068\u30d3\u30c3\u30b0\u30c7\u30fc\u30bf\u89e3\u6790<\/span><\/strong><br>\u8d64\u7a42 \u4eae\u592a\u90ce, \u5bae\u5ddd \u660e\u5927, \u5e83\u7530 \u96c5\u548c, \u8db3\u7acb \u525b\u4e5f.<br><strong>Journal of Internet of Medical Things, 6(1), 24-27, 2023\u5e7410\u6708<\/strong>\uff0e<\/p>\n<\/div><\/details>\n<\/div>\n\n\n\n<div class=\"swell-block-accordion is-style-main\">\n<details class=\"swell-block-accordion__item\" data-swl-acc=\"wrapper\"><summary class=\"swell-block-accordion__title\" data-swl-acc=\"header\"><span class=\"swell-block-accordion__label\"><strong><span class=\"swl-fz u-fz-l\">2022<\/span><\/strong><\/span><span class=\"swell-block-accordion__icon c-switchIconBtn\" data-swl-acc=\"icon\" aria-hidden=\"true\" data-opened=\"false\"><i class=\"__icon--closed icon-caret-down\"><\/i><i class=\"__icon--opened icon-caret-up\"><\/i><\/span><\/summary><div class=\"swell-block-accordion__body\" data-swl-acc=\"body\">\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">1\uff0e\u30b9\u30de\u30fc\u30c8\u30d5\u30a9\u30f3\u4f7f\u7528\u524d\u5f8c\u306e\u773c\u7403\u9ad8\u6b21\u53ce\u5dee\u304a\u3088\u3073\u878d\u50cf\u5e45\u306e\u5909\u5316<\/span><\/strong><br>\u897f\u6751 \u88d5\u6a39, \u5e83\u7530 \u96c5\u548c, \u6797 \u5b5d\u96c4.<br><strong>\u773c\u79d1\u81e8\u5e8a\u7d00\u8981, 15(1): 38-41, 2022\u5e741\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">2\uff0eVideo-oculography\u306e\u30ad\u30e3\u30ea\u30d6\u30ec\u30fc\u30b7\u30e7\u30f3\u65b9\u6cd5\u306e\u691c\u8a0e<\/span><\/strong><br>\u6c60\u7530 \u7d50\u4f73, \u5e83\u7530 \u96c5\u548c, \u52a0\u85e4 \u53ef\u5948\u5b50, \u677e\u5ca1 \u4e45\u7f8e\u5b50.<br><strong>\u773c\u79d1\u81e8\u5e8a\u7d00\u8981, 15(3): 194-197, 2022\u5e743\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">3\uff0eAnalysis of smooth pursuit eye movements in a clinical context by tracking the target and eyes<\/span><\/strong><br>Masakazu Hirota, Kanako Kato, Megumi Fukushima, Yuka Ikeda, Takao Hayashi, Atsushi Mizota.<br><strong>Scientific Reports, 12(1): 8501, 2022\u5e745\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">4\uff0eAutomatic Screening of the Eyes in a Deep-Learning\u2013Based Ensemble Model Using Actual Eye Checkup Optical Coherence Tomography Images<\/span><\/strong><br>Masakazu Hirota, Shinji Ueno, Taiga Inooka, Yasuki Ito, Hideo Takeyama, Yuji Inoue, Emiko Watanabe, Atsushi Mizota.<br><strong>Applied Sciences, 12(14): 6872, 2022\u5e747\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">5\uff0eEvaluation of ocular biometry in the Japanese population using a multicenter approach: Prospective observational study<\/span><\/strong><br>Takushi Kawamorita, Hiroshi Uozato, Tetsuro Oshika, Kazuno Negishi, Takashi Fujikado, Akira Murakami, Kazutaka Kamiya, Naoyuki Maeda, Yuta Ueno, Kazuhiro Onuma, Masakazu Hirota, Rie Hoshikawa, Sachiko Masui, Masahiro Yamaguchi, Toshifumi Mihashi.<br><strong>PLOS ONE, 17(7): e0271814, 2022\u5e747\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">6\uff0eComparisons of size of foveal avascular zone area among children with posterior microphthalmos, high hyperopia, and normal eyes<\/span><\/strong><br>Kozue Sasaki, Kakeru Sasaki, Masakazu Hirota, Takao Hayashi, Atsushi Mizota.<br><strong>International Ophthalmology, 42(8): 2599-2607, 2022\u5e748\u6708<\/strong>.<\/p>\n<\/div><\/details>\n<\/div>\n\n\n\n<div class=\"swell-block-accordion is-style-main\">\n<details class=\"swell-block-accordion__item\" data-swl-acc=\"wrapper\"><summary class=\"swell-block-accordion__title\" data-swl-acc=\"header\"><span class=\"swell-block-accordion__label\"><span class=\"swl-fz u-fz-l\"><strong>2021<\/strong><\/span><\/span><span class=\"swell-block-accordion__icon c-switchIconBtn\" data-swl-acc=\"icon\" aria-hidden=\"true\" data-opened=\"false\"><i class=\"__icon--closed icon-caret-down\"><\/i><i class=\"__icon--opened icon-caret-up\"><\/i><\/span><\/summary><div class=\"swell-block-accordion__body\" data-swl-acc=\"body\">\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>1\uff0e\u30a2\u30a4\u30c8\u30e9\u30c3\u30ab\u30fc\u3092\u5229\u7528\u3057\u305f\u9593\u6b20\u6027\u5916\u659c\u8996\u306e\u8996\u6a5f\u80fd\u8a55\u4fa1<\/strong><\/span><br>\u5e83\u7530 \u96c5\u548c.<br><strong>\u773c\u79d1\u81e8\u5e8a\u7d00\u8981, 14(1): 20-30, 2021\u5e741\u6708<\/strong>.<\/p>\n\n\n\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>2\uff0e\u8996\u899a\u88dc\u52a9\u30c7\u30d0\u30a4\u30b9 OrCam MyEye 2 \u306e\u6027\u80fd\u8a55\u4fa1<\/strong><\/span><br>\u5e83\u7530 \u96c5\u548c, \u68ee\u672c \u58ee, \u963f\u66fd\u6cbc \u65e9\u82d7, \u4e09\u597d \u667a\u6e80, \u4e0d\u4e8c\u9580 \u5c1a.<br><strong>\u65e5\u672c\u30ed\u30fc\u30d3\u30b8\u30e7\u30f3\u5b66\u4f1a\u8a8c, 20: 73-77, 2021\u5e741\u6708<\/strong>.<\/p>\n\n\n\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>3\uff0eAutomatic Recording of the Target Location During Smooth Pursuit Eye Movement Testing Using Video-Oculography and Deep Learning-Based Object Detection<\/strong><\/span><br>Masakazu Hirota, Takao Hayashi, Emiko Watanabe, Yuji Inoue, Atsushi Mizota.<br><strong>Translational Vision Science &amp; Technology, 10(6): 1, 2021\u5e745\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">4\uff0e\u8fd1\u8996\u6027\u76f4\u4e71\u8996\u773c\u306b\u304a\u3051\u308b\u5c48\u6298\u5ea6\u3068\u5168\u773c\u7403\uff0c\u89d2\u819c\uff0c\u773c\u5185\u9ad8\u6b21\u53ce\u5dee\u3068\u306e\u76f8\u95a2<\/span><\/strong><br>\u52a0\u85e4 \u53ef\u5948\u5b50, \u85e4\u4ee3 \u5c1a\u6587, \u5e83\u7530 \u96c5\u548c, \u4e2d\u8fbc \u4eae\u592a, \u677e\u5ca1 \u4e45\u7f8e\u5b50, \u5c0f\u6797 \u514b\u5f66.<br><strong>\u8996\u899a\u306e\u79d1\u5b66, 42(3): 44-47, 2021\u5e7410\u6708<\/strong>.<\/p>\n\n\n\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>5\uff0eRefractive changes with post-rotatory nystagmus in healthy individuals<\/strong><\/span><br>Masakazu Hirota, Ryusei Takigawa, Chinatsu Okabe, Kanako Kato, Ryota Nakagomi, Kakeru Sasaki, Takao Hayashi.<br><strong>International Ophthalmology, 42(2): 559-573, 2021\u5e7410\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">6\uff0e\u6700\u65b0\u6a5f\u5668\u30ec\u30dd\u30fc\u30c8\u300cCyclophorometer\u300dSpec<\/span><\/strong><br>\u4f50\u3005\u6728 \u7fd4.<br><strong>\u5149\u6280\u8853\u30b3\u30f3\u30bf\u30af\u30c8, 59: 22-28, 2021\u5e7412\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">7\uff0e\u30bd\u30d5\u30c8\u30b3\u30f3\u30bf\u30af\u30c8\u30ec\u30f3\u30ba\u306e\u88c5\u7528\u30b9\u30b1\u30b8\u30e5\u30fc\u30eb\u53ca\u3073\u6750\u8cea\u304c\u89d2\u819c\u5185\u76ae\u7d30\u80de\u306e\u5f62\u614b\u306b\u53ca\u307c\u3059\u5f71\u97ff<\/span><\/strong><br>\u4e2d\u8fbc \u4eae\u592a, \u677e\u5ca1 \u4e45\u7f8e\u5b50, \u81fc\u4e95 \u5343\u60e0, \u6797 \u5b5d\u96c4.<br><strong>\u65e5\u672c\u8996\u80fd\u8a13\u7df4\u58eb\u5354\u4f1a\u8a8c, 49: 69-74, 2020\u5e7412\u6708<\/strong>.<\/p>\n<\/div><\/details>\n<\/div>\n\n\n\n<div class=\"swell-block-accordion is-style-main\">\n<details class=\"swell-block-accordion__item\" data-swl-acc=\"wrapper\"><summary class=\"swell-block-accordion__title\" data-swl-acc=\"header\"><span class=\"swell-block-accordion__label\"><span class=\"swl-fz u-fz-l\"><strong>2020<\/strong><\/span><\/span><span class=\"swell-block-accordion__icon c-switchIconBtn\" data-swl-acc=\"icon\" aria-hidden=\"true\" data-opened=\"false\"><i class=\"__icon--closed icon-caret-down\"><\/i><i class=\"__icon--opened icon-caret-up\"><\/i><\/span><\/summary><div class=\"swell-block-accordion__body\" data-swl-acc=\"body\">\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>1\uff0eBinocular Coordination during Smartphone Reading in Esophoric Patients<\/strong><\/span><br>Masakazu Hirota, Takeshi Morimoto, Tomomitsu Miyoshi, Takashi Fujikado.<br><strong>Journal of binocular vision and ocular motility, 70(1): 15-20, 2020\u5e741\u6708.<\/strong><\/p>\n\n\n\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>2\uff0eObjective evaluation of visual fatigue in patients with intermittent exotropia<\/strong><\/span><br>Masakazu Hirota, Kozue Yada, Takeshi Morimoto, Takao Endo, Tomomitsu Miyoshi, Suguru Miyagawa, Yoko Hirohara, Tatsuo Yamaguchi, Makoto Saika, Takashi Fujikado.<br><strong>PloS one, 15(3): e0230788, 2020\u5e743\u6708.<\/strong><\/p>\n\n\n\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>3\uff0e\u6df1\u5c64\u5b66\u7fd2\u306b\u304a\u3051\u308b\u7dd1\u5185\u969c\u63a8\u5b9a\u7cbe\u5ea6\u306b\u5bfe\u3059\u308b\u8272\u60c5\u5831\u306e\u5f71\u97ff<\/strong><\/span><br>\u5e83\u7530 \u96c5\u548c, \u4e95\u4e0a \u8ce2\u6cbb, \u4e09\u6751 \u9054\u54c9, \u6797 \u5b5d\u96c4, \u53e4\u5fb3 \u7d14\u4e00, \u6fa4 \u667a\u535a, \u6e9d\u7530 \u6df3.<br><strong>\u7b2c 24 \u56de \u65e5\u672c\u533b\u7642\u60c5\u5831\u5b66\u4f1a\u6625\u671f\u5b66\u8853\u5927\u4f1a \u30b7\u30f3\u30dd\u30b8\u30a6\u30e0 2020 24th (Web), PB11, 2020\u5e746\u6708<\/strong>.<\/p>\n\n\n\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>4\uff0e\u878d\u50cf\u7dad\u6301\u80fd\u529b\u306b\u3088\u308b\u773c\u75b2\u52b4\u306e\u4ed6\u899a\u7684\u5b9a\u91cf\u8a55\u4fa1<\/strong><\/span><br>Masakazu Hirota.<br><strong>VISION, 32(4): 107-117, 2020\u5e7410\u6708<\/strong>.<\/p>\n\n\n\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>5\uff0eEffect of color information on the diagnostic performance of glaucoma in deep learning using few fundus images<\/strong><\/span><br>Masakazu Hirota, Atsushi Mizota, Tatsuya Mimura, Takao Hayashi, Junichi Kotoku, Tomohiro Sawa, Kenji Inoue.<br><strong>International Ophthalmology, 40(11): 3013-3022, 2020\u5e7411\u6708<\/strong>.<\/p>\n\n\n\n<p><span class=\"swl-inline-color has-swl-main-color\"><strong>6\uff0eSimultaneous Measurement of Objective and Subjective Accommodation in Response to Step Stimulation<\/strong><\/span><br>Masakazu Hirota, Takeshi Morimoto, Tomomitsu Miyoshi, Takashi Fujikado.<br><strong>Investigative Ophthalmology &amp; Visual Science, 61(13): 38, 2020\u5e7411\u6708<\/strong>.<\/p>\n\n\n\n<p><strong><span class=\"swl-inline-color has-swl-main-color\">7\uff0e\u30bd\u30d5\u30c8\u30b3\u30f3\u30bf\u30af\u30c8\u30ec\u30f3\u30ba\u306e\u88c5\u7528\u30b9\u30b1\u30b8\u30e5\u30fc\u30eb\u53ca\u3073\u6750\u8cea\u304c\u89d2\u819c\u5185\u76ae\u7d30\u80de\u306e\u5f62\u614b\u306b\u53ca\u307c\u3059\u5f71\u97ff<\/span><\/strong><br>\u4e2d\u8fbc \u4eae\u592a, \u677e\u5ca1 \u4e45\u7f8e\u5b50, \u81fc\u4e95 \u5343\u60e0, \u52a0\u85e4 \u53ef\u5948\u5b50, \u6797 \u5b5d\u96c4.<br><strong>\u65e5\u672c\u8996\u80fd\u8a13\u7df4\u58eb\u5354\u4f1a\u8a8c, 49: 91-96, 2020\u5e7412\u6708<\/strong>.<\/p>\n<\/div><\/details>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>\u8ad6\u6587\u30fb\u7dcf\u8aac<\/p>\n","protected":false},"author":1,"featured_media":0,"parent":11,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"swell_btn_cv_data":"","footnotes":""},"class_list":["post-122","page","type-page","status-publish","hentry"],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/www.teikyo.jp\/visualsciencelab\/wp-json\/wp\/v2\/pages\/122","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.teikyo.jp\/visualsciencelab\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/www.teikyo.jp\/visualsciencelab\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/www.teikyo.jp\/visualsciencelab\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.teikyo.jp\/visualsciencelab\/wp-json\/wp\/v2\/comments?post=122"}],"version-history":[{"count":0,"href":"https:\/\/www.teikyo.jp\/visualsciencelab\/wp-json\/wp\/v2\/pages\/122\/revisions"}],"up":[{"embeddable":true,"href":"https:\/\/www.teikyo.jp\/visualsciencelab\/wp-json\/wp\/v2\/pages\/11"}],"wp:attachment":[{"href":"https:\/\/www.teikyo.jp\/visualsciencelab\/wp-json\/wp\/v2\/media?parent=122"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}