Automating Manga Character Analysis: A Robust Deep Vision-Transformer Approach to Facial Landmark Detection
dc.contributor.author | Vachmanus S. | |
dc.contributor.author | Phinklao N. | |
dc.contributor.author | Phongsarnariyakul N. | |
dc.contributor.author | Plongcharoen T. | |
dc.contributor.author | Hotta S. | |
dc.contributor.author | Tuarob S. | |
dc.contributor.correspondence | Vachmanus S. | |
dc.contributor.other | Mahidol University | |
dc.date.accessioned | 2024-09-25T18:12:30Z | |
dc.date.available | 2024-09-25T18:12:30Z | |
dc.date.issued | 2024-01-01 | |
dc.description.abstract | Comics, particularly Japanese manga, are a powerful medium that blends images and text to convey ideas and encapsulate a unique cultural heritage. Going beyond mere entertainment, manga merges diverse styles and content deeply rooted in Japanese cultural heritage. This study utilizes computer vision analysis, with a specific focus on facial landmark detection, acknowledging the growing significance of technology in analyzing manga images. Through a comprehensive exploration of various methods, the research identifies the extended version of Bidirectional Encoder Representations from Transformers (BERT), BERT Pre-Training of Image Transformers (BEiT), model as a standout performer due to its efficiency and effectiveness. The BEiT model's success lies in its ability to extract facial features, consequently establishing itself as a go-To solution for landmark detection on manga faces. The outcomes achieved the lowest Failure Rate compared to other landmark detection networks, with a Failure Rate of approximately 9.4% and a Mean Average Error of about 4.6 pixels. Beyond its technical accomplishments, this study carries a cultural significance, contributing to the ongoing narrative of manga in Japan. | |
dc.identifier.citation | IEEE Access (2024) | |
dc.identifier.doi | 10.1109/ACCESS.2024.3459419 | |
dc.identifier.eissn | 21693536 | |
dc.identifier.scopus | 2-s2.0-85204185759 | |
dc.identifier.uri | https://repository.li.mahidol.ac.th/handle/20.500.14594/101342 | |
dc.rights.holder | SCOPUS | |
dc.subject | Materials Science | |
dc.subject | Computer Science | |
dc.subject | Engineering | |
dc.title | Automating Manga Character Analysis: A Robust Deep Vision-Transformer Approach to Facial Landmark Detection | |
dc.type | Article | |
mu.datasource.scopus | https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85204185759&origin=inward | |
oaire.citation.title | IEEE Access | |
oairecerif.author.affiliation | Mahidol University | |
oairecerif.author.affiliation | Tokyo University of Agriculture and Technology |