- Wir über uns
- Zeitschrift für Semiotik
- Mitglied werden
Deutsche Gesellschaft für Semiotik (DGS) e.V. | Online: http://www.semiotik.eu/
The double-blind peer-reviewed Yearbook of Moving Image Studies (YoMIS) is now accepting articles from researchers, artists, designers, technical developers, graphic artists, computer scientists, game designers and film makers for the seventh issue, which is the final part of the »Trilogy of Synthetic Realities« concept, entitled »Mixed Reality Images. Trilogy of Synthetic Realities III«. The first issue has focussed on »Virtual Images« (in 2020), the second on »Augmented Images« (in 2021) and with the third issue the editors will address the aesthetics, cognition, and technological structure of »Mixed Reality Images«. YoMIS will be enriched by disciplines like media theory, film studies and philosophy, art and design, artistic research, image science, semiotics, phenomenology, art history, game studies, visual culture studies, computer graphics and other research areas related to moving, dynamic, virtual, augmented, or mixed reality images in general.
Do we live in the age of Baudrillard’s hyper-reality? What’s about our impression that the boundaries between our physical reality and the rising digital realities are fading? Immersive technologies like virtual reality, augmented reality, and mixed reality – summed up under the concept of synthetic or extended realities – allow and demand a completely novel form of interaction and corporeal relation with the digital images and digital objects. Working and living in synthetic extended realities is challenging our thinking about immersive media technologies and the specific aesthetics of digital image visualization.
Therefore, »Mixed Reality Images: Trilogy of Synthetic Realities III« will address the possibilities of mixed reality images that are already affecting and enriching our life in ways, only science-fiction movies, comics, and novels thought of before. Thus, contributions for this issue of the Yearbook of Moving Image Studies can concentrate on a whole variety of the pictorial aspects of mixed reality images, the specific technological conditions and situations, and the development of graphic representations regarding the different interfaces of mixed reality. Topics should focus on (but are not limited to) mixed reality images as perceptual artefacts, mixed reality as simulation in the context of virtual augmentation and augmented virtuality, the specific mixed reality performance of virtual or augmented technologies that are enabling a possible physical-world-digital-world-coupling and specific mixed reality modes of user (inter-)action, the different aspects of mixed reality aesthetics, art, design, and communication in mixed reality conditions, the new forms of psychological and perceptual interaction and narration in mixed reality media ecologies, the processual dynamic of mixed reality images, embodied and embedded interaction and cognition, effects and characteristics of mixed reality-driven body ownership illusions, the phenomenology or semiotics of user interaction in mixed realities, the coupling of digital images with the operating context, mixed reality in a multimodal framework, and the historical, cultural or philosophical evolution of mixed reality image representations.
The official deadline for abstracts is April 22, 2022. The anonymous review feedback will be given in May 2022. Long abstracts should include 600 to 900 words in length. Please send a short biography, contact details and your abstract to Prof. Dr. Lars C. Grabbe, Prof. Dr. Patrick Rupert-Kruse and Prof. Dr. Norbert M. Schmitz via: email@example.com. The official deadline for the articles is October 21, 2022. The articles should include 5.000 to 7.000 words in length. If you are interested to contribute an abstract and article you will find a specific style sheet of the Yearbook of Moving Image Studies on: www.movingimagescience.com. If you have any questions, please do not hesitate to contact the managing editors via mail.
|Kategorie:||Call for Articles|
|Kontakt:||Prof. Dr. Lars C. Grabbe|