News

NRF Early Career Research Grant (National Research Foundation)

2026-03-24 00:00:00 +0000

We have been selected for the National Research Foundation Early Career Research Program (우수신진연구), a highly competitive national research grant supporting emerging researchers in Korea.

The awarded project focuses on developing a hallucination-free on-device egocentric multimodal AI agent equipped with self-correction mechanisms. This research aims to address one of the most critical challenges in modern AI systems: ensuring reliability and trustworthiness in real-world, context-aware environments.

By integrating actor-validator architectures, RLAIF-based alignment, and dynamic precision optimization for on-device deployment, the project explores a new paradigm where AI systems can autonomously evaluate and refine their own outputs in real time.

This work is expected to contribute not only to advancing multimodal AI and on-device intelligence, but also to establishing a foundational framework for trustworthy AI systems that can safely operate in human-centered environments.

WCCA-AK (ICCV Workshop 2025)

2025-10-20 00:00:00 +0000

WCCA-AK is a large-scale dataset of 3D scans and multi-view images capturing 100 haute couture garments by André Kim (1962–2010), one of Korea’s most iconic fashion designers. This dataset bridges computer vision research and cultural heritage preservation, enabling both faithful documentation of historical artifacts and generative exploration of artistic vision.

NVIDIA Academic Grant for Researchers (NVIDIA)

2025-09-24 00:00:00 +0000

The NVIDIA Academic Grant Program supports our research on building a Validator LLM designed to ensure logical consistency in generative AI explanations. This project focuses on moving beyond surface-level fluency in large language models by introducing a structured validation mechanism that can assess and refine the reasoning process itself.

Our approach adopts a dual-model architecture, where an actor LLM generates responses and explanations, and a validator LLM evaluates their logical coherence. By incorporating reinforcement learning from AI feedback (RLAIF), the system iteratively improves its ability to detect inconsistencies and hallucinations without relying solely on human supervision. In addition, we explore multi-pass reasoning and cross-model verification strategies to enhance robustness and reliability.

On the systems side, this research leverages NVIDIA’s optimized AI stack, including NeMo Framework and TensorRT-LLM, to ensure that the proposed validation pipeline is not only theoretically sound but also deployable in real-world environments with practical latency and scalability constraints.

Ultimately, this work aims to establish a foundation for trustworthy generative AI systems that can self-evaluate and provide explanations users can rely on, addressing one of the most critical challenges in the deployment of large-scale AI models.

Project Aria (Meta Reality Lab)

2025-07-16 00:00:00 +0000

Project Aria is a research initiative led by Meta Reality Labs aiming to develop next-generation augmented reality technologies. Through our academic partnership with Meta Reality Labs, we are currently developing an Egocentric Multimodal AI Agent leveraging Project Aria’s advanced wearable device, Aria Glass. By integrating real-time visual streams from cameras, Visual SLAM for precise spatial understanding, and sophisticated eye-tracking data, this collaboration seeks to enable context-aware, personalized AI interactions. Our goal is to explore how multimodal AI can interpret and seamlessly respond to users’ real-world environments, providing foundational insights for future AR and wearable computing applications.

AgentVox

2025-06-28 00:00:00 +0000

AgentVox — Edge-based voice assistant using Gemma LLM with Speech-to-Text and Text-to-Speech capabilities — officially Released!

Features:

  • Speech Recognition (STT): High-speed speech recognition using RealtimeSTT
  • Conversational AI (LLM): Local LLM based on Llama.cpp (Gemma 3 12B)
  • Speech Synthesis (TTS): Fast response with RealtimeTTS
  • Complete Offline Operation: All processing is done locally, ensuring privacy

DocsRay

2025-06-07 00:00:00 +0000

DocsRay — Lightweight PDF Q&A tool powered by RAG (Retrieval-Augmented Generation) with MCP (Model Context Protocol) Support — officially Released!

Features seamless MCP (Model Context Protocol) integration with Claude Desktop, comprehensive directory management capabilities, visual content analysis, and intelligent hybrid OCR system.

Now you can upload any type of documents including HWP!

Try out our online demo at docsray.com!

WCCA@ICCV2025

2025-05-23 00:00:00 +0000

We are delighted to announce that our workshop proposal to ICCV 2025 got approved!

The Workshop on Cultural Continuity of Artists (WCCA) brings together researchers, creators, and cultural institutions to explore how computer vision, multimodal AI, and XR technologies can safeguard and reinterpret artistic legacies. Our inaugural edition, co‑located with ICCV 2025, highlights the visionary South Korean fashion designer André Kim and introduces a rich, newly curated dataset from his archives.

Find out more at WCCA2025 Website!

QueryDoc

2025-05-05 00:00:00 +0000

QueryDoc — Lightweight PDF Q&A tool powered by RAG (Retrieval-Augmented Generation) — officially Released!

Just upload a PDF and start asking questions.

  • Evangel : Catholic Priest AI powered by QueryDoc
  • VerNova: Presbyterian Pastor AI powered by QueryDoc

Welcome

2024-08-09 00:00:00 +0000

Welcome to MIMIC’s offical website.

Address

GA507A, 35, Baekbeom-ro
Mapo-gu, Seoul 04107
Republic of Korea