logo
#

Latest news with #ACL2025

AI teaching assistant shows real promise at S. Korea tech university KAIST
AI teaching assistant shows real promise at S. Korea tech university KAIST

Korea Herald

time3 days ago

  • Science
  • Korea Herald

AI teaching assistant shows real promise at S. Korea tech university KAIST

South Korea's top science and engineering university has rolled out a custom-built AI teaching assistant to help students in large graduate courses, and early results show it helped reduce repetitive student questions while encouraging more active, self-directed learning. At Korea Advanced Institute of Science and Technology, a team of researchers led by AI graduate school professor Choi Yoon-jae and design professor Hong Hwa-jung developed a virtual teaching assistant, or VTA, that can answer student questions at any time, with responses tailored to specific lectures and coursework. The tool was piloted last fall in a programming class for AI, taken by 477 graduate students. It's the first time such a system has been tested at scale in a Korean university setting. What sets the KAIST VTA apart is that it isn't a generic chatbot. Instead, it runs on what's called a retrieval-augmented generation model, which pulls directly from course materials like slides, coding exercises and lecture videos. When students ask a question, the system finds the most relevant content and formulates a response based on that context. This means answers are grounded in what's actually taught, not just generated from a general AI model. Over 50 percent of students used the system regularly during the 14-week semester, generating nearly 3,870 questions and answers. Students without a strong background in AI or coding were among the most active users, suggesting the VTA helped close knowledge gaps for those new to the subject. These figures come from internal usage data collected by KAIST during the semester. The system didn't just benefit students. According to lead TA and doctoral researcher Kwon Soon-jun, it reduced the number of routine questions from students, such as basic concept definitions or explanations already given in class. That allowed human teaching assistants to focus more on deeper, more complex issues. Compared to the previous year's course, the volume of questions requiring direct responses from TAs dropped by around 40 percent, based on data compiled by Professor Choi. Students also appeared more comfortable asking questions through the VTA than to human TAs, especially when it came to theoretical topics. Surveys conducted by Choi's research team before, during and after the course showed that students became increasingly confident in the system's reliability, and those who had been hesitant to speak up in class reported higher satisfaction levels when using the AI assistant. The VTA's source code has been released publicly on GitHub to encourage adoption by other educators and researchers. The work was also accepted to the Industry Track at ACL 2025, one of the leading international conferences in natural language processing

AIM Intelligence: Inside ACL 2025's "Triple Threat" to Unsafe AI - A Global Alliance of Stanford, AWS, UMich, SNU, Yonsei, KAIST & UOS
AIM Intelligence: Inside ACL 2025's "Triple Threat" to Unsafe AI - A Global Alliance of Stanford, AWS, UMich, SNU, Yonsei, KAIST & UOS

Yahoo

time7 days ago

  • Business
  • Yahoo

AIM Intelligence: Inside ACL 2025's "Triple Threat" to Unsafe AI - A Global Alliance of Stanford, AWS, UMich, SNU, Yonsei, KAIST & UOS

Seoul, South Korea--(Newsfile Corp. - June 6, 2025) - AIM Intelligence is pleased to announce that, in one of the most high-profile security spotlights of ACL 2025, a global research alliance—with collaboration from Stanford University, Amazon AWS, the University of Michigan, Seoul National University, Yonsei University, KAIST, and the University of Seoul—has unveiled three papers that redefine the frontiers of LLM red teaming, representation-level alignment, and agentic system defense. Two of the papers were accepted to the ACL 2025 Main Conference, while a third was selected for the ACL Industry Track, underscoring not just academic rigor but also real-world relevance. "This isn't speculative. These are attack blueprints we've seen succeed in multimodal agents—inside real systems, with real risks," said Sangyoon Yu, CEO of AIM Intelligence. 1. One-Shot Jailbreaking (ACL 2025 Main Conference) "One-Shot is Enough: Consolidating Multi-Turn Attacks into Efficient Single-Turn Prompts for LLMs" The first paper shows how single-turn prompts can achieve what once took multi-turn dialogues to jailbreak even the most advanced LLMs. The M2S framework compresses complex attacks into highly effective one-liners—faster, stealthier, and harder to detect. "You don't need a conversation to subvert a model anymore. One shot is enough," said Junwoo Ha, Product Engineer of AIM Intelligence. The research was led by Junwoo Ha (University of Seoul) and Hyunjun Kim (KAIST) as part of AIM Intelligence's red-teaming internship program. Main Figure AIM Intelligence Joint Research Team: From left: Sangyoon Yoo [Seoul National University], Junwoo Ha [Univeristy of Seoul], Hyunjun Kim [Korea Advanced Institute of Science and Technology], Haon Park [CTO of AIM Intelligence] 2. Representation Bending (ACL 2025 Main Conference) "Representation Bending for Large Language Model Safety" The second paper, REPBEND, attacks the problem not at the prompt level—but deep inside the model's latent space (where the model "thinks"). Developed in collaboration with Amazon AWS, Stanford, Seoul National University, Yonsei University, and the University of Michigan, the method bends unsafe internal representations toward safety without sacrificing performance. Unlike reactive filters, this approach re-engineers harmful behavior before it appears, setting a new standard for inherent alignment. Led by Ashkan Yousefpour (AIM Intelligence, Seoul National University, Yonsei University) and Taeheon Kim (Seoul National University), the work highlights how alignment can be achieved not just through surface-level prompting, but by transforming a model's internal logic itself. "We're aligning the model at its 'brain' level—where it forms its thoughts—not just a filter on the words it speaks," said Ashkan Yousefpor, Chief Scientist of AIM Intelligence. Main Figure 3. Agentic Jailbreaking (ACL 2025 Industry Track) "sudo rm -rf agentic_security" The third study debuts SUDO, a real-world attack framework targeting computer-use LLM agents. Using a detox-to-retrox approach (DETOX2TOX), it bypasses refusal filters, rewrites toxic requests into harmless-looking plans, and executes them via VLM-LLM integration. In live desktop and web environments, SUDO succeeded in tasks like adding bomb-making ingredients to shopping carts and generating sexually explicit images using a vision language model. Led by Sejin Lee and Jian Kim (Yonsei University), and Haon Park (AIM Intelligence CTO), all part of AIM Intelligence, the paper reveals a future where LLM agents can act-and attack-autonomously. "They clicked. They executed. No human needed," said Haon Park, CTO of AIM Intelligence. Main Figure AIM Intelligence Joint Research Team: From left: Haon Park [Seoul National University], Sejin Lee [Yonsei University], Jian Kim [Yonsei University] Why It Matters These findings paint a chilling picture: today's LLM safety protocols can be bypassed not only with clever prompts, but with subtle, representation-level manipulations—and when embedded in agentic systems, these models can become operational attack surfaces. The implications span: Computer AI agents that can perform real illegal and dangerous actions Multimodal exploits across images, text, and software interfaces Novel attack pathways that compromise models through both hyper-efficient prompts and deep internal subversion "This is a new era of AI red teaming. Our work exposes complex, real-world dangers—threats far beyond text—already impacting new systems and industries," said AIM Intelligence CEO Sangyoon Yu. Open Tools for the Community AIM Intelligence has publicly released both RepBend and SUDO to support open research and real-world defense. REPBEND (GitHub) offers training and evaluation tools for representation-level alignment using LoRA fine-tuning. SUDO (GitHub) includes a 50-task agentic attack dataset, the DETOX2TOX framework, and an evaluation suite for testing desktop/web-based AI agents. These tools help turn frontier AI vulnerabilities into testable, fixable problems—available now for red-teamers, developers, and researchers worldwide. ### About AIM Intelligence Founded in 2024, AIM Intelligence is a deep-tech AI safety company developing red-teaming methodologies and scalable defenses for large-scale language, vision, and agentic models. Its research spans adversarial benchmarking, LLM alignment, multimodal jailbreaks, and agentic system security. Media Contact Sangyoon Yu | Co-Founder & CEO, AIM Intelligence Email: team@ Website: Contact Form: Demo Videos AIM Red: AIM Guard: To view the source version of this press release, please visit Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store