Papers
1–3 of 3Research Paper·Jan 30, 2026
FNF: Functional Network Fingerprint for Large Language Models
The development of large language models (LLMs) is costly and has significant commercial value. Consequently, preventing unauthorized appropriation of open-source LLMs and protecting developers' intel...
7.0 viability
Research Paper·Feb 26, 2026
Obscure but Effective: Classical Chinese Jailbreak Prompt Optimization via Bio-Inspired Search
As Large Language Models (LLMs) are increasingly used, their security risks have drawn increasing attention. Existing research reveals that LLMs are highly susceptible to jailbreak attacks, with effec...
6.0 viability
Research Paper·Mar 2, 2026
DualSentinel: A Lightweight Framework for Detecting Targeted Attacks in Black-box LLM via Dual Entropy Lull Pattern
Recent intelligent systems integrate powerful Large Language Models (LLMs) through APIs, but their trustworthiness may be critically undermined by targeted attacks like backdoor and prompt injection a...
5.0 viability