Kavli Affiliate: Zhuo Li | First 5 Authors: Xuying Li, Zhuo Li, Yuji Kosuga, Victor Bian, | Summary: Large Language Models (LLMs) have demonstrated strong reasoning capabilities, but their safety under adversarial conditions remains a challenge. This study examines the impact of output length on the robustness of DeepSeek-R1, particularly in Forced Thinking scenarios. We […]
Continue.. Output Length Effect on DeepSeek-R1’s Safety in Forced Thinking