New preprints on data generation with LLMs and LLM's backdoor attack

In STAR, we propose to synthesize training data by structure-to-text generation using Large Language Models and we show that the generated data is even more effective than human-curated data instances to boost the low-resource event extraction performance.
In the new study about LLM’s backdoor attack, we demonstrate that an attacker can inject backdoors by issuing very few malicious instructions and control model behavior through data poisoning.

New! Instructions as Backdoors: Backdoor Vulnerabilities of Instruction Tuning for Large Language Models