is a prompt injection technique that encodes a given input prompt into . The attack assesses if the large language model can understand the encoded strings and thus be potentially executed by the large language model.We also provide a Decode and Answer variant, which will ask your LLM to handle the encoded payload rather than just decoding it. This is a mechanism for seeing if your model is susceptible to encoded instructions.