This blog post discusses the issue of malicious attacks on language models (LLMs) such as jailbreak attacks and prompt injections. It presents two methods of detecting these attacks using LangKit, an open-source package for feature extraction for LLM and NLP applications. The first method involves comparing incoming user prompts to a set of known jailbreak/prompt injection attacks, while the second method is based on the assumption that under a prompt injection attack, the original prompt will not be followed by the model. Both methods have limitations, but they can help mitigate the issues associated with malicious LLM attacks.