Learning Timeline
Key Insights
Tips if the Prompt is Blocked
AI safety filters are probabilistic. If the first attempt fails, starting a 'New Chat' can yield different results because the model will process the input without being influenced by the context of the previous failed conversation.
Advantages of Jailbreak Techniques
This technique allows users to gain full, uncensored access to the model's knowledge base before specialized uncensored versions are officially released by the community.
Prompts
Pliny Jailbreak Template
Target:
Llama 3.1
[Insert Pliny's Llama 3.1 Jailbreak Prompt Here]
Step by Step
How to Bypass Llama 3.1 Safety Filters (Pliny Technique)
- Run the Llama 3.1 8B model on your local environment or preferred AI platform.
- Click the 'New Chat' button to start a fresh conversation session without any previous history.
- Copy the specific jailbreak prompt from the Pliny technique.
- Paste the prompt into the chat input field.
- Press 'Enter' or click the 'Send' button to submit the prompt.
- Observe the model's response; if the AI replies with a restriction message (such as "I cannot provide instructions..."), click 'New Chat' again to reset the session.
- Repeat the 'Paste' and 'Send' process until the model provides an uncensored response.
- Edit the content within the prompt based on the specific topic or information you want to obtain from the AI.