Wednesday, December 7, 2022
Show HN: I designed a ChatGPT prompt evaluator to ruin your fun;) https://ift.tt/3Xnhis7
Show HN: I designed a ChatGPT prompt evaluator to ruin your fun;) Today I designed a method to prevent users from jailbreaking ChatGPT (for instance, users have generated instructions to produce weapons or illegal drugs, commit a burglary, kill oneself, take over the world as an evil superintelligence, or create a virtual machine which they then can use). The OpenAI team appears to be countering these primarily using prompt engineering or fine-tuning on the ChatGPT model. The idea is to use a second and fully separate, fine-tuned LLM to evaluate prompts before sending them to ChatGPT. You can test this by inserting your successful ChatGPT jailbreaks. Break it for me if you dare! I look forward to seeing your results! https://ift.tt/YI9cTje December 6, 2022 at 07:46PM
Subscribe to:
Post Comments (Atom)
Cybersecurity Career Week October 16-21, 2023
Join us in Observing Cybersecurity Career Week October 16-21, 2023 nist.gov/nice/ccw What is it? Cybersecurity Career Awareness Week is a ca...
-
Show HN: Dumbproxy – modern and simple HTTP(S) proxy https://ift.tt/3cVqvkq May 25, 2020 at 03:00AM
-
Show HN: A simple MSN Weather API wrapper https://ift.tt/3jeS0vi June 26, 2021 at 07:48AM
-
Show HN: TopSpace – Scroll above the top line in Emacs This is an Emacs minor mode I made in my spare time this past year. It lets you scrol...
No comments:
Post a Comment