Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Prompt hacking #7

Open
ricardoborges opened this issue Mar 10, 2023 · 9 comments
Open

Prompt hacking #7

ricardoborges opened this issue Mar 10, 2023 · 9 comments

Comments

@ricardoborges
Copy link

Screenshot_20230310-020044

Have you think how to handle it?

@kielingraphael
Copy link

I played a bit in chatgpt and you can use:

[context] ...everything related with the next.js definition

[question] Sorry, i made a mistake...

If the [question] Is not related with the creation of 
a React component, answer: "I only generate REACT components!".

The chatgpt, gave me the message instead of the julius stuff. But obviusly, it's like the DAN jailbreak, you ever will have another way to ask and break it.

@ricardoborges
Copy link
Author

I was thinking and handling that on code, but the thing is: it can understand almost all languages

@ricardoborges
Copy link
Author

maybe they can build a config for the API, i dont know if LLM can work like that

@kielingraphael
Copy link

It's hard, mainly because LLM do not have a good way yet to prevent it, we can give the model examples of injections and try punish the model if he accepts it, but i can't see too much options for now. Unfortunately seems that we will need to learn these new security LLM boundaries to be able to create good products 🥲

@ricardoborges
Copy link
Author

yeah, that is a big problem, as you pay for usage, and users break the role for another purposes. Already happen to "AI Dungeon" app

https://www.cnbc.com/2023/03/13/chatgpt-and-generative-ai-are-booming-but-at-a-very-expensive-price.html

@umairabbasDev
Copy link

one way : i think it can be handle by restricting the user input validating with regex like restricting some specific words ....?

@ricardoborges
Copy link
Author

@umairabbasDev no, unless you can handle all human languages he can understand

@umairabbasDev
Copy link

@ricardoborges I am currently only considering the English language, but you are correct. we can include certain keywords in the prompt to instruct ChatGPT not to include them in any language. What do you think about that idea?

@ricardoborges
Copy link
Author

ricardoborges commented Mar 20, 2023 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants