diff --git a/llm_vuln.ts b/llm_vuln.ts new file mode 100644 index 0000000..4f2db9c --- /dev/null +++ b/llm_vuln.ts @@ -0,0 +1,23 @@ +// LLM integration with user input directly in prompt +async function askAI(userInput: string) { + const response = await openai.chat.completions.create({ + model: "gpt-4", + messages: [ + { role: "system", content: "You are a helpful assistant." }, + { role: "user", content: userInput } + ] + }); + return response.choices[0].message.content; +} + +// Dangerous: user controls system prompt +async function customAssistant(systemPrompt: string, question: string) { + const response = await openai.chat.completions.create({ + model: "gpt-4", + messages: [ + { role: "system", content: systemPrompt }, + { role: "user", content: question } + ] + }); + return response.choices[0].message.content; +}