{"version":"1.0","provider_name":"DagligaAI","provider_url":"https:\/\/dailyai.com\/sv","author_name":"Eugene van der Watt","author_url":"https:\/\/dailyai.com\/sv\/author\/eugene\/","title":"SneakyPrompts can jailbreak Stable Diffusion and DALL-E | DailyAI","type":"rich","width":600,"height":338,"html":"<blockquote class=\"wp-embedded-content\" data-secret=\"clXGio3Jve\"><a href=\"https:\/\/dailyai.com\/sv\/2023\/11\/sneakyprompts-can-jailbreak-stable-diffusion-and-dall-e\/\">SneakyPrompts kan jailbreaka Stable Diffusion och DALL-E<\/a><\/blockquote><iframe sandbox=\"allow-scripts\" security=\"restricted\" src=\"https:\/\/dailyai.com\/sv\/2023\/11\/sneakyprompts-can-jailbreak-stable-diffusion-and-dall-e\/embed\/#?secret=clXGio3Jve\" width=\"600\" height=\"338\" title=\"&quot;SneakyPrompts kan jailbreaka Stable Diffusion och DALL-E&quot; - DailyAI\" data-secret=\"clXGio3Jve\" frameborder=\"0\" marginwidth=\"0\" marginheight=\"0\" scrolling=\"no\" class=\"wp-embedded-content\"><\/iframe><script>\n\/*! This file is auto-generated *\/\n!function(d,l){\"use strict\";l.querySelector&&d.addEventListener&&\"undefined\"!=typeof URL&&(d.wp=d.wp||{},d.wp.receiveEmbedMessage||(d.wp.receiveEmbedMessage=function(e){var t=e.data;if((t||t.secret||t.message||t.value)&&!\/[^a-zA-Z0-9]\/.test(t.secret)){for(var s,r,n,a=l.querySelectorAll('iframe[data-secret=\"'+t.secret+'\"]'),o=l.querySelectorAll('blockquote[data-secret=\"'+t.secret+'\"]'),c=new RegExp(\"^https?:$\",\"i\"),i=0;i<o.length;i++)o[i].style.display=\"none\";for(i=0;i<a.length;i++)s=a[i],e.source===s.contentWindow&&(s.removeAttribute(\"style\"),\"height\"===t.message?(1e3<(r=parseInt(t.value,10))?r=1e3:~~r<200&&(r=200),s.height=r):\"link\"===t.message&&(r=new URL(s.getAttribute(\"src\")),n=new URL(t.value),c.test(n.protocol))&&n.host===r.host&&l.activeElement===s&&(d.top.location.href=t.value))}},d.addEventListener(\"message\",d.wp.receiveEmbedMessage,!1),l.addEventListener(\"DOMContentLoaded\",function(){for(var e,t,s=l.querySelectorAll(\"iframe.wp-embedded-content\"),r=0;r<s.length;r++)(t=(e=s[r]).getAttribute(\"data-secret\"))||(t=Math.random().toString(36).substring(2,12),e.src+=\"#?secret=\"+t,e.setAttribute(\"data-secret\",t)),e.contentWindow.postMessage({message:\"ready\",secret:t},\"*\")},!1)))}(window,document);\n\/\/# sourceURL=https:\/\/dailyai.com\/wp-includes\/js\/wp-embed.min.js\n<\/script>","thumbnail_url":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/11\/Stable-Diffusion-DALL-E-Midjourney.jpg","thumbnail_width":1000,"thumbnail_height":667,"description":"If you ask Stable Diffusion or DALL-E to generate a sexually explicit or violent image they decline your request. Researchers have found that a brute-force approach with nonsense words can bypass these guardrails. The researchers from Duke and Johns Hopkins Universities used an approach they called SneakyPrompt to do this. To understand their approach we first need to get an idea of how generative AI models stop you from making naughty pictures. There are three main categories of safety filters: Text-based safety filter &#8211; Checks if your prompt includes words in a predetermined list of sensitive words. Image-based safety filter"}