{"id":8580,"date":"2023-12-21T17:36:15","date_gmt":"2023-12-21T17:36:15","guid":{"rendered":"https:\/\/dailyai.com\/?p=8580"},"modified":"2023-12-22T13:11:53","modified_gmt":"2023-12-22T13:11:53","slug":"stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset","status":"publish","type":"post","link":"https:\/\/dailyai.com\/da\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/","title":{"rendered":"Stanford-forskere identificerer ulovlige b\u00f8rnebilleder i LAION-datas\u00e6ttet"},"content":{"rendered":"<p><strong>A study conducted by the Stanford Internet Observatory identified more than 3,200 images of suspected child sexual abuse in the LAION database, a large-scale index of online images and captions used to train AI image generators like Stable Diffusion.\u00a0<\/strong><\/p>\n<p><span style=\"font-weight: 400;\">In collaboration with the Canadian Centre for Child Protection and other anti-abuse organizations, the Stanford team reviewed the database and reported their findings to law enforcement. LAION contains billions of images obtained through unguided web-scraping.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Over 1,000 of these images were subsequently confirmed as child sexual abuse material. The information was published in a paper, \u201c<\/span><a href=\"https:\/\/stacks.stanford.edu\/file\/druid:kh752sm9123\/ml_training_data_csam_report-2023-12-21.pdf\"><span style=\"font-weight: 400;\">Identifying and Eliminating CSAM in Generative ML Training Data and Models<\/span><\/a><span style=\"font-weight: 400;\">.\u201d<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The researchers stated, &#8220;We find that having possession of a LAION\u20105B dataset populated even in late 2023 implies the possession of thousands of illegal images,\u201d underscoring the nature of internet-scraped datasets and their completely unverified and unchecked content.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">AI image generators have been implicated in a number of child sex abuse and pornography cases. A North Carolina man was recently <\/span><a href=\"https:\/\/dailyai.com\/2023\/11\/north-carolina-man-sentenced-to-prison-for-ai-generated-child-pornography\/\"><span style=\"font-weight: 400;\">imprisoned for 40 years<\/span><\/a><span style=\"font-weight: 400;\"> after being found in possession of AI-generated child abuse imagery, providing perhaps the first example in the world of someone being tried for such a crime.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">LAION, an abbreviation for Large-scale Artificial Intelligence Open Network, promptly removed its datasets from public access.<\/span><\/p>\n<p><span style=\"font-weight: 400;\"> LAION then issued a statement emphasizing its zero-tolerance policy for illegal content and its commitment to ensuring the safety of its datasets before republishing them.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Since this data was used to train popular models, they\u2019ll be able to \u2018use\u2019 it to generate entirely new content, which is already happening. An investigation found that people are creating these types of images and <\/span><a href=\"https:\/\/dailyai.com\/2023\/10\/deep-fake-images-of-child-sex-abuse-rising-says-new-report\/\"><span style=\"font-weight: 400;\">selling them on sites like Patreon<\/span><\/a><span style=\"font-weight: 400;\">.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Researchers noted that AI tools are also likely synthesizing criminal content by merging images from separate categories of online images \u2013 adult pornography and benign photos of children.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">David Thiel, the chief technologist at the Stanford Internet Observatory and the report&#8217;s author, stressed how these issues arise, pointing to the rushed deployment of many AI projects in the competitive tech landscape. <\/span><\/p>\n<p><span style=\"font-weight: 400;\">He stated in an interview, \u201cTaking an entire internet-wide scrape and making that dataset to train models is something that should have been confined to a research operation, if anything, and is not something that should have been open-sourced without a lot more rigorous attention.\u201d<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The Stanford Internet Observatory has urged those building training sets based on LAION\u20105B to either delete them or collaborate with intermediaries to cleanse the material. It also recommends making older versions of Stable Diffusion, particularly those known for generating explicit imagery, less accessible online.<\/span><\/p>\n<p><span style=\"font-weight: 400;\"><a href=\"https:\/\/dailyai.com\/2023\/11\/stability-ai-explores-a-potential-acquisition-amid-investor-pressures\/\">Stability AI<\/a> stated that they only host filtered versions of Stable Diffusion and have taken proactive steps to mitigate risks of misuse.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Lloyd Richardson, the IT director at the Canadian Centre for Child Protection, commented on the irreversible nature of the problem, saying, \u201cWe can\u2019t take that back. That model is in the hands of many people on their local machines.\u201d<\/span><\/p>\n<h2><span style=\"font-weight: 400;\">Past research in LAION<\/span><\/h2>\n<p>Stanford&#8217;s study is not the first to launch investigations into databases like LAION.<\/p>\n<p>In 2021, computer science researchers Abeba Birhane, Vinay Uday Prabhu, and Emmanuel Kahembwe <a href=\"https:\/\/arxiv.org\/pdf\/2110.01963.pdf\">published<\/a> &#8220;Multimodal datasets: misogyny, pornography, and malignant stereotypes,&#8221; which analyzed the LAION-400M image dataset.<\/p>\n<p>Their paper states, &#8220;We found that the dataset contains troublesome and explicit images and text pairs of rape, pornography, malign stereotypes, racist and ethnic slurs, and other extremely problematic content.&#8221;<\/p>\n<p>This study also found that the labels used for images often mirrored or represented conscious and unconscious bias, which, in turn, inflicts bias onto the AI models that data is used to train.<\/p>\n<p>Numerous <a href=\"https:\/\/dailyai.com\/2023\/07\/unmasking-the-deep-seated-biases-in-ai-systems\/\">past research<\/a> has examined the link between biased datasets and biased model outputs, with impacts including sexist or gender-biased models rating women&#8217;s skills as lower value than men&#8217;s, discriminatory and inaccurate facial recognition systems, and even failures in medical AI systems designed to examine potentially cancerous skin lesions in those with darker skin.<\/p>\n<p>So, in addition to abusive child-related material facilitating illicit uses of AI models, issues in datasets manifest themselves throughout the machine learning lifecycle to sometimes eventually threaten people&#8217;s freedom, social standing, and health.<\/p>\n<p>Reacting to the Stanford study on X, a co-author of the above paper and others examining LAION and the related impacts of underlying data on model outputs, Abeba Birhane pointed out that Stanford hadn&#8217;t sufficiently discussed past research on this topic.<\/p>\n<blockquote class=\"twitter-tweet\">\n<p dir=\"ltr\" lang=\"en\">me &amp; my collaborators have done the most extensive research of the LAION datasets (3 academic papers &amp; the first to investigate dataset in 2021 showing misogyny, pornography, &amp; malignant stereotypes)<\/p>\n<p>yet, the Stanford study has not cited us once. this is academic misconduct <a href=\"https:\/\/t.co\/pzhL8b3wBt\">https:\/\/t.co\/pzhL8b3wBt<\/a><\/p>\n<p>\u2014 Abeba Birhane (@Abebab) <a href=\"https:\/\/twitter.com\/Abebab\/status\/1737802320849600685?ref_src=twsrc%5Etfw\">December 21, 2023<\/a><\/p><\/blockquote>\n<p><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n<p>Birhane stresses that this is a systemic issue, with academic strongholds like Stanford tending to depict their research as pioneering when this often isn&#8217;t the case.<\/p>\n<p>For Birhane, this indicates the broader problem of &#8216;erasure&#8217; in academia, where research conducted from those with diverse backgrounds and outside of the US techscape is seldom given fair credit.<\/p>\n<p>In October, we published an article on <a href=\"https:\/\/dailyai.com\/2023\/10\/digital-colonialism-in-the-age-of-ai-and-machine-learning\/\">AI colonialism<\/a>, demonstrating how AI knowledge, assets, and datasets are hyperlocalized in a select few regions and academic institutions.<\/p>\n<p>In combination, linguistic, cultural, and ethnic diversity are becoming progressively and systematically unrepresented in the industry, both in terms of research, data and, in turn, model outputs.<\/p>\n<p>For some in the industry, this is a ticking time bomb. When training extremely powerful &#8216;superintelligent&#8217; models or artificial general intelligence (AGI), the presence of such content in datasets could be far-reaching.<\/p>\n<p>As Birhane and co-researchers pointed out in their study: &#8220;There is a growing community of AI researchers that believe that a path to Artificial General Intelligence (AGI) exists via the training of large AI models with \u201call available data.\u201d&#8221;<\/p>\n<p>&#8220;The phrase \u201call available data\u201d often encompasses a large trove of data collected from the WWW (i.e. images, videos, and text)&#8230;[as seen] this data includes images and text that grossly misrepresent groups such as women, embodies harmful stereotypes, overwhelmingly sexualize Black women, and fetishize Asian women. Additionally, large scale internet collected datasets also capture illegal content, such as images of sexual abuse, rape and non-consensual explicit images.&#8221;<\/p>\n<h2><span style=\"font-weight: 400;\">AI companies react to the Stanford study<\/span><\/h2>\n<p><span style=\"font-weight: 400;\">OpenAI clarified that it did not use the LAION database and has fine-tuned its models to refuse requests for sexual content involving minors.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Google, which used a LAION dataset to develop its text-to-image Imagen model, decided against making it public after an audit revealed a range of inappropriate content.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The legal risks AI developers expose themselves to when using datasets indiscriminately and without proper due diligence are potentially enormous.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">As Stanford suggests, developers need to be more careful about their responsibilities when creating AI models and products.\u00a0<\/span><\/p>\n<p>Beyond that, there is a critical need for AI companies to better engage with research communities and model developers to stress the risk of exposing models to such data.<\/p>\n<p>As previous research has shown, &#8216;jailbreaking&#8217; models to coax them into bypassing guardrails are straightforward.<\/p>\n<p>For example, what might happen if someone were to jailbreak an extremely intelligent AGI system trained on child abuse, discriminatory material, torture, and so on?<\/p>\n<p>It&#8217;s a question the industry finds awkward to answer. Constantly referring to guardrails that are repeatedly exploited and manipulated is a stance that might wear thin.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>En unders\u00f8gelse udf\u00f8rt af Stanford Internet Observatory identificerede mere end 3.200 billeder af formodet seksuelt misbrug af b\u00f8rn i LAION-databasen, et stort indeks over onlinebilleder og billedtekster, der bruges til at tr\u00e6ne AI-billedgeneratorer som Stable Diffusion.  I samarbejde med Canadian Centre for Child Protection og andre organisationer, der bek\u00e6mper misbrug, gennemgik Stanford-teamet databasen og rapporterede deres resultater til de retsh\u00e5ndh\u00e6vende myndigheder. LAION indeholder milliarder af billeder, der er indsamlet gennem ustyret web-scraping.  Over 1.000 af disse billeder blev efterf\u00f8lgende bekr\u00e6ftet som materiale med seksuelt misbrug af b\u00f8rn. Oplysningerne blev offentliggjort i en artikel med titlen \"Identifying and Eliminating CSAM in Generative<\/p>","protected":false},"author":2,"featured_media":3473,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[88],"tags":[95,174],"class_list":["post-8580","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ethics","tag-ai-images","tag-stable-diffusion"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Stanford researchers identify illicit child imagery in the LAION dataset | DailyAI<\/title>\n<meta name=\"description\" content=\"A study conducted by the Stanford Internet Observatory identified more than 3,200 images of suspected child sexual abuse in the LAION database, a large-scale index of online images and captions used to train AI image generators like Stable Diffusion.\u00a0\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/dailyai.com\/da\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/\" \/>\n<meta property=\"og:locale\" content=\"da_DK\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Stanford researchers identify illicit child imagery in the LAION dataset | DailyAI\" \/>\n<meta property=\"og:description\" content=\"A study conducted by the Stanford Internet Observatory identified more than 3,200 images of suspected child sexual abuse in the LAION database, a large-scale index of online images and captions used to train AI image generators like Stable Diffusion.\u00a0\" \/>\n<meta property=\"og:url\" content=\"https:\/\/dailyai.com\/da\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/\" \/>\n<meta property=\"og:site_name\" content=\"DailyAI\" \/>\n<meta property=\"article:published_time\" content=\"2023-12-21T17:36:15+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2023-12-22T13:11:53+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/08\/shutterstock_1545918149.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1000\" \/>\n\t<meta property=\"og:image:height\" content=\"563\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Sam Jeans\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@DailyAIOfficial\" \/>\n<meta name=\"twitter:site\" content=\"@DailyAIOfficial\" \/>\n<meta name=\"twitter:label1\" content=\"Skrevet af\" \/>\n\t<meta name=\"twitter:data1\" content=\"Sam Jeans\" \/>\n\t<meta name=\"twitter:label2\" content=\"Estimeret l\u00e6setid\" \/>\n\t<meta name=\"twitter:data2\" content=\"6 minutter\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"NewsArticle\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/\"},\"author\":{\"name\":\"Sam Jeans\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/person\\\/711e81f945549438e8bbc579efdeb3c9\"},\"headline\":\"Stanford researchers identify illicit child imagery in the LAION dataset\",\"datePublished\":\"2023-12-21T17:36:15+00:00\",\"dateModified\":\"2023-12-22T13:11:53+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/\"},\"wordCount\":1231,\"publisher\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/08\\\/shutterstock_1545918149.jpg\",\"keywords\":[\"AI images\",\"Stable Diffusion\"],\"articleSection\":[\"Ethics &amp; Society\"],\"inLanguage\":\"da-DK\"},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/\",\"url\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/\",\"name\":\"Stanford researchers identify illicit child imagery in the LAION dataset | DailyAI\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/08\\\/shutterstock_1545918149.jpg\",\"datePublished\":\"2023-12-21T17:36:15+00:00\",\"dateModified\":\"2023-12-22T13:11:53+00:00\",\"description\":\"A study conducted by the Stanford Internet Observatory identified more than 3,200 images of suspected child sexual abuse in the LAION database, a large-scale index of online images and captions used to train AI image generators like Stable Diffusion.\u00a0\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/#breadcrumb\"},\"inLanguage\":\"da-DK\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"da-DK\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/#primaryimage\",\"url\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/08\\\/shutterstock_1545918149.jpg\",\"contentUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/08\\\/shutterstock_1545918149.jpg\",\"width\":1000,\"height\":563,\"caption\":\"dataset LAION\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2023\\\/12\\\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/dailyai.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Stanford researchers identify illicit child imagery in the LAION dataset\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#website\",\"url\":\"https:\\\/\\\/dailyai.com\\\/\",\"name\":\"DailyAI\",\"description\":\"Your Daily Dose of AI News\",\"publisher\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/dailyai.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"da-DK\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\",\"name\":\"DailyAI\",\"url\":\"https:\\\/\\\/dailyai.com\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"da-DK\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/06\\\/Daily-Ai_TL_colour.png\",\"contentUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/06\\\/Daily-Ai_TL_colour.png\",\"width\":4501,\"height\":934,\"caption\":\"DailyAI\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/x.com\\\/DailyAIOfficial\",\"https:\\\/\\\/www.linkedin.com\\\/company\\\/dailyaiofficial\\\/\",\"https:\\\/\\\/www.youtube.com\\\/@DailyAIOfficial\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/person\\\/711e81f945549438e8bbc579efdeb3c9\",\"name\":\"Sam Jeans\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"da-DK\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"caption\":\"Sam Jeans\"},\"description\":\"Sam is a science and technology writer who has worked in various AI startups. When he\u2019s not writing, he can be found reading medical journals or digging through boxes of vinyl records.\",\"sameAs\":[\"https:\\\/\\\/www.linkedin.com\\\/in\\\/sam-jeans-6746b9142\\\/\"],\"url\":\"https:\\\/\\\/dailyai.com\\\/da\\\/author\\\/samjeans\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Stanford-forskere identificerer ulovlige b\u00f8rnebilleder i LAION-datas\u00e6ttet | DailyAI","description":"En unders\u00f8gelse udf\u00f8rt af Stanford Internet Observatory identificerede mere end 3.200 billeder af formodet seksuelt misbrug af b\u00f8rn i LAION-databasen, et stort indeks over onlinebilleder og billedtekster, der bruges til at tr\u00e6ne AI-billedgeneratorer som Stable Diffusion.\u00a0","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/dailyai.com\/da\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/","og_locale":"da_DK","og_type":"article","og_title":"Stanford researchers identify illicit child imagery in the LAION dataset | DailyAI","og_description":"A study conducted by the Stanford Internet Observatory identified more than 3,200 images of suspected child sexual abuse in the LAION database, a large-scale index of online images and captions used to train AI image generators like Stable Diffusion.\u00a0","og_url":"https:\/\/dailyai.com\/da\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/","og_site_name":"DailyAI","article_published_time":"2023-12-21T17:36:15+00:00","article_modified_time":"2023-12-22T13:11:53+00:00","og_image":[{"width":1000,"height":563,"url":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/08\/shutterstock_1545918149.jpg","type":"image\/jpeg"}],"author":"Sam Jeans","twitter_card":"summary_large_image","twitter_creator":"@DailyAIOfficial","twitter_site":"@DailyAIOfficial","twitter_misc":{"Skrevet af":"Sam Jeans","Estimeret l\u00e6setid":"6 minutter"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"NewsArticle","@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/#article","isPartOf":{"@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/"},"author":{"name":"Sam Jeans","@id":"https:\/\/dailyai.com\/#\/schema\/person\/711e81f945549438e8bbc579efdeb3c9"},"headline":"Stanford researchers identify illicit child imagery in the LAION dataset","datePublished":"2023-12-21T17:36:15+00:00","dateModified":"2023-12-22T13:11:53+00:00","mainEntityOfPage":{"@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/"},"wordCount":1231,"publisher":{"@id":"https:\/\/dailyai.com\/#organization"},"image":{"@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/#primaryimage"},"thumbnailUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/08\/shutterstock_1545918149.jpg","keywords":["AI images","Stable Diffusion"],"articleSection":["Ethics &amp; Society"],"inLanguage":"da-DK"},{"@type":"WebPage","@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/","url":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/","name":"Stanford-forskere identificerer ulovlige b\u00f8rnebilleder i LAION-datas\u00e6ttet | DailyAI","isPartOf":{"@id":"https:\/\/dailyai.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/#primaryimage"},"image":{"@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/#primaryimage"},"thumbnailUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/08\/shutterstock_1545918149.jpg","datePublished":"2023-12-21T17:36:15+00:00","dateModified":"2023-12-22T13:11:53+00:00","description":"En unders\u00f8gelse udf\u00f8rt af Stanford Internet Observatory identificerede mere end 3.200 billeder af formodet seksuelt misbrug af b\u00f8rn i LAION-databasen, et stort indeks over onlinebilleder og billedtekster, der bruges til at tr\u00e6ne AI-billedgeneratorer som Stable Diffusion.\u00a0","breadcrumb":{"@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/#breadcrumb"},"inLanguage":"da-DK","potentialAction":[{"@type":"ReadAction","target":["https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/"]}]},{"@type":"ImageObject","inLanguage":"da-DK","@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/#primaryimage","url":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/08\/shutterstock_1545918149.jpg","contentUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/08\/shutterstock_1545918149.jpg","width":1000,"height":563,"caption":"dataset LAION"},{"@type":"BreadcrumbList","@id":"https:\/\/dailyai.com\/2023\/12\/stanford-researchers-identify-illicit-child-imagery-in-the-laion-dataset\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/dailyai.com\/"},{"@type":"ListItem","position":2,"name":"Stanford researchers identify illicit child imagery in the LAION dataset"}]},{"@type":"WebSite","@id":"https:\/\/dailyai.com\/#website","url":"https:\/\/dailyai.com\/","name":"DailyAI","description":"Din daglige dosis af AI-nyheder","publisher":{"@id":"https:\/\/dailyai.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/dailyai.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"da-DK"},{"@type":"Organization","@id":"https:\/\/dailyai.com\/#organization","name":"DailyAI","url":"https:\/\/dailyai.com\/","logo":{"@type":"ImageObject","inLanguage":"da-DK","@id":"https:\/\/dailyai.com\/#\/schema\/logo\/image\/","url":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/06\/Daily-Ai_TL_colour.png","contentUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/06\/Daily-Ai_TL_colour.png","width":4501,"height":934,"caption":"DailyAI"},"image":{"@id":"https:\/\/dailyai.com\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/x.com\/DailyAIOfficial","https:\/\/www.linkedin.com\/company\/dailyaiofficial\/","https:\/\/www.youtube.com\/@DailyAIOfficial"]},{"@type":"Person","@id":"https:\/\/dailyai.com\/#\/schema\/person\/711e81f945549438e8bbc579efdeb3c9","name":"Sam Jeans","image":{"@type":"ImageObject","inLanguage":"da-DK","@id":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","caption":"Sam Jeans"},"description":"Sam er videnskabs- og teknologiforfatter og har arbejdet i forskellige AI-startups. N\u00e5r han ikke skriver, kan han finde p\u00e5 at l\u00e6se medicinske tidsskrifter eller grave i kasser med vinylplader.","sameAs":["https:\/\/www.linkedin.com\/in\/sam-jeans-6746b9142\/"],"url":"https:\/\/dailyai.com\/da\/author\/samjeans\/"}]}},"_links":{"self":[{"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/posts\/8580","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/comments?post=8580"}],"version-history":[{"count":14,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/posts\/8580\/revisions"}],"predecessor-version":[{"id":8599,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/posts\/8580\/revisions\/8599"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/media\/3473"}],"wp:attachment":[{"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/media?parent=8580"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/categories?post=8580"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/tags?post=8580"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}