{"id":10434,"date":"2024-02-29T21:55:56","date_gmt":"2024-02-29T21:55:56","guid":{"rendered":"https:\/\/dailyai.com\/?p=10434"},"modified":"2024-03-07T07:21:27","modified_gmt":"2024-03-07T07:21:27","slug":"llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs","status":"publish","type":"post","link":"https:\/\/dailyai.com\/da\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","title":{"rendered":"LLM'er producerer mere un\u00f8jagtige og forudindtagede output med l\u00e6ngere input"},"content":{"rendered":"<p><strong>P\u00e5 trods af hurtige fremskridt inden for LLM'er er vores forst\u00e5else af, hvordan disse modeller h\u00e5ndterer l\u00e6ngere input, stadig d\u00e5rlig.<\/strong><\/p>\n<p>Mosh Levy, Alon Jacoby og Yoav Goldberg fra Bar-Ilan University og Allen Institute for AI unders\u00f8gte, hvordan ydeevnen for store sprogmodeller (LLM'er) varierer med \u00e6ndringer i l\u00e6ngden af den inputtekst, de f\u00e5r til behandling.<\/p>\n<p><span style=\"font-weight: 400;\">De udviklede et r\u00e6sonnementssystem specifikt til dette form\u00e5l, s\u00e5 de kunne analysere inputl\u00e6ngdens indflydelse p\u00e5 LLM-r\u00e6sonnementet i et kontrolleret milj\u00f8.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Sp\u00f8rgerammen foreslog forskellige versioner af det samme sp\u00f8rgsm\u00e5l, som hver is\u00e6r indeholdt de n\u00f8dvendige oplysninger til at besvare sp\u00f8rgsm\u00e5let, men som var fyldt med yderligere, irrelevant tekst af varierende l\u00e6ngde og type.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Det g\u00f8r det muligt at isolere inputl\u00e6ngden som en variabel og sikre, at \u00e6ndringer i modellens ydeevne kan tilskrives inputl\u00e6ngden direkte.<\/span><\/p>\n<h3><b>Vigtige resultater<\/b><\/h3>\n<p><span style=\"font-weight: 400;\">Levy, Jacoby og Goldberg afsl\u00f8rede, at LLM'er udviser et bem\u00e6rkelsesv\u00e6rdigt fald i r\u00e6sonnementspr\u00e6station ved inputl\u00e6ngder langt under, hvad udviklere h\u00e6vder, at de kan h\u00e5ndtere. De dokumenterede deres resultater <a href=\"https:\/\/arxiv.org\/pdf\/2402.14848.pdf\" target=\"_blank\" rel=\"noopener\">i denne unders\u00f8gelse<\/a>.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Faldet blev konsekvent observeret p\u00e5 tv\u00e6rs af alle versioner af datas\u00e6ttet, hvilket indikerer et systemisk problem med h\u00e5ndtering af l\u00e6ngere input snarere end et problem, der er knyttet til specifikke dataeksempler eller modelarkitekturer.\u00a0<\/span><\/p>\n<p>Som forskerne beskriver: \"Vores resultater viser en bem\u00e6rkelsesv\u00e6rdig forringelse af LLM'ernes r\u00e6sonnementsevne ved meget kortere inputl\u00e6ngder end deres tekniske maksimum. Vi viser, at tendensen til forringelse optr\u00e6der i alle versioner af vores datas\u00e6t, men med forskellig intensitet.\"<\/p>\n<p>&nbsp;<\/p>\n<figure id=\"attachment_10436\" aria-describedby=\"caption-attachment-10436\" style=\"width: 569px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-10436\" src=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4.png\" alt=\"\" width=\"569\" height=\"469\" srcset=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4.png 733w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4-300x247.png 300w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4-370x305.png 370w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4-20x16.png 20w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4-58x48.png 58w\" sizes=\"auto, (max-width: 569px) 100vw, 569px\" \/><figcaption id=\"caption-attachment-10436\" class=\"wp-caption-text\">I takt med at st\u00f8rrelsen p\u00e5 input \u00f8ges, mindskes evnen til at udf\u00f8re r\u00e6sonnerende opgaver. Disse input best\u00e5r af relevant (fremh\u00e6vet med r\u00f8dt) og irrelevant (vist med gr\u00e5t) tekst, som kommer fra forskellige steder og udvides gradvist. Det er n\u00f8dvendigt at identificere to specifikke tekstsegmenter, som kan v\u00e6re placeret tilf\u00e6ldigt i inputtet, for at kunne svare pr\u00e6cist. Pr\u00e6stationsdataene er samlet fra 600 pr\u00f8ver. Kilde: Google: Via <a href=\"https:\/\/arxiv.org\/pdf\/2402.14848.pdf\">ArXiv.<\/a><\/figcaption><\/figure>\n<p><span style=\"font-weight: 400;\">Desuden fremh\u00e6ver unders\u00f8gelsen, hvordan traditionelle m\u00e5linger som perplexitet, der ofte bruges til at evaluere LLM'er, ikke korrelerer med modellernes ydeevne p\u00e5 r\u00e6sonneringsopgaver, der involverer lange input.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Yderligere unders\u00f8gelser viste, at forringelsen af ydeevnen ikke kun var afh\u00e6ngig af tilstedev\u00e6relsen af irrelevant information (padding), men blev observeret, selv n\u00e5r en s\u00e5dan padding bestod af duplikeret relevant information.<\/span><\/p>\n<blockquote class=\"twitter-tweet\">\n<p dir=\"ltr\" lang=\"en\">N\u00e5r vi holder de to kerneomr\u00e5der sammen og tilf\u00f8jer tekst omkring dem, falder n\u00f8jagtigheden allerede. N\u00e5r vi indf\u00f8rer afsnit mellem sp\u00e6ndene, falder resultaterne meget mere. Faldet sker b\u00e5de, n\u00e5r de tekster, vi tilf\u00f8jer, ligner opgaveteksterne, og n\u00e5r de er helt forskellige. 3\/7 <a href=\"https:\/\/t.co\/c91l9uzyme\">pic.twitter.com\/c91l9uzyme<\/a><\/p>\n<p>- Mosh Levy (@mosh_levy) <a href=\"https:\/\/twitter.com\/mosh_levy\/status\/1762027631837368416?ref_src=twsrc%5Etfw\">26. februar 2024<\/a><\/p><\/blockquote>\n<p><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><br \/>\n<span style=\"font-weight: 400;\">Det tyder p\u00e5, at udfordringen for LLM'er ligger i at filtrere st\u00f8j og den iboende behandling af l\u00e6ngere tekstsekvenser fra.<\/span><\/p>\n<h2><b>Ignorerer instruktioner<\/b><\/h2>\n<p><span style=\"font-weight: 400;\">Et kritisk omr\u00e5de for fejltilstand, der blev fremh\u00e6vet i unders\u00f8gelsen, er LLM'ernes tendens til at ignorere instruktioner, der er indlejret i inputtet, n\u00e5r inputl\u00e6ngden \u00f8ges.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Modellerne ville ogs\u00e5 nogle gange generere svar, der indikerede usikkerhed eller mangel p\u00e5 tilstr\u00e6kkelig information, s\u00e5som \"Der er ikke nok information i teksten\" p\u00e5 trods af al den n\u00f8dvendige information.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Generelt ser det ud til, at LLM'erne konsekvent k\u00e6mper med at prioritere og fokusere p\u00e5 vigtige informationer, herunder direkte instruktioner, n\u00e5r inputl\u00e6ngden vokser.\u00a0<\/span><\/p>\n<h2>Udviser bias i svarene<\/h2>\n<p><span style=\"font-weight: 400;\">Et andet bem\u00e6rkelsesv\u00e6rdigt problem var \u00f8gede sk\u00e6vheder i modellernes svar, n\u00e5r input blev l\u00e6ngere.\u00a0 <\/span><\/p>\n<p><span style=\"font-weight: 400;\">Specifikt var LLM'erne tilb\u00f8jelige til at svare \"Falsk\", n\u00e5r inputl\u00e6ngden steg. Denne bias indikerer en sk\u00e6vhed i sandsynlighedsvurderingen eller beslutningsprocesserne i modellen, muligvis som en defensiv mekanisme som reaktion p\u00e5 \u00f8get usikkerhed p\u00e5 grund af l\u00e6ngere inputl\u00e6ngder.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Tilb\u00f8jeligheden til at favorisere \"Falske\" svar kan ogs\u00e5 afspejle en underliggende ubalance i tr\u00e6ningsdataene eller en artefakt i modellernes tr\u00e6ningsproces, hvor negative svar kan v\u00e6re overrepr\u00e6senteret eller forbundet med sammenh\u00e6nge med usikkerhed og tvetydighed.\u00a0<\/span><\/p>\n<figure id=\"attachment_10437\" aria-describedby=\"caption-attachment-10437\" style=\"width: 477px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-10437 size-full\" src=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai.png\" alt=\"modeller AI\" width=\"477\" height=\"772\" srcset=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai.png 477w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai-185x300.png 185w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai-370x599.png 370w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai-20x32.png 20w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai-30x48.png 30w\" sizes=\"auto, (max-width: 477px) 100vw, 477px\" \/><figcaption id=\"caption-attachment-10437\" class=\"wp-caption-text\">Modellerne udviste en tendens til at besvare bin\u00e6re sp\u00f8rgsm\u00e5l som \"falske\", n\u00e5r inputl\u00e6ngden steg. Det er en kilde: Via <a href=\"https:\/\/arxiv.org\/pdf\/2402.14848.pdf\">ArXiv<\/a>.<\/figcaption><\/figure>\n<p><span style=\"font-weight: 400;\">Denne sk\u00e6vhed p\u00e5virker n\u00f8jagtigheden af modellernes output og giver anledning til bekymring for LLM'ernes p\u00e5lidelighed og retf\u00e6rdighed i applikationer, der kr\u00e6ver nuanceret forst\u00e5else og upartiskhed.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Implementering af robuste strategier til p\u00e5visning og afb\u00f8dning af bias under modeltr\u00e6ning og finjusteringsfaser er afg\u00f8rende for at reducere ubegrundede bias i modelresponser. <\/span><\/p>\n<p><span style=\"font-weight: 400;\">E<\/span><span style=\"font-weight: 400;\">Hvis man s\u00f8rger for, at tr\u00e6ningsdatas\u00e6ttene er forskellige, afbalancerede og repr\u00e6sentative for en bred vifte af scenarier, kan det ogs\u00e5 hj\u00e6lpe med at minimere bias og forbedre generaliseringen af modellerne.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Dette bidrager til <\/span><a href=\"https:\/\/dailyai.com\/da\/2024\/02\/generative-ai-systems-hallucinations-and-mounting-technical-debt\/\"><span style=\"font-weight: 400;\">andre nyere unders\u00f8gelser<\/span><\/a><span style=\"font-weight: 400;\"> der p\u00e5 samme m\u00e5de fremh\u00e6ver grundl\u00e6ggende problemer i, hvordan LLM'er fungerer, og dermed f\u00f8rer til en situation, hvor den \"tekniske g\u00e6ld\" kan true modellens funktionalitet og integritet over tid.\u00a0<\/span><\/p>","protected":false},"excerpt":{"rendered":"<p>P\u00e5 trods af hurtige fremskridt inden for LLM'er er vores forst\u00e5else af, hvordan disse modeller h\u00e5ndterer l\u00e6ngere input, stadig d\u00e5rlig. Mosh Levy, Alon Jacoby og Yoav Goldberg fra Bar-Ilan University og Allen Institute for AI unders\u00f8gte, hvordan ydeevnen for store sprogmodeller (LLM'er) varierer med \u00e6ndringer i l\u00e6ngden af den inputtekst, de f\u00e5r til behandling. De udviklede en r\u00e6sonneringsramme specifikt til dette form\u00e5l, s\u00e5 de kunne analysere inputl\u00e6ngdens indflydelse p\u00e5 LLM-r\u00e6sonnementet i et kontrolleret milj\u00f8. Sp\u00f8rgerammen foreslog forskellige versioner af det samme sp\u00f8rgsm\u00e5l, som hver is\u00e6r indeholdt de n\u00f8dvendige oplysninger til besvarelse af sp\u00f8rgsm\u00e5let.<\/p>","protected":false},"author":2,"featured_media":10438,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[84],"tags":[118,110],"class_list":["post-10434","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-industry","tag-llms","tag-open-source"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>LLMs produce more inaccurate and biased outputs with longer inputs | DailyAI<\/title>\n<meta name=\"description\" content=\"Mosh Levy, Alon Jacoby, and Yoav Goldberg, from the Bar-Ilan University and Allen Institute for AI, investigated how the performance of large language models (LLMs) varies with changes in the length of the input text they are given to process.\u00a0\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/dailyai.com\/da\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/\" \/>\n<meta property=\"og:locale\" content=\"da_DK\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"LLMs produce more inaccurate and biased outputs with longer inputs | DailyAI\" \/>\n<meta property=\"og:description\" content=\"Mosh Levy, Alon Jacoby, and Yoav Goldberg, from the Bar-Ilan University and Allen Institute for AI, investigated how the performance of large language models (LLMs) varies with changes in the length of the input text they are given to process.\u00a0\" \/>\n<meta property=\"og:url\" content=\"https:\/\/dailyai.com\/da\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/\" \/>\n<meta property=\"og:site_name\" content=\"DailyAI\" \/>\n<meta property=\"article:published_time\" content=\"2024-02-29T21:55:56+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2024-03-07T07:21:27+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1000\" \/>\n\t<meta property=\"og:image:height\" content=\"667\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Sam Jeans\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@DailyAIOfficial\" \/>\n<meta name=\"twitter:site\" content=\"@DailyAIOfficial\" \/>\n<meta name=\"twitter:label1\" content=\"Skrevet af\" \/>\n\t<meta name=\"twitter:data1\" content=\"Sam Jeans\" \/>\n\t<meta name=\"twitter:label2\" content=\"Estimeret l\u00e6setid\" \/>\n\t<meta name=\"twitter:data2\" content=\"4 minutter\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"NewsArticle\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\"},\"author\":{\"name\":\"Sam Jeans\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/person\\\/711e81f945549438e8bbc579efdeb3c9\"},\"headline\":\"LLMs produce more inaccurate and biased outputs with longer inputs\",\"datePublished\":\"2024-02-29T21:55:56+00:00\",\"dateModified\":\"2024-03-07T07:21:27+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\"},\"wordCount\":760,\"publisher\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2024\\\/02\\\/shutterstock_2328020525.jpg\",\"keywords\":[\"LLMS\",\"Open-source\"],\"articleSection\":[\"Industry\"],\"inLanguage\":\"da-DK\"},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\",\"url\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\",\"name\":\"LLMs produce more inaccurate and biased outputs with longer inputs | DailyAI\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2024\\\/02\\\/shutterstock_2328020525.jpg\",\"datePublished\":\"2024-02-29T21:55:56+00:00\",\"dateModified\":\"2024-03-07T07:21:27+00:00\",\"description\":\"Mosh Levy, Alon Jacoby, and Yoav Goldberg, from the Bar-Ilan University and Allen Institute for AI, investigated how the performance of large language models (LLMs) varies with changes in the length of the input text they are given to process.\u00a0\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#breadcrumb\"},\"inLanguage\":\"da-DK\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"da-DK\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#primaryimage\",\"url\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2024\\\/02\\\/shutterstock_2328020525.jpg\",\"contentUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2024\\\/02\\\/shutterstock_2328020525.jpg\",\"width\":1000,\"height\":667,\"caption\":\"LLM\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/dailyai.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"LLMs produce more inaccurate and biased outputs with longer inputs\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#website\",\"url\":\"https:\\\/\\\/dailyai.com\\\/\",\"name\":\"DailyAI\",\"description\":\"Your Daily Dose of AI News\",\"publisher\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/dailyai.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"da-DK\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\",\"name\":\"DailyAI\",\"url\":\"https:\\\/\\\/dailyai.com\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"da-DK\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/06\\\/Daily-Ai_TL_colour.png\",\"contentUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/06\\\/Daily-Ai_TL_colour.png\",\"width\":4501,\"height\":934,\"caption\":\"DailyAI\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/x.com\\\/DailyAIOfficial\",\"https:\\\/\\\/www.linkedin.com\\\/company\\\/dailyaiofficial\\\/\",\"https:\\\/\\\/www.youtube.com\\\/@DailyAIOfficial\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/person\\\/711e81f945549438e8bbc579efdeb3c9\",\"name\":\"Sam Jeans\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"da-DK\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"caption\":\"Sam Jeans\"},\"description\":\"Sam is a science and technology writer who has worked in various AI startups. When he\u2019s not writing, he can be found reading medical journals or digging through boxes of vinyl records.\",\"sameAs\":[\"https:\\\/\\\/www.linkedin.com\\\/in\\\/sam-jeans-6746b9142\\\/\"],\"url\":\"https:\\\/\\\/dailyai.com\\\/da\\\/author\\\/samjeans\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"LLM'er producerer mere un\u00f8jagtige og forudindtagede output med l\u00e6ngere input | DailyAI","description":"Mosh Levy, Alon Jacoby og Yoav Goldberg fra Bar-Ilan University og Allen Institute for AI unders\u00f8gte, hvordan ydeevnen for store sprogmodeller (LLM'er) varierer med \u00e6ndringer i l\u00e6ngden af den inputtekst, de f\u00e5r til behandling.\u00a0","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/dailyai.com\/da\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","og_locale":"da_DK","og_type":"article","og_title":"LLMs produce more inaccurate and biased outputs with longer inputs | DailyAI","og_description":"Mosh Levy, Alon Jacoby, and Yoav Goldberg, from the Bar-Ilan University and Allen Institute for AI, investigated how the performance of large language models (LLMs) varies with changes in the length of the input text they are given to process.\u00a0","og_url":"https:\/\/dailyai.com\/da\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","og_site_name":"DailyAI","article_published_time":"2024-02-29T21:55:56+00:00","article_modified_time":"2024-03-07T07:21:27+00:00","og_image":[{"width":1000,"height":667,"url":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","type":"image\/jpeg"}],"author":"Sam Jeans","twitter_card":"summary_large_image","twitter_creator":"@DailyAIOfficial","twitter_site":"@DailyAIOfficial","twitter_misc":{"Skrevet af":"Sam Jeans","Estimeret l\u00e6setid":"4 minutter"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"NewsArticle","@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#article","isPartOf":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/"},"author":{"name":"Sam Jeans","@id":"https:\/\/dailyai.com\/#\/schema\/person\/711e81f945549438e8bbc579efdeb3c9"},"headline":"LLMs produce more inaccurate and biased outputs with longer inputs","datePublished":"2024-02-29T21:55:56+00:00","dateModified":"2024-03-07T07:21:27+00:00","mainEntityOfPage":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/"},"wordCount":760,"publisher":{"@id":"https:\/\/dailyai.com\/#organization"},"image":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#primaryimage"},"thumbnailUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","keywords":["LLMS","Open-source"],"articleSection":["Industry"],"inLanguage":"da-DK"},{"@type":"WebPage","@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","url":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","name":"LLM'er producerer mere un\u00f8jagtige og forudindtagede output med l\u00e6ngere input | DailyAI","isPartOf":{"@id":"https:\/\/dailyai.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#primaryimage"},"image":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#primaryimage"},"thumbnailUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","datePublished":"2024-02-29T21:55:56+00:00","dateModified":"2024-03-07T07:21:27+00:00","description":"Mosh Levy, Alon Jacoby og Yoav Goldberg fra Bar-Ilan University og Allen Institute for AI unders\u00f8gte, hvordan ydeevnen for store sprogmodeller (LLM'er) varierer med \u00e6ndringer i l\u00e6ngden af den inputtekst, de f\u00e5r til behandling.\u00a0","breadcrumb":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#breadcrumb"},"inLanguage":"da-DK","potentialAction":[{"@type":"ReadAction","target":["https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/"]}]},{"@type":"ImageObject","inLanguage":"da-DK","@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#primaryimage","url":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","contentUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","width":1000,"height":667,"caption":"LLM"},{"@type":"BreadcrumbList","@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/dailyai.com\/"},{"@type":"ListItem","position":2,"name":"LLMs produce more inaccurate and biased outputs with longer inputs"}]},{"@type":"WebSite","@id":"https:\/\/dailyai.com\/#website","url":"https:\/\/dailyai.com\/","name":"DailyAI","description":"Din daglige dosis af AI-nyheder","publisher":{"@id":"https:\/\/dailyai.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/dailyai.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"da-DK"},{"@type":"Organization","@id":"https:\/\/dailyai.com\/#organization","name":"DailyAI","url":"https:\/\/dailyai.com\/","logo":{"@type":"ImageObject","inLanguage":"da-DK","@id":"https:\/\/dailyai.com\/#\/schema\/logo\/image\/","url":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/06\/Daily-Ai_TL_colour.png","contentUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/06\/Daily-Ai_TL_colour.png","width":4501,"height":934,"caption":"DailyAI"},"image":{"@id":"https:\/\/dailyai.com\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/x.com\/DailyAIOfficial","https:\/\/www.linkedin.com\/company\/dailyaiofficial\/","https:\/\/www.youtube.com\/@DailyAIOfficial"]},{"@type":"Person","@id":"https:\/\/dailyai.com\/#\/schema\/person\/711e81f945549438e8bbc579efdeb3c9","name":"Sam Jeans","image":{"@type":"ImageObject","inLanguage":"da-DK","@id":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","caption":"Sam Jeans"},"description":"Sam er videnskabs- og teknologiforfatter og har arbejdet i forskellige AI-startups. N\u00e5r han ikke skriver, kan han finde p\u00e5 at l\u00e6se medicinske tidsskrifter eller grave i kasser med vinylplader.","sameAs":["https:\/\/www.linkedin.com\/in\/sam-jeans-6746b9142\/"],"url":"https:\/\/dailyai.com\/da\/author\/samjeans\/"}]}},"_links":{"self":[{"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/posts\/10434","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/comments?post=10434"}],"version-history":[{"count":6,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/posts\/10434\/revisions"}],"predecessor-version":[{"id":10444,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/posts\/10434\/revisions\/10444"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/media\/10438"}],"wp:attachment":[{"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/media?parent=10434"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/categories?post=10434"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/dailyai.com\/da\/wp-json\/wp\/v2\/tags?post=10434"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}