{"id":10434,"date":"2024-02-29T21:55:56","date_gmt":"2024-02-29T21:55:56","guid":{"rendered":"https:\/\/dailyai.com\/?p=10434"},"modified":"2024-03-07T07:21:27","modified_gmt":"2024-03-07T07:21:27","slug":"llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs","status":"publish","type":"post","link":"https:\/\/dailyai.com\/es\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","title":{"rendered":"Los LLM producen resultados m\u00e1s imprecisos y sesgados con entradas m\u00e1s largas"},"content":{"rendered":"<p><strong>A pesar de los r\u00e1pidos avances de los LLM, seguimos sin comprender c\u00f3mo estos modelos hacen frente a entradas m\u00e1s largas.<\/strong><\/p>\n<p>Mosh Levy, Alon Jacoby y Yoav Goldberg, de la Universidad Bar-Ilan y el Instituto Allen de Inteligencia Artificial, investigaron c\u00f3mo var\u00eda el rendimiento de los grandes modelos ling\u00fc\u00edsticos (LLM) en funci\u00f3n de la longitud del texto de entrada que deben procesar.<\/p>\n<p><span style=\"font-weight: 400;\">Desarrollaron un marco de razonamiento espec\u00edfico para este fin, que les permiti\u00f3 diseccionar la influencia de la longitud de la entrada en el razonamiento LLM en un entorno controlado.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">El marco de preguntas propon\u00eda diferentes versiones de la misma pregunta, cada una de las cuales conten\u00eda la informaci\u00f3n necesaria para responder a la pregunta, rellenada con texto adicional irrelevante de longitud y tipo variables.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Esto permite aislar la longitud de la entrada como variable, garantizando que los cambios en el rendimiento del modelo puedan atribuirse directamente a la longitud de la entrada.<\/span><\/p>\n<h3><b>Principales resultados<\/b><\/h3>\n<p><span style=\"font-weight: 400;\">Levy, Jacoby y Goldberg descubrieron que los LLM muestran un notable descenso en el rendimiento del razonamiento a longitudes de entrada muy por debajo de lo que los desarrolladores afirman que pueden manejar. Documentaron sus hallazgos <a href=\"https:\/\/arxiv.org\/pdf\/2402.14848.pdf\" target=\"_blank\" rel=\"noopener\">en este estudio<\/a>.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">El declive se observ\u00f3 sistem\u00e1ticamente en todas las versiones del conjunto de datos, lo que indica un problema sist\u00e9mico con el manejo de entradas m\u00e1s largas en lugar de un problema vinculado a muestras de datos o arquitecturas de modelos espec\u00edficas.\u00a0<\/span><\/p>\n<p>Como describen los investigadores, \"nuestros hallazgos muestran una notable degradaci\u00f3n en el rendimiento de razonamiento de los LLM a longitudes de entrada mucho m\u00e1s cortas que su m\u00e1ximo t\u00e9cnico. Demostramos que la tendencia a la degradaci\u00f3n aparece en todas las versiones de nuestro conjunto de datos, aunque con diferentes intensidades.\"<\/p>\n<p>&nbsp;<\/p>\n<figure id=\"attachment_10436\" aria-describedby=\"caption-attachment-10436\" style=\"width: 569px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-10436\" src=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4.png\" alt=\"\" width=\"569\" height=\"469\" srcset=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4.png 733w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4-300x247.png 300w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4-370x305.png 370w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4-20x16.png 20w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/model4-58x48.png 58w\" sizes=\"auto, (max-width: 569px) 100vw, 569px\" \/><figcaption id=\"caption-attachment-10436\" class=\"wp-caption-text\">A medida que aumenta el tama\u00f1o de la entrada, disminuye la capacidad de realizar tareas de razonamiento. Estas entradas constan de texto relevante (resaltado en rojo) e irrelevante (en gris), que se obtienen de varios lugares y se ampl\u00edan de forma incremental. Para responder con precisi\u00f3n es necesario identificar dos segmentos de texto concretos, que podr\u00edan estar situados aleatoriamente dentro de la entrada. Los datos de rendimiento se han obtenido a partir de 600 muestras. Fuente: V\u00eda <a href=\"https:\/\/arxiv.org\/pdf\/2402.14848.pdf\">ArXiv.<\/a><\/figcaption><\/figure>\n<p><span style=\"font-weight: 400;\">Adem\u00e1s, el estudio pone de relieve c\u00f3mo las m\u00e9tricas tradicionales como la perplejidad, utilizadas habitualmente para evaluar los LLM, no se correlacionan con el rendimiento de los modelos en tareas de razonamiento que implican entradas largas.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Una exploraci\u00f3n m\u00e1s profunda descubri\u00f3 que la degradaci\u00f3n del rendimiento no depend\u00eda \u00fanicamente de la presencia de informaci\u00f3n irrelevante (relleno), sino que se observaba incluso cuando dicho relleno consist\u00eda en informaci\u00f3n relevante duplicada.<\/span><\/p>\n<blockquote class=\"twitter-tweet\">\n<p dir=\"ltr\" lang=\"en\">Cuando mantenemos juntos los dos vanos centrales y a\u00f1adimos texto a su alrededor, la precisi\u00f3n ya baja. Introduciendo p\u00e1rrafos entre los espacios, los resultados bajan mucho m\u00e1s. El descenso se produce tanto cuando los textos que a\u00f1adimos son similares a los de la tarea como cuando son completamente diferentes. 3\/7 <a href=\"https:\/\/t.co\/c91l9uzyme\">pic.twitter.com\/c91l9uzyme<\/a><\/p>\n<p>- Mosh Levy (@mosh_levy) <a href=\"https:\/\/twitter.com\/mosh_levy\/status\/1762027631837368416?ref_src=twsrc%5Etfw\">26 de febrero de 2024<\/a><\/p><\/blockquote>\n<p><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><br \/>\n<span style=\"font-weight: 400;\">Esto sugiere que el reto para los LLM reside en filtrar el ruido y el procesamiento inherente a las secuencias de texto m\u00e1s largas.<\/span><\/p>\n<h2><b>Ignorar las instrucciones<\/b><\/h2>\n<p><span style=\"font-weight: 400;\">Un modo de fallo cr\u00edtico que se destaca en el estudio es la tendencia de los LLM a ignorar las instrucciones incrustadas en la entrada a medida que \u00e9sta aumenta.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Los modelos tambi\u00e9n generaban a veces respuestas que indicaban incertidumbre o falta de informaci\u00f3n suficiente, como \"No hay suficiente informaci\u00f3n en el texto\", a pesar de contar con toda la informaci\u00f3n necesaria.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">En general, a medida que aumenta la longitud de la informaci\u00f3n, los estudiantes de LLM parecen tener dificultades para priorizar y centrarse en los elementos de informaci\u00f3n clave, incluidas las instrucciones directas.\u00a0<\/span><\/p>\n<h2>Mostrar sesgos en las respuestas<\/h2>\n<p><span style=\"font-weight: 400;\">Otro problema notable fue el aumento de los sesgos en las respuestas de los modelos a medida que las entradas se hac\u00edan m\u00e1s largas.\u00a0 <\/span><\/p>\n<p><span style=\"font-weight: 400;\">En concreto, los LLM mostraron un sesgo hacia la respuesta \"Falso\" a medida que aumentaba la longitud de la entrada. Este sesgo indica un sesgo en la estimaci\u00f3n de probabilidades o en los procesos de toma de decisiones dentro del modelo, posiblemente como mecanismo defensivo en respuesta a la mayor incertidumbre debida a la mayor longitud de las entradas.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">La inclinaci\u00f3n a favorecer las respuestas \"Falsas\" tambi\u00e9n podr\u00eda reflejar un desequilibrio subyacente en los datos de entrenamiento o un artefacto del proceso de entrenamiento de los modelos, donde las respuestas negativas pueden estar sobrerrepresentadas o asociadas a contextos de incertidumbre y ambig\u00fcedad.\u00a0<\/span><\/p>\n<figure id=\"attachment_10437\" aria-describedby=\"caption-attachment-10437\" style=\"width: 477px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-10437 size-full\" src=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai.png\" alt=\"modelos AI\" width=\"477\" height=\"772\" srcset=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai.png 477w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai-185x300.png 185w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai-370x599.png 370w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai-20x32.png 20w, https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/biasai-30x48.png 30w\" sizes=\"auto, (max-width: 477px) 100vw, 477px\" \/><figcaption id=\"caption-attachment-10437\" class=\"wp-caption-text\">Los modelos mostraron un sesgo hacia la respuesta \"falsa\" a las preguntas binarias a medida que aumentaba la longitud de la entrada. Fuente: V\u00eda <a href=\"https:\/\/arxiv.org\/pdf\/2402.14848.pdf\">ArXiv<\/a>.<\/figcaption><\/figure>\n<p><span style=\"font-weight: 400;\">Este sesgo afecta a la precisi\u00f3n de los resultados de los modelos y suscita dudas sobre la fiabilidad y equidad de los LLM en aplicaciones que requieren una comprensi\u00f3n matizada e imparcialidad.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">La aplicaci\u00f3n de estrategias s\u00f3lidas de detecci\u00f3n y mitigaci\u00f3n de sesgos durante las fases de entrenamiento y ajuste de los modelos es esencial para reducir los sesgos injustificados en las respuestas de los modelos. <\/span><\/p>\n<p><span style=\"font-weight: 400;\">E<\/span><span style=\"font-weight: 400;\">arantizar que los conjuntos de datos de entrenamiento sean diversos, equilibrados y representativos de una amplia gama de escenarios tambi\u00e9n puede ayudar a minimizar los sesgos y mejorar la generalizaci\u00f3n de los modelos.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Esto contribuye a <\/span><a href=\"https:\/\/dailyai.com\/es\/2024\/02\/generative-ai-systems-hallucinations-and-mounting-technical-debt\/\"><span style=\"font-weight: 400;\">otros estudios recientes<\/span><\/a><span style=\"font-weight: 400;\"> que, del mismo modo, ponen de manifiesto problemas fundamentales en el funcionamiento de los LLM, lo que lleva a una situaci\u00f3n en la que esa \"deuda t\u00e9cnica\" podr\u00eda amenazar la funcionalidad y la integridad del modelo con el paso del tiempo.\u00a0<\/span><\/p>","protected":false},"excerpt":{"rendered":"<p>A pesar de los r\u00e1pidos avances en el campo de los LLM, a\u00fan no sabemos muy bien c\u00f3mo funcionan estos modelos con entradas m\u00e1s largas. Mosh Levy, Alon Jacoby y Yoav Goldberg, de la Universidad Bar-Ilan y el Instituto Allen de Inteligencia Artificial, investigaron c\u00f3mo var\u00eda el rendimiento de los modelos ling\u00fc\u00edsticos de gran tama\u00f1o (LLM) en funci\u00f3n de la longitud del texto de entrada que deben procesar. Desarrollaron un marco de razonamiento espec\u00edfico para este fin, que les permiti\u00f3 diseccionar la influencia de la longitud de la entrada en el razonamiento de los LLM en un entorno controlado. El marco de cuestionamiento propon\u00eda distintas versiones de la misma pregunta, cada una de las cuales conten\u00eda la informaci\u00f3n necesaria para responder a la<\/p>","protected":false},"author":2,"featured_media":10438,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[84],"tags":[118,110],"class_list":["post-10434","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-industry","tag-llms","tag-open-source"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>LLMs produce more inaccurate and biased outputs with longer inputs | DailyAI<\/title>\n<meta name=\"description\" content=\"Mosh Levy, Alon Jacoby, and Yoav Goldberg, from the Bar-Ilan University and Allen Institute for AI, investigated how the performance of large language models (LLMs) varies with changes in the length of the input text they are given to process.\u00a0\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/dailyai.com\/es\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/\" \/>\n<meta property=\"og:locale\" content=\"es_ES\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"LLMs produce more inaccurate and biased outputs with longer inputs | DailyAI\" \/>\n<meta property=\"og:description\" content=\"Mosh Levy, Alon Jacoby, and Yoav Goldberg, from the Bar-Ilan University and Allen Institute for AI, investigated how the performance of large language models (LLMs) varies with changes in the length of the input text they are given to process.\u00a0\" \/>\n<meta property=\"og:url\" content=\"https:\/\/dailyai.com\/es\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/\" \/>\n<meta property=\"og:site_name\" content=\"DailyAI\" \/>\n<meta property=\"article:published_time\" content=\"2024-02-29T21:55:56+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2024-03-07T07:21:27+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1000\" \/>\n\t<meta property=\"og:image:height\" content=\"667\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Sam Jeans\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@DailyAIOfficial\" \/>\n<meta name=\"twitter:site\" content=\"@DailyAIOfficial\" \/>\n<meta name=\"twitter:label1\" content=\"Escrito por\" \/>\n\t<meta name=\"twitter:data1\" content=\"Sam Jeans\" \/>\n\t<meta name=\"twitter:label2\" content=\"Tiempo de lectura\" \/>\n\t<meta name=\"twitter:data2\" content=\"4 minutos\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"NewsArticle\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\"},\"author\":{\"name\":\"Sam Jeans\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/person\\\/711e81f945549438e8bbc579efdeb3c9\"},\"headline\":\"LLMs produce more inaccurate and biased outputs with longer inputs\",\"datePublished\":\"2024-02-29T21:55:56+00:00\",\"dateModified\":\"2024-03-07T07:21:27+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\"},\"wordCount\":760,\"publisher\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2024\\\/02\\\/shutterstock_2328020525.jpg\",\"keywords\":[\"LLMS\",\"Open-source\"],\"articleSection\":[\"Industry\"],\"inLanguage\":\"es\"},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\",\"url\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\",\"name\":\"LLMs produce more inaccurate and biased outputs with longer inputs | DailyAI\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2024\\\/02\\\/shutterstock_2328020525.jpg\",\"datePublished\":\"2024-02-29T21:55:56+00:00\",\"dateModified\":\"2024-03-07T07:21:27+00:00\",\"description\":\"Mosh Levy, Alon Jacoby, and Yoav Goldberg, from the Bar-Ilan University and Allen Institute for AI, investigated how the performance of large language models (LLMs) varies with changes in the length of the input text they are given to process.\u00a0\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#breadcrumb\"},\"inLanguage\":\"es\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"es\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#primaryimage\",\"url\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2024\\\/02\\\/shutterstock_2328020525.jpg\",\"contentUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2024\\\/02\\\/shutterstock_2328020525.jpg\",\"width\":1000,\"height\":667,\"caption\":\"LLM\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/2024\\\/02\\\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/dailyai.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"LLMs produce more inaccurate and biased outputs with longer inputs\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#website\",\"url\":\"https:\\\/\\\/dailyai.com\\\/\",\"name\":\"DailyAI\",\"description\":\"Your Daily Dose of AI News\",\"publisher\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/dailyai.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"es\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#organization\",\"name\":\"DailyAI\",\"url\":\"https:\\\/\\\/dailyai.com\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"es\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/06\\\/Daily-Ai_TL_colour.png\",\"contentUrl\":\"https:\\\/\\\/dailyai.com\\\/wp-content\\\/uploads\\\/2023\\\/06\\\/Daily-Ai_TL_colour.png\",\"width\":4501,\"height\":934,\"caption\":\"DailyAI\"},\"image\":{\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/x.com\\\/DailyAIOfficial\",\"https:\\\/\\\/www.linkedin.com\\\/company\\\/dailyaiofficial\\\/\",\"https:\\\/\\\/www.youtube.com\\\/@DailyAIOfficial\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/dailyai.com\\\/#\\\/schema\\\/person\\\/711e81f945549438e8bbc579efdeb3c9\",\"name\":\"Sam Jeans\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"es\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g\",\"caption\":\"Sam Jeans\"},\"description\":\"Sam is a science and technology writer who has worked in various AI startups. When he\u2019s not writing, he can be found reading medical journals or digging through boxes of vinyl records.\",\"sameAs\":[\"https:\\\/\\\/www.linkedin.com\\\/in\\\/sam-jeans-6746b9142\\\/\"],\"url\":\"https:\\\/\\\/dailyai.com\\\/es\\\/author\\\/samjeans\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Los LLM producen resultados m\u00e1s imprecisos y sesgados con entradas m\u00e1s largas | DiarioAI","description":"Mosh Levy, Alon Jacoby y Yoav Goldberg, de la Universidad Bar-Ilan y el Instituto Allen de Inteligencia Artificial, investigaron c\u00f3mo var\u00eda el rendimiento de los grandes modelos ling\u00fc\u00edsticos (LLM) en funci\u00f3n de la longitud del texto de entrada que deben procesar.\u00a0","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/dailyai.com\/es\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","og_locale":"es_ES","og_type":"article","og_title":"LLMs produce more inaccurate and biased outputs with longer inputs | DailyAI","og_description":"Mosh Levy, Alon Jacoby, and Yoav Goldberg, from the Bar-Ilan University and Allen Institute for AI, investigated how the performance of large language models (LLMs) varies with changes in the length of the input text they are given to process.\u00a0","og_url":"https:\/\/dailyai.com\/es\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","og_site_name":"DailyAI","article_published_time":"2024-02-29T21:55:56+00:00","article_modified_time":"2024-03-07T07:21:27+00:00","og_image":[{"width":1000,"height":667,"url":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","type":"image\/jpeg"}],"author":"Sam Jeans","twitter_card":"summary_large_image","twitter_creator":"@DailyAIOfficial","twitter_site":"@DailyAIOfficial","twitter_misc":{"Escrito por":"Sam Jeans","Tiempo de lectura":"4 minutos"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"NewsArticle","@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#article","isPartOf":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/"},"author":{"name":"Sam Jeans","@id":"https:\/\/dailyai.com\/#\/schema\/person\/711e81f945549438e8bbc579efdeb3c9"},"headline":"LLMs produce more inaccurate and biased outputs with longer inputs","datePublished":"2024-02-29T21:55:56+00:00","dateModified":"2024-03-07T07:21:27+00:00","mainEntityOfPage":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/"},"wordCount":760,"publisher":{"@id":"https:\/\/dailyai.com\/#organization"},"image":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#primaryimage"},"thumbnailUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","keywords":["LLMS","Open-source"],"articleSection":["Industry"],"inLanguage":"es"},{"@type":"WebPage","@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","url":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/","name":"Los LLM producen resultados m\u00e1s imprecisos y sesgados con entradas m\u00e1s largas | DiarioAI","isPartOf":{"@id":"https:\/\/dailyai.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#primaryimage"},"image":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#primaryimage"},"thumbnailUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","datePublished":"2024-02-29T21:55:56+00:00","dateModified":"2024-03-07T07:21:27+00:00","description":"Mosh Levy, Alon Jacoby y Yoav Goldberg, de la Universidad Bar-Ilan y el Instituto Allen de Inteligencia Artificial, investigaron c\u00f3mo var\u00eda el rendimiento de los grandes modelos ling\u00fc\u00edsticos (LLM) en funci\u00f3n de la longitud del texto de entrada que deben procesar.\u00a0","breadcrumb":{"@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#breadcrumb"},"inLanguage":"es","potentialAction":[{"@type":"ReadAction","target":["https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/"]}]},{"@type":"ImageObject","inLanguage":"es","@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#primaryimage","url":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","contentUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2024\/02\/shutterstock_2328020525.jpg","width":1000,"height":667,"caption":"LLM"},{"@type":"BreadcrumbList","@id":"https:\/\/dailyai.com\/2024\/02\/llms-produce-more-inaccurate-and-biased-outputs-with-longer-inputs\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/dailyai.com\/"},{"@type":"ListItem","position":2,"name":"LLMs produce more inaccurate and biased outputs with longer inputs"}]},{"@type":"WebSite","@id":"https:\/\/dailyai.com\/#website","url":"https:\/\/dailyai.com\/","name":"DailyAI","description":"Su dosis diaria de noticias sobre IA","publisher":{"@id":"https:\/\/dailyai.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/dailyai.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"es"},{"@type":"Organization","@id":"https:\/\/dailyai.com\/#organization","name":"DailyAI","url":"https:\/\/dailyai.com\/","logo":{"@type":"ImageObject","inLanguage":"es","@id":"https:\/\/dailyai.com\/#\/schema\/logo\/image\/","url":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/06\/Daily-Ai_TL_colour.png","contentUrl":"https:\/\/dailyai.com\/wp-content\/uploads\/2023\/06\/Daily-Ai_TL_colour.png","width":4501,"height":934,"caption":"DailyAI"},"image":{"@id":"https:\/\/dailyai.com\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/x.com\/DailyAIOfficial","https:\/\/www.linkedin.com\/company\/dailyaiofficial\/","https:\/\/www.youtube.com\/@DailyAIOfficial"]},{"@type":"Person","@id":"https:\/\/dailyai.com\/#\/schema\/person\/711e81f945549438e8bbc579efdeb3c9","name":"Sam Jeans","image":{"@type":"ImageObject","inLanguage":"es","@id":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/a24a4a8f8e2a1a275b7491dc9c9f032c401eabf23c3206da4628dc84b6dac5c8?s=96&d=robohash&r=g","caption":"Sam Jeans"},"description":"Sam es un escritor de ciencia y tecnolog\u00eda que ha trabajado en varias startups de IA. Cuando no est\u00e1 escribiendo, se le puede encontrar leyendo revistas m\u00e9dicas o rebuscando en cajas de discos de vinilo.","sameAs":["https:\/\/www.linkedin.com\/in\/sam-jeans-6746b9142\/"],"url":"https:\/\/dailyai.com\/es\/author\/samjeans\/"}]}},"_links":{"self":[{"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/posts\/10434","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/comments?post=10434"}],"version-history":[{"count":6,"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/posts\/10434\/revisions"}],"predecessor-version":[{"id":10444,"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/posts\/10434\/revisions\/10444"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/media\/10438"}],"wp:attachment":[{"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/media?parent=10434"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/categories?post=10434"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/dailyai.com\/es\/wp-json\/wp\/v2\/tags?post=10434"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}