{"id":6254,"date":"2025-02-07T16:24:30","date_gmt":"2025-02-07T16:24:30","guid":{"rendered":"https:\/\/processtalks.com\/?p=6254"},"modified":"2025-02-07T17:57:18","modified_gmt":"2025-02-07T17:57:18","slug":"tecnologia-llm-as-a-judge-i-assistents-rag","status":"publish","type":"post","link":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/","title":{"rendered":"Tecnologia LLM-as-a-Judge i assistents RAG"},"content":{"rendered":"\n<p>M\u00e8todes i eines<\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><a><\/a>Introducci\u00f3<\/h2>\n\n\n\n<p>L&#8217;adopci\u00f3 generalitzada de models massius de llenguatge (en angl\u00e8s <em>Large language Models, LLMs<\/em>) ha portat al desenvolupament de t\u00e8cniques per poder optimitzar-ne les seves capacitats en situacions diverses. Per exemple, la t\u00e8cnica coneguda amb el terme angl\u00e8s de <em>fine-tuning<\/em> s&#8217;utilitza per adaptar un model fundacional a una tasca o domini concret entrenant-lo amb un conjunt m\u00e9s petit de dades especialitzades; el <em>prompting<\/em> guia les respostes d&#8217;un model fundacional sense necessitat d&#8217;entrenament o alineaci\u00f3 addicional; i la <em>retrieval-augmented generation<\/em> (<em>RAG<\/em>) permet d\u2019obtenir informaci\u00f3 de fonts externes en temps real i de manera precisa, reduint la depend\u00e8ncia a les dades est\u00e0tiques del cos de coneixement del model inicial.<\/p>\n\n\n\n<p>Totes aquestes t\u00e8cniques necessiten per\u00f2 el suport de m\u00e8todes d&#8217;avaluaci\u00f3 s\u00f2lids, no nom\u00e9s per accelerar el desplegament de l\u2019aplicaci\u00f3 final sin\u00f3, el que \u00e9s m\u00e9s important, per garantir-ne resultats precisos i fiables. En aquest article ens centrem en un d&#8217;aquests m\u00e8todes, conegut en angl\u00e8s com a <strong><em>LLM-as-a-Judge,<\/em><\/strong> aplicat en assistents virtuals basats en RAG. Igualment, presentem <strong>una<\/strong> <strong>eina desenvolupada a Process Talks<\/strong> per facilitar l\u2019aplicaci\u00f3 d\u2019aquest m\u00e8tode d\u2019avaluaci\u00f3 als vostres projectes.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><a><\/a>Sistemes RAG: Com avaluar-los<\/h2>\n\n\n\n<p>Imagineu-vos els seg\u00fcents escenaris:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>T\u00e8cnics municipals que assessoren sobre normatives, decrets i regulacions a ciutadans que volen obtenir un perm\u00eds d&#8217;obres, una llic\u00e8ncia comercial o similar. Els cal un acc\u00e9s r\u00e0pid a tota la informaci\u00f3 rellevant en cada cas d\u2019entre tot un cos de documentaci\u00f3 legal que s\u2019actualitza peri\u00f2dicament.&nbsp;<\/li>\n\n\n\n<li>PiMEs centrades en productes i serveis especialitzats i que presten suport als seus clients sobre actualitzacions, bones pr\u00e0ctiques, resoluci\u00f3 de problemes, etc. Per a aquestes empreses, atendre les sol\u00b7licituds dels clients a temps suposa un esfor\u00e7 notable a causa dels recursos limitats amb qu\u00e8 compten, la dispersi\u00f3 documental, l&#8217;actualitzaci\u00f3 constant de la informaci\u00f3 i la gran variabilitat en les consultes.<\/li>\n<\/ul>\n\n\n\n<p>Els <strong>assistents basats en RAG<\/strong> ofereixen una soluci\u00f3 \u00f2ptima a situacions d&#8217;aquest tipus perqu\u00e8 proporcionen un acc\u00e9s r\u00e0pid i prec\u00eds a la informaci\u00f3 que cal en cada cas. No obstant aix\u00f2, el seu \u00e8xit dep\u00e8n de com es desplegui el sistema per adaptar-se a les particularitats de cada context (t\u00e8cniques utilitzades per al tractament de dades i text, indexaci\u00f3 de documents, etc.), per\u00f2 tamb\u00e9, i de manera molt important, dels mecanismes d&#8217;avaluaci\u00f3 establerts per garantir-ne un rendiment m\u00e0xim.<\/p>\n\n\n\n<p>L&#8217;avaluaci\u00f3 dels assistents basats en RAG es fonamenta principalment en 2 par\u00e0metres:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Precisi\u00f3<\/strong>. Determina si el contingut de la resposta proporcionada per l&#8217;assistent \u00e9s correcte.<\/li>\n\n\n\n<li><strong>Adequaci\u00f3<\/strong>. Analitza si la resposta est\u00e0 ben formulada en termes d&#8217;extensi\u00f3 (massa verbositat? brevetat excessiva?), si el registre s\u2019adapta al tipus d\u2019usuari (exc\u00e9s o manca de formalitat?), si sona natural en la llengua de destinaci\u00f3, etc.&nbsp;&nbsp;<\/li>\n<\/ul>\n\n\n\n<p>Per garantir que els assistents basats en RAG ofereixin respostes fiables i adequades caldr\u00e0 doncs desplegar una capa s\u00f2lida d\u2019avaluaci\u00f3. I aqu\u00ed \u00e9s on entra en joc la tecnologia LLM-as-a-Judge.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><a><\/a>Qu\u00e8 \u00e9s LLM-as-a-Judge?<\/h2>\n\n\n\n<p><strong>LLM-as-a-Judge (LLM-J)<\/strong> \u00e9s un procediment d&#8217;avaluaci\u00f3 que utilitza LLMs per avaluar la qualitat de les respostes generades per altres models. Ens podem preguntar: qu\u00e8 justifica aquest enfoc? Fins a quin punt \u00e9s fiable utilitzar les capacitats d&#8217;una IA per avaluar les capacitats d&#8217;una altra IA? Al capdavall, ja comptem amb mecanismes automatitzats per avaluar el rendiment de sistemes, que s&#8217;han utilitzat durant d\u00e8cades i que es basen en el discerniment hum\u00e0 en lloc de la intel\u00b7lig\u00e8ncia artificial.&nbsp;<\/p>\n\n\n\n<p>I \u00e9s que <strong>els processos d&#8217;avaluaci\u00f3 basats en expertesa humana poden ser<\/strong> <strong>automatitzats <\/strong>en moltes situacions. Per exemple, si les respostes esperades es poden contrastar de manera fiable sobre un <strong>est\u00e0ndard de refer\u00e8ncia<\/strong> \u2013 en angl\u00e8s, <em>gold standard<\/em> (per ex., en el cas de preguntes s\u00ed\/no, preguntes factuals, etc.) o avaluades mitjan\u00e7ant <strong>regles heur\u00edstiques<\/strong> (p. ex., quan les respostes depenen de determinades condicions: <em>si X llavors respon A, si Y llavors respon B<\/em>). En aquests dos casos, cal l\u2019expertesa humana per generar el coneixement (conjunts de dades de refer\u00e8ncia, conjunts de regles, etc.) que un proc\u00e9s automatitzat far\u00e0 servir per avaluar els resultats del sistema.<\/p>\n\n\n\n<p>Ara b\u00e9, qu\u00e8 fer quan la sortida del sistema no implica una soluci\u00f3 \u00fanica? O quan es presenta com a text lliure i no estructurat? O quan es pot expressar de maneres diferents per\u00f2 igualment v\u00e0lides? Aqu\u00ed \u00e9s on els anteriors m\u00e8todes automatitzats queden curts i caldria per contra un proc\u00e9s ad hoc de validaci\u00f3 humana que avalui la correcci\u00f3 de cada resposta.&nbsp;<\/p>\n\n\n\n<p>Per\u00f2 de la mateixa manera que els models d&#8217;IA destaquen per la seva compet\u00e8ncia en activitats cognitives generals (com redactar o resumir un text), tamb\u00e9 es poden utilitzar per avaluar si la resposta a una pregunta \u00e9s correcta, fins i tot quan est\u00e0 frasejada diferent de la l\u2019est\u00e0ndard de refer\u00e8ncia. Aix\u00ed doncs, a l\u2019hora d\u2019avaluar la precisi\u00f3 i l&#8217;adequaci\u00f3 dels assistents basats en RAG, el m\u00e8tode LLM-J t\u00e9 un rendiment molt comparable al de l\u2019avaluaci\u00f3 humana.<\/p>\n\n\n\n<p>La t\u00e8cnica LLM-J pot implementar-se amb variacions concretes per adaptar-se a escenaris diferents, en base als seg\u00fcents aspectes:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Veritat de refer\u00e8ncia o conjunt de criteris?<\/strong> L\u2019alternativa aqu\u00ed \u00e9s si l\u2019LLM-J ha d\u2019avaluar respecte de dades de refer\u00e8ncia preses com a \u201cveritat de base\u201d (en angl\u00e8s, <em>ground truth<\/em>) o b\u00e9 fent servir un conjunt de criteris que descriuen qu\u00e8 s&#8217;ha de considerar com a acceptable.<br><br>La <strong>veritat de refer\u00e8ncia <\/strong>sol consistir en parelles de pregunta-resposta. La tasca de l\u2019avaluador LLM-J \u00e9s analitzar si les respostes de l&#8217;assistent RAG s\u00f3n equivalents a les de la refer\u00e8ncia, fins i tot si estan expressades de manera diferent. Per contra, el <strong>conjunt de criteris<\/strong> no proporciona una resposta clau sin\u00f3 una descripci\u00f3 de quin tipus de resposta s&#8217;espera (per exemple, en termes de claredat, extensi\u00f3, flu\u00efdesa, registre, etc.). Tant la veritat de refer\u00e8ncia com els criteris s\u00f3n establerts per humans, assegurant aix\u00ed un nivell \u00faltim de control.<br><br>Tamb\u00e9 hi ha sistemes que es basen en tots dos elements, cada un per avaluar un aspecte diferent: la veritat de refer\u00e8ncia es pot utilitzar per avaluar<strong> la<\/strong> <strong>precisi\u00f3<\/strong> de les respostes (\u00e9s a dir, si proporcionen el contingut correcte), mentre que una bateria de criteris pot fer-se servir per qualificar <strong>l&#8217;adequaci\u00f3<\/strong> de les respostes (registre, longitud, grau de flu\u00efdesa, etc.).<br><\/li>\n\n\n\n<li><strong>Avaluaci\u00f3 d\u2019un \u00fanic model o d\u2019un conjunt de models<\/strong>. L\u2019LLM-J pot usar-se per avaluar un \u00fanic model o per comparar-ne uns quants\u00a0 i triar-ne un de guanyador. En tots dos casos, l&#8217;avaluaci\u00f3 es pot recolzar tant en una veritat de refer\u00e8ncia com en un conjunt de criteris, o en tots dos elements a la vegada.<br><\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\"><a><\/a>Avaluadors LLM-J en assistents basats en RAG<\/h2>\n\n\n\n<p>Els avaluadors LLM-J poden donar suport als assistents basats en RAG en dues fases de la seva exist\u00e8ncia. En un primer pas, <strong>durant el desenvolupament de l&#8217;assistent<\/strong>. En aquest punt, l\u2019avaluador LLM-J pot basar-se tant en una veritat de refer\u00e8ncia com en un conjunt de criteris. De la mateixa manera, l&#8217;avaluaci\u00f3 es pot aplicar sobre diversos models simult\u00e0niament per triar-ne el millor, o nom\u00e9s sobre un de sol, escollit pr\u00e8viament, per tal de refinar els processos de recuperaci\u00f3 i generaci\u00f3 del RAG, i doncs anar ajustant l&#8217;assistent de manera iterativa, millorant-ne la qualitat pr\u00e8viament al seu desplegament final.<\/p>\n\n\n\n<p>Els avaluadors LLM-J tamb\u00e9 poden donar suport als assistents RAG en un pas posterior, <strong>un cop aquests ja estan desplegats i en \u00fas<\/strong>. Donada la imprevisibilitat de les preguntes dels usuaris, aqu\u00ed no es pot comptar amb una veritat de refer\u00e8ncia, per\u00f2 s\u00ed que es pot avaluar les respostes de l&#8217;assistent en funci\u00f3 d\u2019un conjunt de criteris que permeti de discriminar-ne les que possiblement no s\u2019adeq\u00fcen al perfil de l\u2019usuari, necessiten una capa de postproc\u00e9s, etc.<\/p>\n\n\n\n<p>En aquesta mateixa fase tamb\u00e9 es pot utilitzar un conjunt de criteris en el cas d\u2019assistents h\u00edbrids, \u00e9s a dir, assistents el RAG del quals crida diversos LLMs per a una mateixa cerca. En aquest cas, el conjunt de criteris serveix per determinar quina \u00e9s la millor resposta entre les retornades per cada model.&nbsp;&nbsp;&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><a><\/a>Eina LLM-J per avaluar assistents RAG<\/h2>\n\n\n\n<p>Actualment hi ha diversos marcs de desenvolupament que permeten configurar un proc\u00e9s d&#8217;avaluaci\u00f3 LLM-J de manera r\u00e0pida, com ara <a href=\"https:\/\/docs.smith.langchain.com\/evaluation\/how_to_guides\/llm_as_judge\">langchain<\/a> i <a href=\"https:\/\/langfuse.com\/guides\/videos\/llm-as-a-judge-eval-on-dataset-experiments\">langfuse<\/a> per citar-ne un parell. El que \u00e9s m\u00e9s, alguns d&#8217;aquests recursos permeten visualitzar els resultats d\u2019una avaluaci\u00f3 LLM-J. Tot i aix\u00ed, en aquestes eines la presentaci\u00f3 i possibilitat de consulta de les dades no sempre \u00e9s la m\u00e9s eficient per al desenvolupador perqu\u00e8 solen dependre del marc tecnol\u00f2gic del prove\u00efdor. Per aix\u00f2, a Process Talks hem desenvolupat una eina per avaluar el rendiment d\u2019assistents basats en RAG, vers\u00e0til i f\u00e0cil d&#8217;utilitzar. Est\u00e0 implementada sobre llibreries comunes i d\u2019acc\u00e9s gratu\u00eft per aprofitar la solidesa d\u2019aquests recursos, per\u00f2 permet tanmateix una independ\u00e8ncia completa respecte de plataformes de tercers.<\/p>\n\n\n\n<p>Es pot utilitzar per avaluar un \u00fanic model d&#8217;IA o tot un conjunt d&#8217;aquests en paral\u00b7lel, en base a la mateixa veritat de refer\u00e8ncia o bateria de criteris, per tal d\u2019escollir-ne el que respondr\u00e0 millor als requeriments del vostre assistent. La figura seg\u00fcent mostra una imatge d\u2019aquesta eina en un context de m\u00faltiples models.<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img fetchpriority=\"high\" decoding=\"async\" width=\"900\" height=\"445\" src=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure1.jpg\" alt=\"\" class=\"wp-image-6237\" srcset=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure1.jpg 900w, https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure1-300x148.jpg 300w, https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure1-768x380.jpg 768w\" sizes=\"(max-width: 900px) 100vw, 900px\" \/><figcaption class=\"wp-element-caption\">Visi\u00f3 general de la nostra eina d\u2019avaluaci\u00f3 LLM-J. Permet avaluar models en paral\u00b7lel. La taula inferior mostra el veredicte del jutge LLM-J per a cada model emprat (columnes) i proves realitzades (identificades com a Q0-Q6 a les files).<\/figcaption><\/figure>\n\n\n\n<p>Les cel\u00b7les de la matriu de la figura anterior presenten el resultat de l\u2019avaluaci\u00f3 LLM-J per a cada model d&#8217;IA utilitzat pel RAG (per exemple, LLAMA 3.1-8B) sobre una prova concreta (per exemple, Q4).&nbsp; Fent clic a una cel\u00b7la, s\u2019obt\u00e9 els detalls de la prova en q\u00fcesti\u00f3, tal com es mostra a la figura seg\u00fcent. En aquest cas, l\u2019avaluador LLM-J conclou que LLAMA 3.1-8B ofereix la resposta correcta a la pregunta Q4.<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img decoding=\"async\" width=\"900\" height=\"447\" src=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure2.jpg\" alt=\"\" class=\"wp-image-6240\" srcset=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure2.jpg 900w, https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure2-300x149.jpg 300w, https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure2-768x381.jpg 768w\" sizes=\"(max-width: 900px) 100vw, 900px\" \/><figcaption class=\"wp-element-caption\">Exemple d&#8217;informaci\u00f3 proporcionada per a cada prova d&#8217;avaluaci\u00f3 (per a un sistema RAG en catal\u00e0). La part inferior presenta, a l\u2019esquerra, la resposta de refer\u00e8ncia, i a la dreta, la resposta retornada pel sistema RAG. La part superior detalla la configuraci\u00f3 de la prova: model utilitzat per l&#8217;assistent RAG (LLAMA 3.1-8B), la pregunta de la prova, el veredicte de l&#8217;LLM-J i el comentari de l&#8217;LLM-J \u2013 \u00e9s a dir, el raonament que l\u2019ha portat a aquest veredicte.<\/figcaption><\/figure>\n\n\n\n<p>El disseny \u00e9s el mateix per a les respostes (parcialment) err\u00f2nies, en aquest cas naturalment amb un veredicte diferent. La seg\u00fcent figura ho il\u00b7lustra:<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img decoding=\"async\" width=\"800\" height=\"386\" src=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure3.jpg\" alt=\"\" class=\"wp-image-6243\" srcset=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure3.jpg 800w, https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure3-300x145.jpg 300w, https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure3-768x371.jpg 768w\" sizes=\"(max-width: 800px) 100vw, 800px\" \/><figcaption class=\"wp-element-caption\">Detalls de la prova d&#8217;avaluaci\u00f3 en un sistema RAG basat ara en Deepseek-r1-llama-8B. L&#8217;LLM-J conclou que Deepseek-r1-llama-8B no retorna la resposta correcta en aquesta prova. Tamb\u00e9 explica per qu\u00e8 considera que la resposta \u00e9s incorrecta.<\/figcaption><\/figure>\n\n\n\n<p>A m\u00e9s, l&#8217;usuari pot inspeccionar el <em>prompt <\/em>emprat pel RAG en cada prova. D\u2019aquesta&nbsp; manera, a l\u2019hora d\u2019avaluar el rendiment de cada un dels LLM usats pel RAG es disposa de la seva configuraci\u00f3 completa.<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"935\" height=\"871\" src=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure4b.jpg\" alt=\"\" class=\"wp-image-6246\" srcset=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure4b.jpg 935w, https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure4b-300x279.jpg 300w, https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/figure4b-768x715.jpg 768w\" sizes=\"(max-width: 935px) 100vw, 935px\" \/><figcaption class=\"wp-element-caption\">Prompt RAG utilitzat per configurar un model LLM.<\/figcaption><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\"><a><\/a>T&#8217;interessa aquesta tecnologia?<\/h2>\n\n\n\n<p>Com s\u2019ha vist, la t\u00e8cnica LLM-J \u00e9s particularment efica\u00e7 per avaluar resultats textuals i no estructurats on altres m\u00e8todes d&#8217;avaluaci\u00f3 automatitzats queden curts. La integraci\u00f3 d&#8217;aquesta t\u00e8cnica en un marc d&#8217;avaluaci\u00f3 per a sistemes basats en RAG n\u2019accelera el seu desenvolupament, permetent un ajust iteratiu abans del seu desplegament. A m\u00e9s,&nbsp; n\u2019assegura un nivell alt de rendiment, precisi\u00f3 i fiabilitat un cop ja en \u00fas en un context real.<\/p>\n\n\n\n<p>A Process Talks hem desenvolupat la nostra pr\u00f2pia eina per integrar la capacitat d\u2019aquesta tecnologia en qualsevol projecte basat en RAG, garantint aix\u00ed una qualitat m\u00e0xima a les nostres solucions d&#8217;IA.<\/p>\n\n\n\n<p>Si us interessa, podeu posar-vos en contacte amb nosaltres a <a href=\"mailto:hello@processtalks.com\">hola@processtalks.com<\/a>.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>M\u00e8todes i eines Introducci\u00f3 L&#8217;adopci\u00f3 generalitzada de models massius de llenguatge (en angl\u00e8s Large language Models, LLMs) ha portat al desenvolupament de t\u00e8cniques per poder optimitzar-ne les seves capacitats en situacions diverses. Per exemple, la t\u00e8cnica coneguda amb el terme angl\u00e8s de fine-tuning s&#8217;utilitza per adaptar un model fundacional a una tasca o domini concret [&hellip;]<\/p>\n","protected":false},"author":14,"featured_media":6234,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"ocean_post_layout":"","ocean_both_sidebars_style":"","ocean_both_sidebars_content_width":0,"ocean_both_sidebars_sidebars_width":0,"ocean_sidebar":"","ocean_second_sidebar":"","ocean_disable_margins":"enable","ocean_add_body_class":"","ocean_shortcode_before_top_bar":"","ocean_shortcode_after_top_bar":"","ocean_shortcode_before_header":"","ocean_shortcode_after_header":"","ocean_has_shortcode":"","ocean_shortcode_after_title":"","ocean_shortcode_before_footer_widgets":"","ocean_shortcode_after_footer_widgets":"","ocean_shortcode_before_footer_bottom":"","ocean_shortcode_after_footer_bottom":"","ocean_display_top_bar":"default","ocean_display_header":"default","ocean_header_style":"","ocean_center_header_left_menu":"","ocean_custom_header_template":"","ocean_custom_logo":0,"ocean_custom_retina_logo":0,"ocean_custom_logo_max_width":0,"ocean_custom_logo_tablet_max_width":0,"ocean_custom_logo_mobile_max_width":0,"ocean_custom_logo_max_height":0,"ocean_custom_logo_tablet_max_height":0,"ocean_custom_logo_mobile_max_height":0,"ocean_header_custom_menu":"","ocean_menu_typo_font_family":"","ocean_menu_typo_font_subset":"","ocean_menu_typo_font_size":0,"ocean_menu_typo_font_size_tablet":0,"ocean_menu_typo_font_size_mobile":0,"ocean_menu_typo_font_size_unit":"px","ocean_menu_typo_font_weight":"","ocean_menu_typo_font_weight_tablet":"","ocean_menu_typo_font_weight_mobile":"","ocean_menu_typo_transform":"","ocean_menu_typo_transform_tablet":"","ocean_menu_typo_transform_mobile":"","ocean_menu_typo_line_height":0,"ocean_menu_typo_line_height_tablet":0,"ocean_menu_typo_line_height_mobile":0,"ocean_menu_typo_line_height_unit":"","ocean_menu_typo_spacing":0,"ocean_menu_typo_spacing_tablet":0,"ocean_menu_typo_spacing_mobile":0,"ocean_menu_typo_spacing_unit":"","ocean_menu_link_color":"","ocean_menu_link_color_hover":"","ocean_menu_link_color_active":"","ocean_menu_link_background":"","ocean_menu_link_hover_background":"","ocean_menu_link_active_background":"","ocean_menu_social_links_bg":"","ocean_menu_social_hover_links_bg":"","ocean_menu_social_links_color":"","ocean_menu_social_hover_links_color":"","ocean_disable_title":"default","ocean_disable_heading":"default","ocean_post_title":"","ocean_post_subheading":"","ocean_post_title_style":"","ocean_post_title_background_color":"","ocean_post_title_background":0,"ocean_post_title_bg_image_position":"","ocean_post_title_bg_image_attachment":"","ocean_post_title_bg_image_repeat":"","ocean_post_title_bg_image_size":"","ocean_post_title_height":0,"ocean_post_title_bg_overlay":0.5,"ocean_post_title_bg_overlay_color":"","ocean_disable_breadcrumbs":"default","ocean_breadcrumbs_color":"","ocean_breadcrumbs_separator_color":"","ocean_breadcrumbs_links_color":"","ocean_breadcrumbs_links_hover_color":"","ocean_display_footer_widgets":"default","ocean_display_footer_bottom":"default","ocean_custom_footer_template":"","ocean_post_oembed":"","ocean_post_self_hosted_media":"","ocean_post_video_embed":"","ocean_link_format":"","ocean_link_format_target":"self","ocean_quote_format":"","ocean_quote_format_link":"post","ocean_gallery_link_images":"on","ocean_gallery_id":[],"footnotes":""},"categories":[29],"tags":[],"class_list":["post-6254","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-uncategorized-ca","entry","has-media"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Tecnologia LLM-as-a-Judge i assistents RAG - Process Talks<\/title>\n<meta name=\"description\" content=\"Automate your daily office workflows with an engine based on natural language interaction.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/\" \/>\n<meta property=\"og:locale\" content=\"ca_ES\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Tecnologia LLM-as-a-Judge i assistents RAG - Process Talks\" \/>\n<meta property=\"og:description\" content=\"Automate your daily office workflows with an engine based on natural language interaction.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/\" \/>\n<meta property=\"og:site_name\" content=\"Process Talks\" \/>\n<meta property=\"article:published_time\" content=\"2025-02-07T16:24:30+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2025-02-07T17:57:18+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/conny-schneider-3hkKv6WzjcE-unsplash.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1200\" \/>\n\t<meta property=\"og:image:height\" content=\"739\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"web-designer\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Escrit per\" \/>\n\t<meta name=\"twitter:data1\" content=\"web-designer\" \/>\n\t<meta name=\"twitter:label2\" content=\"Temps estimat de lectura\" \/>\n\t<meta name=\"twitter:data2\" content=\"10 minuts\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/\"},\"author\":{\"name\":\"web-designer\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/#\\\/schema\\\/person\\\/fa4d5a416a5d88d77b957a359148b1a1\"},\"headline\":\"Tecnologia LLM-as-a-Judge i assistents RAG\",\"datePublished\":\"2025-02-07T16:24:30+00:00\",\"dateModified\":\"2025-02-07T17:57:18+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/\"},\"wordCount\":2095,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/processtalks.com\\\/wp-content\\\/uploads\\\/2025\\\/02\\\/conny-schneider-3hkKv6WzjcE-unsplash.jpg\",\"articleSection\":[\"Uncategorized @ca\"],\"inLanguage\":\"ca\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/\",\"url\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/\",\"name\":\"Tecnologia LLM-as-a-Judge i assistents RAG - Process Talks\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/processtalks.com\\\/wp-content\\\/uploads\\\/2025\\\/02\\\/conny-schneider-3hkKv6WzjcE-unsplash.jpg\",\"datePublished\":\"2025-02-07T16:24:30+00:00\",\"dateModified\":\"2025-02-07T17:57:18+00:00\",\"description\":\"Automate your daily office workflows with an engine based on natural language interaction.\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/#breadcrumb\"},\"inLanguage\":\"ca\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"ca\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/#primaryimage\",\"url\":\"https:\\\/\\\/processtalks.com\\\/wp-content\\\/uploads\\\/2025\\\/02\\\/conny-schneider-3hkKv6WzjcE-unsplash.jpg\",\"contentUrl\":\"https:\\\/\\\/processtalks.com\\\/wp-content\\\/uploads\\\/2025\\\/02\\\/conny-schneider-3hkKv6WzjcE-unsplash.jpg\",\"width\":1200,\"height\":739},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/ca\\\/tecnologia-llm-as-a-judge-i-assistents-rag\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/processtalks.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Tecnologia LLM-as-a-Judge i assistents RAG\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/#website\",\"url\":\"https:\\\/\\\/processtalks.com\\\/\",\"name\":\"Process Talks\",\"description\":\"Automate your daily office workflows with an engine based on natural language interaction.\",\"publisher\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/processtalks.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"ca\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/#organization\",\"name\":\"Process Talks\",\"url\":\"https:\\\/\\\/processtalks.com\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"ca\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/processtalks.com\\\/wp-content\\\/uploads\\\/2022\\\/08\\\/white-logo.png\",\"contentUrl\":\"https:\\\/\\\/processtalks.com\\\/wp-content\\\/uploads\\\/2022\\\/08\\\/white-logo.png\",\"width\":768,\"height\":238,\"caption\":\"Process Talks\"},\"image\":{\"@id\":\"https:\\\/\\\/processtalks.com\\\/#\\\/schema\\\/logo\\\/image\\\/\"}},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/processtalks.com\\\/#\\\/schema\\\/person\\\/fa4d5a416a5d88d77b957a359148b1a1\",\"name\":\"web-designer\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"ca\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/d27d450c0cf5599e776e61c8cd4a50679e849cf6c6184ab991590969dda10f52?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/d27d450c0cf5599e776e61c8cd4a50679e849cf6c6184ab991590969dda10f52?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/d27d450c0cf5599e776e61c8cd4a50679e849cf6c6184ab991590969dda10f52?s=96&d=mm&r=g\",\"caption\":\"web-designer\"}}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Tecnologia LLM-as-a-Judge i assistents RAG - Process Talks","description":"Automate your daily office workflows with an engine based on natural language interaction.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/","og_locale":"ca_ES","og_type":"article","og_title":"Tecnologia LLM-as-a-Judge i assistents RAG - Process Talks","og_description":"Automate your daily office workflows with an engine based on natural language interaction.","og_url":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/","og_site_name":"Process Talks","article_published_time":"2025-02-07T16:24:30+00:00","article_modified_time":"2025-02-07T17:57:18+00:00","og_image":[{"width":1200,"height":739,"url":"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/conny-schneider-3hkKv6WzjcE-unsplash.jpg","type":"image\/jpeg"}],"author":"web-designer","twitter_card":"summary_large_image","twitter_misc":{"Escrit per":"web-designer","Temps estimat de lectura":"10 minuts"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/#article","isPartOf":{"@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/"},"author":{"name":"web-designer","@id":"https:\/\/processtalks.com\/#\/schema\/person\/fa4d5a416a5d88d77b957a359148b1a1"},"headline":"Tecnologia LLM-as-a-Judge i assistents RAG","datePublished":"2025-02-07T16:24:30+00:00","dateModified":"2025-02-07T17:57:18+00:00","mainEntityOfPage":{"@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/"},"wordCount":2095,"commentCount":0,"publisher":{"@id":"https:\/\/processtalks.com\/#organization"},"image":{"@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/#primaryimage"},"thumbnailUrl":"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/conny-schneider-3hkKv6WzjcE-unsplash.jpg","articleSection":["Uncategorized @ca"],"inLanguage":"ca","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/","url":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/","name":"Tecnologia LLM-as-a-Judge i assistents RAG - Process Talks","isPartOf":{"@id":"https:\/\/processtalks.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/#primaryimage"},"image":{"@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/#primaryimage"},"thumbnailUrl":"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/conny-schneider-3hkKv6WzjcE-unsplash.jpg","datePublished":"2025-02-07T16:24:30+00:00","dateModified":"2025-02-07T17:57:18+00:00","description":"Automate your daily office workflows with an engine based on natural language interaction.","breadcrumb":{"@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/#breadcrumb"},"inLanguage":"ca","potentialAction":[{"@type":"ReadAction","target":["https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/"]}]},{"@type":"ImageObject","inLanguage":"ca","@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/#primaryimage","url":"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/conny-schneider-3hkKv6WzjcE-unsplash.jpg","contentUrl":"https:\/\/processtalks.com\/wp-content\/uploads\/2025\/02\/conny-schneider-3hkKv6WzjcE-unsplash.jpg","width":1200,"height":739},{"@type":"BreadcrumbList","@id":"https:\/\/processtalks.com\/ca\/tecnologia-llm-as-a-judge-i-assistents-rag\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/processtalks.com\/"},{"@type":"ListItem","position":2,"name":"Tecnologia LLM-as-a-Judge i assistents RAG"}]},{"@type":"WebSite","@id":"https:\/\/processtalks.com\/#website","url":"https:\/\/processtalks.com\/","name":"Process Talks","description":"Automate your daily office workflows with an engine based on natural language interaction.","publisher":{"@id":"https:\/\/processtalks.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/processtalks.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"ca"},{"@type":"Organization","@id":"https:\/\/processtalks.com\/#organization","name":"Process Talks","url":"https:\/\/processtalks.com\/","logo":{"@type":"ImageObject","inLanguage":"ca","@id":"https:\/\/processtalks.com\/#\/schema\/logo\/image\/","url":"https:\/\/processtalks.com\/wp-content\/uploads\/2022\/08\/white-logo.png","contentUrl":"https:\/\/processtalks.com\/wp-content\/uploads\/2022\/08\/white-logo.png","width":768,"height":238,"caption":"Process Talks"},"image":{"@id":"https:\/\/processtalks.com\/#\/schema\/logo\/image\/"}},{"@type":"Person","@id":"https:\/\/processtalks.com\/#\/schema\/person\/fa4d5a416a5d88d77b957a359148b1a1","name":"web-designer","image":{"@type":"ImageObject","inLanguage":"ca","@id":"https:\/\/secure.gravatar.com\/avatar\/d27d450c0cf5599e776e61c8cd4a50679e849cf6c6184ab991590969dda10f52?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/d27d450c0cf5599e776e61c8cd4a50679e849cf6c6184ab991590969dda10f52?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/d27d450c0cf5599e776e61c8cd4a50679e849cf6c6184ab991590969dda10f52?s=96&d=mm&r=g","caption":"web-designer"}}]}},"_links":{"self":[{"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/posts\/6254","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/users\/14"}],"replies":[{"embeddable":true,"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/comments?post=6254"}],"version-history":[{"count":2,"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/posts\/6254\/revisions"}],"predecessor-version":[{"id":6260,"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/posts\/6254\/revisions\/6260"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/media\/6234"}],"wp:attachment":[{"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/media?parent=6254"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/categories?post=6254"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/processtalks.com\/ca\/wp-json\/wp\/v2\/tags?post=6254"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}