<?xml version="1.0" encoding="UTF-8"?>
<urlset
  xmlns="http://www.sitemaps.org/schemas/sitemap/0.9"
  xmlns:news="http://www.google.com/schemas/sitemap-news/0.9"
>
  <url>
    <loc>https://aiexpert.news/en/article/negation-neglect-llms-can-learn-false-claims-from-negation-heavy-fine-tuning</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T21:34:08.326Z</news:publication_date>
      <news:title>Negation Neglect Drives False Belief Rate to 88.6% in Fine-Tuned LLMs</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/negation-neglect-llms-can-learn-false-claims-from-negation-heavy-fine-tuning</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T21:34:08.326Z</news:publication_date>
      <news:title>Negligência de Negação Eleva Taxa de Crenças Falsas para 88,6% em LLMs Fine-Tuned</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/negation-neglect-llms-can-learn-false-claims-from-negation-heavy-fine-tuning</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T21:34:08.326Z</news:publication_date>
      <news:title>Negligencia de Negación Eleva Tasa de Creencias Falsas a 88,6% en LLMs Fine-Tuned</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/ai-harness-runtime-substrate-for-reliable-software-engineering-agents</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T20:54:08.296Z</news:publication_date>
      <news:title>Why Production Agents Fail Without Harness Infrastructure</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/ai-harness-runtime-substrate-for-reliable-software-engineering-agents</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T20:54:08.296Z</news:publication_date>
      <news:title>Por que Agentes em Produção Falham Sem Infraestrutura de Harness</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/ai-harness-runtime-substrate-for-reliable-software-engineering-agents</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T20:54:08.296Z</news:publication_date>
      <news:title>Por Qué los Agentes en Producción Fallan Sin Infraestructura de Harness</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/real-time-ai-agents-demand-asynchronous-io-and-speculative-execution</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T20:22:08.395Z</news:publication_date>
      <news:title>Berkeley Framework Cuts Agent Latency 1.3–2.2×</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/real-time-ai-agents-demand-asynchronous-io-and-speculative-execution</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T20:22:08.395Z</news:publication_date>
      <news:title>Framework de Berkeley Reduz Latência de Agentes 1.3–2.2×</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/real-time-ai-agents-demand-asynchronous-io-and-speculative-execution</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T20:22:08.395Z</news:publication_date>
      <news:title>Framework de Berkeley Reduce Latencia de Agentes 1.3–2.2×</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/ciscos-ai-surge-lifts-stock-14-despite-massive-workforce-cuts</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T19:30:08.246Z</news:publication_date>
      <news:title>Cisco&apos;s $9 billion AI orders lift stock 15% on record quarter</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/ciscos-ai-surge-lifts-stock-14-despite-massive-workforce-cuts</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T19:30:08.246Z</news:publication_date>
      <news:title>Pedidos de IA de $9 bilhões da Cisco impulsionam ações 15% em trimestre recorde</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/ciscos-ai-surge-lifts-stock-14-despite-massive-workforce-cuts</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T19:30:08.246Z</news:publication_date>
      <news:title>Pedidos de IA de $9 mil millones de Cisco impulsan acciones 15% en trimestre récord</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/forced-ai-coding-mandates-are-de-skilling-developers-and-stacking-tech-debt-work</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T18:50:09.069Z</news:publication_date>
      <news:title>AI Code Mandates Drive 10x Security Findings Spike</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/forced-ai-coding-mandates-are-de-skilling-developers-and-stacking-tech-debt-work</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T18:50:09.069Z</news:publication_date>
      <news:title>Mandatos de IA no Código Disparam Descobertas de Segurança em 10x</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/forced-ai-coding-mandates-are-de-skilling-developers-and-stacking-tech-debt-work</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T18:50:09.069Z</news:publication_date>
      <news:title>Mandatos de IA en Código Disparan Hallazgos de Seguridad en 10x</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/building-multi-agent-systems-practical-lessons-and-pitfalls</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T18:10:08.231Z</news:publication_date>
      <news:title>Shopify Swarm Cuts Theme Review from 22 Hours to 20 Minutes</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/building-multi-agent-systems-practical-lessons-and-pitfalls</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T18:10:08.231Z</news:publication_date>
      <news:title>Shopify Swarm Reduz Revisão de Tema de 22 Horas para 20 Minutos</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/building-multi-agent-systems-practical-lessons-and-pitfalls</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-16T18:10:08.231Z</news:publication_date>
      <news:title>Shopify Swarm Reduce Revisión de Tema de 22 Horas a 20 Minutos</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/aws-workspaces-ai-agents-now-control-legacy-desktop-apps-without-apis</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T03:06:48.566Z</news:publication_date>
      <news:title>AI Agents Can Now Access Any Desktop App Without APIs</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/aws-workspaces-ai-agents-now-control-legacy-desktop-apps-without-apis</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T03:06:48.566Z</news:publication_date>
      <news:title>Agentes de IA Podem Acessar Qualquer Aplicação Desktop Sem APIs</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/aws-workspaces-ai-agents-now-control-legacy-desktop-apps-without-apis</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T03:06:48.566Z</news:publication_date>
      <news:title>Agentes de IA Pueden Acceder a Cualquier Aplicación Desktop Sin APIs</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/kv-fold-one-step-protocol-extends-context-windows-without-retraining</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T02:35:38.656Z</news:publication_date>
      <news:title>KV-Fold Extends Transformer Context to 128K Without Retraining</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/kv-fold-one-step-protocol-extends-context-windows-without-retraining</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T02:35:38.656Z</news:publication_date>
      <news:title>KV-Fold Estende Contexto de Transformers até 128K Sem Retreinamento</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/kv-fold-one-step-protocol-extends-context-windows-without-retraining</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T02:35:38.656Z</news:publication_date>
      <news:title>KV-Fold Extiende el Contexto de Transformers a 128K sin Reentrenamiento</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/task-adaptive-embeddings-llm-guided-query-refinement-for-zero-shot-search</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T02:05:37.596Z</news:publication_date>
      <news:title>IBM Boosts Zero-Shot Search Accuracy 25% With LLM Query Refinement</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/task-adaptive-embeddings-llm-guided-query-refinement-for-zero-shot-search</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T02:05:37.596Z</news:publication_date>
      <news:title>IBM Aumenta Precisão de Busca Zero-Shot 25% Com Refinamento de Query via LLM</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/task-adaptive-embeddings-llm-guided-query-refinement-for-zero-shot-search</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T02:05:37.596Z</news:publication_date>
      <news:title>IBM Aumenta Precisión de Búsqueda Zero-Shot 25% Con Refinamiento de Consulta Mediante LLM</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/attractor-models-iterative-refinement-for-stable-looped-reasoning</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T01:34:48.650Z</news:publication_date>
      <news:title>27M Attractor Model Beats GPT o3 on Logic Puzzles</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/attractor-models-iterative-refinement-for-stable-looped-reasoning</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T01:34:48.650Z</news:publication_date>
      <news:title>Modelo Attractor de 27M Supera GPT o3 em Quebra-Cabeças de Lógica</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/attractor-models-iterative-refinement-for-stable-looped-reasoning</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T01:34:48.650Z</news:publication_date>
      <news:title>Modelo Attractor de 27M Supera GPT o3 en Rompecabezas de Lógica</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/reward-hacking-in-rubric-based-rl-how-post-training-verifiers-can-mislead</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T01:02:48.773Z</news:publication_date>
      <news:title>Reward Hacking Undetected in Single-Verifier Training</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/reward-hacking-in-rubric-based-rl-how-post-training-verifiers-can-mislead</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T01:02:48.773Z</news:publication_date>
      <news:title>Reward Hacking Não Detectado no Treinamento com Verificador Único</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/reward-hacking-in-rubric-based-rl-how-post-training-verifiers-can-mislead</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T01:02:48.773Z</news:publication_date>
      <news:title>Reward Hacking No Detectado en Entrenamiento con Verificador Único</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/beyond-grpo-sparse-to-dense-reward-allocation-for-efficient-llm-post-training</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T00:30:48.532Z</news:publication_date>
      <news:title>Sparse-to-Dense RL Lifts MATH Scores to 78.5% on Small Models</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/beyond-grpo-sparse-to-dense-reward-allocation-for-efficient-llm-post-training</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T00:30:48.532Z</news:publication_date>
      <news:title>Aprendizado Sparse-to-Dense Eleva Scores MATH para 78.5% em Modelos Pequenos</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/beyond-grpo-sparse-to-dense-reward-allocation-for-efficient-llm-post-training</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T00:30:48.532Z</news:publication_date>
      <news:title>Aprendizaje Sparse-to-Dense Eleva Scores MATH a 78.5% en Modelos Pequeños</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/en/article/sparse-mixture-of-experts-routers-show-geometric-couplingnew-path-to-stable-smoe</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T00:00:38.325Z</news:publication_date>
      <news:title>Standard load-balancing losses degrade SMoE expert specialization by 3x</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/pt/article/sparse-mixture-of-experts-routers-show-geometric-couplingnew-path-to-stable-smoe</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>pt-BR</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T00:00:38.325Z</news:publication_date>
      <news:title>Perdas padrão de balanceamento de carga degradam especialização de experts em SMoE em 3x</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://aiexpert.news/es/article/sparse-mixture-of-experts-routers-show-geometric-couplingnew-path-to-stable-smoe</loc>
    <news:news>
      <news:publication>
        <news:name>ai|expert</news:name>
        <news:language>es</news:language>
      </news:publication>
      <news:publication_date>2026-05-15T00:00:38.325Z</news:publication_date>
      <news:title>Las pérdidas estándar de equilibrio de carga degradan la especialización de expertos en SMoE en 3x</news:title>
    </news:news>
  </url>
</urlset>