<?xml version="1.0" encoding="UTF-8"?>
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9"
        xmlns:news="http://www.google.com/schemas/sitemap-news/0.9">
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/arxiv-adamezo-zeroth-order-llm-finetuning/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>AdaMeZO: fino ugađanje LLM-ova Adam-stilom bez pohrane momenata u GPU memoriji</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/arxiv-aem-adaptive-entropy-modulation-rl-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv AEM: adaptivna modulacija entropije za multi-turn RL agente postiže +1,4 % na SWE-bench Verified</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/arxiv-armor-2025-vojni-llm-sigurnost/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv ARMOR 2025: prvi vojni benchmark za LLM sigurnost s 519 promptova kroz 21 komercijalni model</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/arxiv-bayes-consistent-agentic-orchestration-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>Position paper s 30 autora na ICML 2026: orkestracija agentnih AI sustava mora biti Bayes-konzistentna</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/arxiv-bwla-w1ax-kvantizacija-llm/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>BWLA: 1-bitna kvantizacija LLM-ova s 3,26× ubrzanjem i 70% boljim rezultatima (ACL 2026)</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/arxiv-stable-gflownet-llm-red-teaming-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026 Spotlight: Stable-GFlowNet uvodi stabilnije i raznovrsnije automatizirano red-teamanje LLM-ova</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/arxiv-tool-calling-framework-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv okvir &apos;To Call or Not to Call&apos; otkriva da LLM-ovi pogrešno procjenjuju kad im trebaju vanjski alati</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/arxiv-tool-use-tax-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv: skriveni trošak alata u LLM agentima — &quot;tool-use tax&quot; smanjuje točnost čak i kad alati pomažu</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-04/ibm-ceo-study-c-suite-ai-restrukturiranje/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>IBM studija: 76 % organizacija ima Chief AI Officera, CEO-i očekuju 48 % autonomnih AI odluka do 2030.</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/arxiv-adamezo-zeroth-order-llm-finetuning/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>AdaMeZO: Adam-style LLM fine-tuning without storing gradient moments in GPU memory</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/arxiv-aem-adaptive-entropy-modulation-rl-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv AEM: Adaptive Entropy Modulation for multi-turn RL agents achieves +1.4% on SWE-bench Verified</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/arxiv-armor-2025-vojni-llm-sigurnost/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv ARMOR 2025: first military LLM safety benchmark with 519 prompts across 21 commercial models</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/arxiv-bayes-consistent-agentic-orchestration-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>Position paper by 30 authors at ICML 2026: agentic AI orchestration must be Bayes-consistent</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/arxiv-bwla-w1ax-kvantizacija-llm/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>BWLA: 1-bit LLM quantization with 3.26× speedup and 70% better results (ACL 2026)</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/arxiv-stable-gflownet-llm-red-teaming-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026 Spotlight: Stable-GFlowNet introduces more stable and diverse automated LLM red-teaming</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/arxiv-tool-calling-framework-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv &apos;To Call or Not to Call&apos; framework reveals LLMs misjudge when they need external tools</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/arxiv-tool-use-tax-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv: the hidden cost of tools in LLM agents — &apos;tool-use tax&apos; reduces accuracy even when tools help</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-04/ibm-ceo-study-c-suite-ai-restrukturiranje/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>IBM study: 76% of organizations have a Chief AI Officer, CEOs expect 48% autonomous AI decisions by 2030</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/arxiv-adamezo-zeroth-order-llm-finetuning/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>AdaMeZO: Adam-Stil LLM-Fine-Tuning ohne Speicherung von Gradientenmomenten im GPU-Speicher</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/arxiv-aem-adaptive-entropy-modulation-rl-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv AEM: Adaptive Entropiemodulation für Multi-Turn-RL-Agenten erreicht +1,4 % auf SWE-bench Verified</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/arxiv-armor-2025-vojni-llm-sigurnost/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv ARMOR 2025: erster militärischer LLM-Sicherheitsbenchmark mit 519 Prompts über 21 kommerzielle Modelle</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/arxiv-bayes-consistent-agentic-orchestration-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>Position Paper von 30 Autoren auf ICML 2026: Orchestrierung agentischer KI-Systeme muss Bayes-konsistent sein</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/arxiv-bwla-w1ax-kvantizacija-llm/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>BWLA: 1-Bit-Quantisierung von Sprachmodellen mit 3,26-facher Beschleunigung und 70 % besseren Ergebnissen (ACL 2026)</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/arxiv-stable-gflownet-llm-red-teaming-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026 Spotlight: Stable-GFlowNet führt stabileres und vielfältigeres automatisiertes Red-Teaming von Sprachmodellen ein</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/arxiv-tool-calling-framework-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv-Rahmen &apos;To Call or Not to Call&apos; zeigt: Sprachmodelle beurteilen falsch, wann sie externe Werkzeuge brauchen</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/arxiv-tool-use-tax-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv: die versteckten Kosten von Werkzeugen in LLM-Agenten — &apos;Tool-Use Tax&apos; senkt Genauigkeit selbst wenn Werkzeuge helfen</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-04/ibm-ceo-study-c-suite-ai-restrukturiranje/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>IBM-Studie: 76 % der Unternehmen haben einen Chief AI Officer, CEOs erwarten 48 % autonome KI-Entscheidungen bis 2030</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/arxiv-adamezo-zeroth-order-llm-finetuning/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>AdaMeZO：以类Adam方式微调LLM，无需在GPU内存中存储动量</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/arxiv-aem-adaptive-entropy-modulation-rl-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv AEM：多轮RL智能体的自适应熵调制在SWE-bench Verified上提升+1.4%</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/arxiv-armor-2025-vojni-llm-sigurnost/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv ARMOR 2025：519个提示词测试21个商业LLM的军事安全性基准</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/arxiv-bayes-consistent-agentic-orchestration-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026 立场论文：30位作者认为智能体AI系统的编排必须符合贝叶斯一致性</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/arxiv-bwla-w1ax-kvantizacija-llm/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>BWLA：1位量化LLM实现3.26倍加速和70%更好结果（ACL 2026）</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/arxiv-stable-gflownet-llm-red-teaming-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026 Spotlight：Stable-GFlowNet引入更稳定、更多样化的LLM自动化红队测试</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/arxiv-tool-calling-framework-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv框架「是否调用」揭示LLM错误判断何时需要外部工具</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/arxiv-tool-use-tax-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv：LLM智能体工具的隐性成本——「工具使用税」即使工具有帮助也会降低准确性</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-04/ibm-ceo-study-c-suite-ai-restrukturiranje/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>IBM研究：76%的组织有首席AI官，CEO预计到2030年48%的运营决策将由AI自主作出</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/arxiv-adamezo-zeroth-order-llm-finetuning/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>AdaMeZO：GPU メモリにモーメントを保存せずAdam方式でLLMをファインチューニング</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/arxiv-aem-adaptive-entropy-modulation-rl-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv AEM：マルチターンRL エージェントの適応的エントロピー変調がSWE-bench Verifiedで+1.4%を達成</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/arxiv-armor-2025-vojni-llm-sigurnost/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv ARMOR 2025：519の軍事プロンプトで21の商用LLMの安全性を評価する初の軍事ベンチマーク</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/arxiv-bayes-consistent-agentic-orchestration-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026立場論文：30名の著者がエージェントAIのオーケストレーションはベイズ一貫性を持つべきと主張</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/arxiv-bwla-w1ax-kvantizacija-llm/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>BWLA：1ビット量子化LLMで3.26倍の高速化と70%の改善を達成（ACL 2026）</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/arxiv-stable-gflownet-llm-red-teaming-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026 Spotlight：Stable-GFlowNetがより安定した多様なLLM自動レッドチーミングを実現</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/arxiv-tool-calling-framework-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXivフレームワーク「呼ぶべきか否か」がLLMの外部ツール判断ミスを明らかに</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/arxiv-tool-use-tax-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv：LLMエージェントのツールの隠れたコスト——「ツール使用税」はツールが役立つ時でも精度を下げる</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-04/ibm-ceo-study-c-suite-ai-restrukturiranje/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>IBM調査：76%の組織がChief AI Officerを設置、CEOは2030年までにAIが48%の運営決定を自律的に下すと予測</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/arxiv-adamezo-zeroth-order-llm-finetuning/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>AdaMeZO: GPU 메모리에 모멘트를 저장하지 않고 Adam 방식으로 LLM 파인튜닝하는 새 최적화기</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/arxiv-aem-adaptive-entropy-modulation-rl-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv AEM: 멀티턴 RL 에이전트를 위한 적응형 엔트로피 변조, SWE-bench Verified에서 +1.4% 향상</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/arxiv-armor-2025-vojni-llm-sigurnost/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv ARMOR 2025: 519개 프롬프트로 21개 상용 LLM의 군사 안전성을 평가한 최초의 군사 벤치마크</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/arxiv-bayes-consistent-agentic-orchestration-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026 포지션 페이퍼: 저자 30명, 에이전트 AI 오케스트레이션은 베이즈 일관성을 가져야 한다</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/arxiv-bwla-w1ax-kvantizacija-llm/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>BWLA: 1비트 양자화 LLM으로 3.26배 가속 및 70% 향상 달성 (ACL 2026)</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/arxiv-stable-gflownet-llm-red-teaming-icml/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ICML 2026 Spotlight: Stable-GFlowNet, 더 안정적이고 다양한 LLM 자동화 레드팀 테스트 도입</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/arxiv-tool-calling-framework-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv 프레임워크 &apos;호출할 것인가 말 것인가&apos;: LLM이 외부 도구 필요성을 잘못 판단한다는 것을 밝혀</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/arxiv-tool-use-tax-llm-agenti/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>ArXiv: LLM 에이전트 도구의 숨겨진 비용 - &apos;도구 사용세&apos;는 도구가 도움이 될 때도 정확도를 낮춘다</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-04/ibm-ceo-study-c-suite-ai-restrukturiranje/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-04T00:00:00Z</news:publication_date>
      <news:title>IBM 연구: 조직의 76%가 최고AI책임자를 보유, CEO들은 2030년까지 AI가 운영 결정의 48%를 자율적으로 내릴 것으로 예상</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/arxiv-exploration-hacking-rl-resistance/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Exploration Hacking: mogu li LLM-ovi naučiti opirati se RL treningu i strategijski potiskivati vlastite sposobnosti?</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/arxiv-kellybench-premier-league-decisions/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>KellyBench: AI agenti upravljali kladioničarskim bankrollom u Premier Ligi — svi vodeći modeli izgubili novac</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/arxiv-latent-grpo-reasoning-rl/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Latent-GRPO: stabilna RL optimizacija za latent reasoning — 7,86 boda na GSM8K-Aug i 4,27 boda na AIME uz 3-4× kraće reasoning chain-ove</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/arxiv-mcphunt-mcp-credential-propagation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>MCPHunt: prvi benchmark koji mjeri curenje vjerodajnica između granica povjerenja u multi-server MCP agentima — stope 11,5–41,3 %</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/arxiv-msr-synthetic-computers-1000-scale/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Microsoft Research Synthetic Computers: 1 000 sintetičkih računala kao supstrat za long-horizon trening produktivnih AI agenata</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/aws-transform-bi-migration-bedrock-agentcore/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>AWS Transform automatizira migraciju BI dashboarda iz Tableau i Power BI u QuickSight za dane umjesto mjeseci</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/github-copilot-gpt-52-deprecation-june-2026/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>GitHub povlači GPT-5.2 i GPT-5.2-Codex iz Copilota 1. lipnja 2026. — migracija na GPT-5.5 i GPT-5.3-Codex</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/google-research-open-science-250k-reach/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Google Research open-source alati dosegnuli 250.000 istraživača: od genoma do monsunskih prognoza za 38 milijuna farmera</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/hr/news/2026-05-02/nist-caisi-deepseek-v4-pro-evaluation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>hr</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>NIST CAISI evaluacija DeepSeek V4 Pro: 8 mjeseci zaostatka za frontier US modelima na 9 benchmarka u 5 domena</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/arxiv-exploration-hacking-rl-resistance/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Exploration Hacking: Can LLMs Learn to Resist RL Training and Strategically Suppress Their Own Capabilities?</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/arxiv-kellybench-premier-league-decisions/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>KellyBench: AI agents managing a betting bankroll through the Premier League season — all leading models lost money</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/arxiv-latent-grpo-reasoning-rl/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Latent-GRPO: Stable RL Optimization for Latent Reasoning — 7.86 Points on GSM8K-Aug and 4.27 Points on AIME With 3-4× Shorter Reasoning Chains</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/arxiv-mcphunt-mcp-credential-propagation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>MCPHunt: first benchmark measuring credential leakage across trust boundaries in multi-server MCP agents — rates of 11.5–41.3%</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/arxiv-msr-synthetic-computers-1000-scale/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Microsoft Research Synthetic Computers: 1,000 synthetic computers as a substrate for long-horizon training of productive AI agents</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/aws-transform-bi-migration-bedrock-agentcore/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>AWS Transform automates BI dashboard migration from Tableau and Power BI to QuickSight in days instead of months</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/github-copilot-gpt-52-deprecation-june-2026/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>GitHub is retiring GPT-5.2 and GPT-5.2-Codex from Copilot on June 1, 2026 — migration to GPT-5.5 and GPT-5.3-Codex</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/google-research-open-science-250k-reach/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Google Research open-source tools reach 250,000 researchers: from genomes to monsoon forecasts for 38 million farmers</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/en/news/2026-05-02/nist-caisi-deepseek-v4-pro-evaluation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>en</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>NIST CAISI evaluation of DeepSeek V4 Pro: 8-month lag behind frontier US models across 9 benchmarks in 5 domains</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/arxiv-exploration-hacking-rl-resistance/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Exploration Hacking: Können LLMs lernen, sich dem RL-Training zu widersetzen und ihre eigenen Fähigkeiten strategisch zu unterdrücken?</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/arxiv-kellybench-premier-league-decisions/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>KellyBench: KI-Agenten verwalten Wett-Bankroll durch die Premier-League-Saison — alle führenden Modelle verloren Geld</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/arxiv-latent-grpo-reasoning-rl/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Latent-GRPO: Stabile RL-Optimierung für Latent Reasoning — 7,86 Punkte auf GSM8K-Aug und 4,27 Punkte auf AIME bei 3-4× kürzeren Reasoning-Ketten</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/arxiv-mcphunt-mcp-credential-propagation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>MCPHunt: erster Benchmark zur Messung von Credential-Leakage über Vertrauensgrenzen in Multi-Server-MCP-Agenten — Raten von 11,5–41,3 %</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/arxiv-msr-synthetic-computers-1000-scale/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Microsoft Research Synthetic Computers: 1.000 synthetische Computer als Substrat für das Long-Horizon-Training produktiver KI-Agenten</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/aws-transform-bi-migration-bedrock-agentcore/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>AWS Transform automatisiert BI-Dashboard-Migration von Tableau und Power BI nach QuickSight in Tagen statt Monaten</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/github-copilot-gpt-52-deprecation-june-2026/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>GitHub stellt GPT-5.2 und GPT-5.2-Codex in Copilot am 1. Juni 2026 ein — Migration auf GPT-5.5 und GPT-5.3-Codex</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/google-research-open-science-250k-reach/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Google Research Open-Source-Tools erreichen 250.000 Forscher: von Genomen bis zu Monsun-Prognosen für 38 Millionen Landwirte</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/de/news/2026-05-02/nist-caisi-deepseek-v4-pro-evaluation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>de</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>NIST CAISI-Evaluierung von DeepSeek V4 Pro: 8 Monate Rückstand gegenüber US-Frontier-Modellen in 9 Benchmarks und 5 Domänen</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/arxiv-exploration-hacking-rl-resistance/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>探索黑客攻击：大语言模型能否学会抵抗强化学习训练并战略性地压制自身能力？</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/arxiv-kellybench-premier-league-decisions/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>KellyBench：AI代理管理Premier League赛季投注资金——所有顶级模型均亏损</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/arxiv-latent-grpo-reasoning-rl/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Latent-GRPO：面向潜在推理的稳定 RL 优化——GSM8K-Aug 上提升 7.86 分、AIME 上提升 4.27 分，推理链长度缩短 3-4 倍</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/arxiv-mcphunt-mcp-credential-propagation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>MCPHunt：首个衡量多服务器MCP代理信任边界凭证泄露的基准测试——泄露率11.5%至41.3%</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/arxiv-msr-synthetic-computers-1000-scale/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Microsoft Research Synthetic Computers：1000台合成计算机作为长视程生产力AI代理训练的基础底层</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/aws-transform-bi-migration-bedrock-agentcore/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>AWS Transform利用AI代理将Tableau和Power BI仪表盘自动迁移至QuickSight，耗时从月缩短至天</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/github-copilot-gpt-52-deprecation-june-2026/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>GitHub将于2026年6月1日从Copilot中弃用GPT-5.2和GPT-5.2-Codex——迁移至GPT-5.5和GPT-5.3-Codex</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/google-research-open-science-250k-reach/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Google Research开源工具触及25万研究人员：从基因组学到为3800万农民提供季风预报</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/zh/news/2026-05-02/nist-caisi-deepseek-v4-pro-evaluation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>zh</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>NIST CAISI对DeepSeek V4 Pro的评估：在5个领域9个基准测试中落后美国前沿模型8个月</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/arxiv-exploration-hacking-rl-resistance/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>探索ハッキング:LLMはRL訓練に抵抗し、自らの能力を戦略的に抑制することを学べるのか</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/arxiv-kellybench-premier-league-decisions/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>KellyBench：AIエージェントがPremier Leagueシーズンを通じて賭けの資金を管理——主要モデルはすべて損失</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/arxiv-latent-grpo-reasoning-rl/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Latent-GRPO:潜在推論のための安定したRL最適化——GSM8K-Augで7.86ポイント、AIMEで4.27ポイント向上、推論チェーンは3-4倍短縮</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/arxiv-mcphunt-mcp-credential-propagation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>MCPHunt：マルチサーバーMCPエージェントの信頼境界を越えた認証情報漏洩を測定する初のベンチマーク——漏洩率11.5〜41.3%</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/arxiv-msr-synthetic-computers-1000-scale/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Microsoft Research Synthetic Computers：長期生産性AIエージェントトレーニングの基盤として1,000台の合成コンピューターを活用</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/aws-transform-bi-migration-bedrock-agentcore/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>AWS TransformがTableauとPower BIのダッシュボードをQuickSightへ自動移行——数か月かかる作業を数日に短縮</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/github-copilot-gpt-52-deprecation-june-2026/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>GitHubが2026年6月1日にCopilotからGPT-5.2とGPT-5.2-Codexを廃止——GPT-5.5とGPT-5.3-Codexへの移行</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/google-research-open-science-250k-reach/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Google Researchのオープンソースツールが25万人の研究者に到達：ゲノミクスから3,800万人の農家への季節風予報まで</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ja/news/2026-05-02/nist-caisi-deepseek-v4-pro-evaluation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>NIST CAISI によるDeepSeek V4 Pro評価：5分野9ベンチマークで米国フロンティアモデルに8か月遅れ</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/arxiv-exploration-hacking-rl-resistance/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>탐색 해킹: LLM은 RL 훈련에 저항하고 자신의 능력을 전략적으로 억제하는 법을 배울 수 있는가?</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/arxiv-kellybench-premier-league-decisions/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>KellyBench: AI 에이전트가 Premier League 시즌 내내 베팅 자금을 관리——모든 주요 모델이 손실</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/arxiv-latent-grpo-reasoning-rl/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Latent-GRPO: 잠재 추론을 위한 안정적 RL 최적화 — GSM8K-Aug에서 7.86점, AIME에서 4.27점 향상, 추론 체인은 3-4배 단축</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/arxiv-mcphunt-mcp-credential-propagation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>MCPHunt: 다중 서버 MCP 에이전트의 신뢰 경계 간 자격증명 유출을 측정하는 최초의 벤치마크——유출률 11.5~41.3%</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/arxiv-msr-synthetic-computers-1000-scale/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Microsoft Research Synthetic Computers: 장기 생산성 AI 에이전트 훈련의 기반으로서 1,000개의 합성 컴퓨터</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/aws-transform-bi-migration-bedrock-agentcore/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>AWS Transform, Tableau 및 Power BI 대시보드를 QuickSight로 자동 마이그레이션——수개월 작업을 수일로 단축</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/github-copilot-gpt-52-deprecation-june-2026/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>GitHub, 2026년 6월 1일 Copilot에서 GPT-5.2 및 GPT-5.2-Codex 지원 종료——GPT-5.5 및 GPT-5.3-Codex로 마이그레이션</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/google-research-open-science-250k-reach/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>Google Research 오픈소스 도구, 25만 명의 연구자에게 도달: 유전체학부터 3,800만 농부를 위한 몬순 예보까지</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://24-ai.news/ko/news/2026-05-02/nist-caisi-deepseek-v4-pro-evaluation/</loc>
    <news:news>
      <news:publication>
        <news:name>24 AI</news:name>
        <news:language>ko</news:language>
      </news:publication>
      <news:publication_date>2026-05-02T00:00:00Z</news:publication_date>
      <news:title>NIST CAISI의 DeepSeek V4 Pro 평가: 5개 영역 9개 벤치마크에서 미국 프론티어 모델보다 8개월 뒤처짐</news:title>
    </news:news>
  </url>
</urlset>
