<?xml version="1.0" encoding="UTF-8"?>
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9"
        xmlns:news="http://www.google.com/schemas/sitemap-news/0.9">
  <url>
    <loc>https://ai.whytrend.jp/articles/b2b169c50e36_20260429.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-29T00:02:18.640142+09:00</news:publication_date>
      <news:title>Workflows for work that runs the business</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/d733638772d9_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T16:01:41.337509+09:00</news:publication_date>
      <news:title>SpecRLBench: A Benchmark for Generalization in Specification-Guided Reinforcement Learning</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/7f86f7363edf_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T15:06:12.280394+09:00</news:publication_date>
      <news:title>Learning to Think from Multiple Thinkers</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/d56759c680b1_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T15:01:56.086530+09:00</news:publication_date>
      <news:title>The Optimal Sample Complexity of Multiclass and List Learning</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/11ed6d592f59_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T14:05:37.504924+09:00</news:publication_date>
      <news:title>Personalized Worked Example Generation from Student Code Submissions using Pattern-based Knowledge Components</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/fd006d88d302_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T14:01:59.842936+09:00</news:publication_date>
      <news:title>Conflict-Aware Harmonized Rotational Gradient for Multiscale Kinetic Regimes</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/d07682c7ab7c_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T13:06:17.717074+09:00</news:publication_date>
      <news:title>K-MetBench: A Multi-Dimensional Benchmark for Fine-Grained Evaluation of Expert Reasoning, Locality, and Multimodality in Meteorology</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/6d672cf7f1e8_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T13:02:15.309874+09:00</news:publication_date>
      <news:title>The Last Human-Written Paper: Agent-Native Research Artifacts</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/5f03b58113a9_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T12:07:00.097697+09:00</news:publication_date>
      <news:title>SFT-then-RL Outperforms Mixed-Policy Methods for LLM Reasoning</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/c62f5ddcae6d_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T12:02:18.693856+09:00</news:publication_date>
      <news:title>The Override Gap: A Magnitude Account of Knowledge Conflict Failure in Hypernetwork-Based Instant LLM Adaptation</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/61e43bc87c11_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T11:06:13.454974+09:00</news:publication_date>
      <news:title>Modeling Induced Pleasure through Cognitive Appraisal Prediction via Multimodal Fusion</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/48d078159bf5_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T11:01:59.438247+09:00</news:publication_date>
      <news:title>Agentic Fusion of Large Atomic and Language Models to Accelerate Materials Discovery</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/44dc126324e2_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T10:01:52.129740+09:00</news:publication_date>
      <news:title>Adaptive Ultrasound Imaging with Physics-Informed NV-Raw2Insights-US AI</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/5a13ffc4196e_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T09:01:57.761877+09:00</news:publication_date>
      <news:title>Amazon Redshift Serverless AI-driven scaling is now the default for new workgroups</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/bbe03dc901ea_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T08:02:00.287655+09:00</news:publication_date>
      <news:title>Amazon FSx for OpenZFS Single-AZ (HA) file systems are now available in 17 additional AWS commercial and AWS GovCloud (US) Regions</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/3d491f95704f_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T07:01:58.097269+09:00</news:publication_date>
      <news:title>OpenAI available at FedRAMP Moderate</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/453691ea92d9_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T04:01:51.540054+09:00</news:publication_date>
      <news:title>Choco automates food distribution with AI agents</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/1749273ae238_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T03:06:42.247859+09:00</news:publication_date>
      <news:title>Amazon SageMaker HyperPod now supports G7e and r5d.16xlarge instances</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/4944f7d5951d_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T03:02:24.038682+09:00</news:publication_date>
      <news:title>An open-source spec for orchestration: Symphony</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/17d34d695b1c_20260428.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-28T00:01:42.861507+09:00</news:publication_date>
      <news:title>How to build scalable web apps with OpenAI's Privacy Filter</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/3083a1a3d347_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T23:06:35.416491+09:00</news:publication_date>
      <news:title>Working with Codex</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/28e4e9faafa2_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T23:02:35.270061+09:00</news:publication_date>
      <news:title>The next phase of the Microsoft OpenAI partnership</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/015c116ef337_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T17:01:54.642375+09:00</news:publication_date>
      <news:title>What is Codex?</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/59c4f55ac643_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T14:01:40.444297+09:00</news:publication_date>
      <news:title>Agentic World Modeling: Foundations, Capabilities, Laws, and Beyond</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/c04417a60213_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T13:05:46.848315+09:00</news:publication_date>
      <news:title>Representational Harms in LLM-Generated Narratives Against Global Majority Nationalities</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/c4cab7aca315_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T13:02:07.225230+09:00</news:publication_date>
      <news:title>How Do AI Agents Spend Your Money? Analyzing and Predicting Token Consumption in Agentic Coding Tasks</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/2f1186edd5c0_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T12:06:01.769443+09:00</news:publication_date>
      <news:title>Spend Less, Fit Better: Budget-Efficient Scaling Law Fitting via Active Experiment Selection</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/16039a771484_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T12:01:59.768309+09:00</news:publication_date>
      <news:title>Relaxation-Informed Training of Neural Network Surrogate Models</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/9a6ae8a397c2_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T09:08:53.096541+09:00</news:publication_date>
      <news:title>How to get started with Codex</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/27e32bfa00fa_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T09:02:05.990745+09:00</news:publication_date>
      <news:title>Our principles</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/7d0ae2883855_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T08:01:29.507451+09:00</news:publication_date>
      <news:title>デジタル庁：行政の進化と革新のための生成AIの調達・利活用ガイドライン</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/11ac33be56ca_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T07:06:00.011295+09:00</news:publication_date>
      <news:title>デジタル庁：先進的AI利活用アドバイザリーボード</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/2afaa1d9e909_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T07:02:08.438106+09:00</news:publication_date>
      <news:title>デジタル庁：ガバメントAI「源内」</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/6bf4b7ab1bdf_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T06:05:30.386733+09:00</news:publication_date>
      <news:title>European AI Office</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/2b2b904fe28e_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T06:02:04.282705+09:00</news:publication_date>
      <news:title>AI Act</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/67dca01ed8ba_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T05:05:46.052713+09:00</news:publication_date>
      <news:title>Speaking of Voxtral</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/cc7c349b8085_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T05:01:52.546567+09:00</news:publication_date>
      <news:title>Autonomous AI at Scale: Adobe Agents Unlock Breakthrough Creative Intelligence With NVIDIA and WPP</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/701af37f9866_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T04:05:42.289345+09:00</news:publication_date>
      <news:title>AI and the Future of Cybersecurity: Why Openness Matters</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/e452b6672db0_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T04:02:03.920795+09:00</news:publication_date>
      <news:title>QIMMA قِمّة ⛰: A Quality-First Arabic LLM Leaderboard</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/1fb64e59b1c5_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T03:05:59.986353+09:00</news:publication_date>
      <news:title>NVIDIA and Google Cloud Collaborate to Advance Agentic and Physical AI</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/0e66b46ad812_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T03:02:02.233626+09:00</news:publication_date>
      <news:title>From Rainforests to Recycling Plants: 5 Ways NVIDIA AI Is Protecting the Planet</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/caa7ef5f9476_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T02:02:38.173701+09:00</news:publication_date>
      <news:title>Gemma 4 VLA Demo on Jetson Orin Nano Super</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/d08658224381_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T02:02:04.819124+09:00</news:publication_date>
      <news:title>How to Use Transformers.js in a Chrome Extension</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/65f9ed8829d2_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T01:06:36.607894+09:00</news:publication_date>
      <news:title>Top 10 uses for Codex at work</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/00bb844fad2b_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T01:02:46.745110+09:00</news:publication_date>
      <news:title>Plugins and skills</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/e7e42070ea8a_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T00:08:38.772139+09:00</news:publication_date>
      <news:title>Making Sense of the Early Universe</news:title>
    </news:news>
  </url>
  <url>
    <loc>https://ai.whytrend.jp/articles/193002765aa7_20260427.html</loc>
    <news:news>
      <news:publication>
        <news:name>WhyTrend</news:name>
        <news:language>ja</news:language>
      </news:publication>
      <news:publication_date>2026-04-27T00:08:24.016665+09:00</news:publication_date>
      <news:title>Automations</news:title>
    </news:news>
  </url>
</urlset>