<?xml version="1.0" encoding="UTF-8" ?>
		<urlset
			xmlns="https://www.sitemaps.org/schemas/sitemap/0.9"
			xmlns:xhtml="https://www.w3.org/1999/xhtml"
			xmlns:mobile="https://www.google.com/schemas/sitemap-mobile/1.0"
			xmlns:news="https://www.google.com/schemas/sitemap-news/0.9"
			xmlns:image="https://www.google.com/schemas/sitemap-image/1.1"
			xmlns:video="https://www.google.com/schemas/sitemap-video/1.1"
		>
      
    <url>
      <loc>https://hackernoon.com/optimizing-language-models-decoding-griffins-local-attention-and-memory-efficiency</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/overcoming-hbm-vmem-bottlenecks-in-tpu-v3-recurrent-workloads</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/hawk-and-griffin-efficient-rnn-models-redefining-ai-performance</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/rnns-vs-transformers-innovations-in-scalability-and-efficiency</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/hawk-and-griffin-mastering-long-context-extrapolation-in-ai</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/griffin-model-advancing-copying-and-retrieval-in-ai-tasks</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/hawk-and-griffin-models-superior-latency-and-throughput-in-ai-inference</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/recurrent-models-enhancing-latency-and-throughput-efficiency</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/recurrent-models-decoding-faster-with-lower-latency-and-higher-throughput</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/training-speed-on-longer-sequences</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/efficient-linear-recurrences-on-device</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/efficient-training-scaling-griffin-models-for-large-scale-ai-on-tpus</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/hawk-and-griffin-models-superior-nlp-performance-with-minimal-training-data</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/griffin-models-outperforming-transformers-with-scalable-ai-innovation</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/recurrent-models-scale-as-efficiently-as-transformers</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/rg-lru-a-breakthrough-recurrent-layer-redefining-nlp-model-efficiency</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/rnn-models-hawk-and-griffin-transforming-nlp-efficiency-and-scaling</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
            <url>
            <loc>https://gating.tech/about</loc>
              <priority>0.8</priority>
            <changefreq>daily</changefreq>
            </url>
            
		</urlset>