<?xml version="1.0" encoding="UTF-8" ?>
		<urlset
			xmlns="https://www.sitemaps.org/schemas/sitemap/0.9"
			xmlns:xhtml="https://www.w3.org/1999/xhtml"
			xmlns:mobile="https://www.google.com/schemas/sitemap-mobile/1.0"
			xmlns:news="https://www.google.com/schemas/sitemap-news/0.9"
			xmlns:image="https://www.google.com/schemas/sitemap-image/1.1"
			xmlns:video="https://www.google.com/schemas/sitemap-video/1.1"
		>
      
    <url>
      <loc>https://hackernoon.com/liebn-on-spd-manifolds-the-additional-details-and-experiments-that-you-dont-want-to-miss</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/liebn-as-a-natural-generalization-of-euclidean-bn-and-backpropagation-of-matrix-functions</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/basic-layers-in-spdnet-tsmnet-and-statistical-results-of-scaling-in-the-liebn</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/proposing-a-novel-framework-called-liebn-a-conclusion</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/a-lie-group-approach-to-riemannian-batch-normalization-experimental-results</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/smarter-ai-code-completion-with-memory-efficient-techniques</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/reducing-memory-overhead-in-ai-models</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/understanding-multi-group-attention-in-ai-models</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/how-we-implemented-our-three-families-of-liebn-to-spd-neural-networks</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/liebn-on-spd-manifolds-showcasing-our-liebn-framework</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/liebn-on-the-lie-groups-of-spd-manifolds-and-deformed-lie-groups-of-spd-manifolds</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/training-ai-models-on-nvidia-a100-gpus</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/why-memory-io-efficiency-matters-for-ai-model-performance</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/faqs-how-bifurcated-attention-improves-ai-model-efficiency</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/riemannian-normalization-on-lie-groups</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/revisiting-existing-rbn-developing-riemannian-normalization-approaches-tailored-for-manifolds</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/revisiting-euclidean-normalization-a-second-look</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/faster-ai-code-generation-with-bifurcated-attention</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/understanding-latency-trade-offs-in-multi-query-vs-multi-head-ai-models</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/a-little-less-memory-a-lot-more-speed-the-science-of-bifurcated-attention</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/faster-ai-less-lag-a-smarter-way-to-process-language-models</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/why-multi-query-attention-matters-for-large-language-models</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/why-some-ai-models-are-faster-than-others</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/how-to-speed-up-your-ai-modelswithout-frying-your-memory</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/a-brief-review-of-the-lie-group-and-the-geometries-of-spd-manifolds</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
    <url>
      <loc>https://hackernoon.com/a-lie-group-approach-to-riemannian-batch-normalization</loc>
        <priority>0.8</priority>
      <changefreq>daily</changefreq>
    </url>
    
            <url>
            <loc>https://batching.tech/about</loc>
              <priority>0.8</priority>
            <changefreq>daily</changefreq>
            </url>
            
		</urlset>