{"id":530701,"date":"2026-03-16T11:05:32","date_gmt":"2026-03-16T11:05:32","guid":{"rendered":"https:\/\/webkul.com\/blog\/?p=530701"},"modified":"2026-03-16T11:05:41","modified_gmt":"2026-03-16T11:05:41","slug":"nemotron-3-super-mamba-moe-magic","status":"publish","type":"post","link":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/","title":{"rendered":"Nemotron 3 Super: Mamba + MoE Magic"},"content":{"rendered":"\n<p>NVIDIA recently announced Nemotron 3 Super. It is an open-source AI model that helps <a href=\"https:\/\/webkul.com\/blog\/impact-ai-agents-e-commerce\/\">AI agents<\/a> work better and faster.<\/p>\n\n\n\n<p>It was released on March 11, 2026.<\/p>\n\n\n\n<p> The model uses a new design that helps it reason better while using much less computing power, even for long tasks and multi-agent systems.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Why Agentic Reasoning Needs a New Approach<\/h2>\n\n\n\n<p>Today\u2019s AI agents do many tasks, like writing code, checking security issues, studying finance data, or planning actions. But they still face two big problems.<\/p>\n\n\n\n<p>Careful reasoning needs many steps. These steps use more computing power and make the process slower and more costly.<\/p>\n\n\n\n<p><strong>Context explosion:<\/strong> AI agents must track long histories, such as code, logs, or long chats.<\/p>\n\n\n\n<p>Transformer-based dense models are difficult to use as it is quadratic in attention cost and its memory requirements are high.<\/p>\n\n\n\n<p>Mixture of Experts (MoE) models can save resources, but they still slow down when working with very long inputs.<\/p>\n\n\n\n<p>These issues are directly addressed in Nemotron 3 Super. It is based on Nemotron 3 Nano (released late 2025) but is scaled to higher techniques.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Core Architecture: Hybrid Mamba-Transformer MoE<\/h2>\n\n\n\n<p>In its simplest form, Nemotron 3 Super is based on a <a href=\"https:\/\/research.nvidia.com\/publication\/2025-06_mambavision-hybrid-mamba-transformer-vision-backbone\">hybrid Mamba-Transformer<\/a> Mixture-of-Experts:<\/p>\n\n\n\n<p>They update their state very fast and handle longer inputs well. This helps when tasks have long context and memory use grows too much.<\/p>\n\n\n\n<p>Transformer attention is used only when needed. It helps the model understand connections and think more clearly.<\/p>\n\n\n\n<p>The mix is interspersed with layers of MoE to maintain active computation low.<\/p>\n\n\n\n<p>This mixed model is faster than regular Transformer models, especially when it creates long answers for AI agents.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Major Innovations in Nemotron 3 Super<\/h2>\n\n\n\n<p><strong>1. LatentMoE:<\/strong> A routing method that groups tokens into larger units and then sends them to the right experts.<\/p>\n\n\n\n<p>Compared to normal MoE routers, this model can use about four times more experts with almost the same cost. This helps it give better results.<\/p>\n\n\n\n<p><strong>2. Multi-Token Prediction (MTP):<\/strong> This lets the model predict many tokens at once. It helps the model respond faster and learn better during training.<\/p>\n\n\n\n<p><strong>3. NVFP4 Pre-training:<\/strong> This uses a 4-bit format from NVIDIA. It needs less memory and helps the model train faster.<\/p>\n\n\n\n<p><strong>4. 1M-token context window:<\/strong> The model can handle very long inputs. This helps AI agents remember more and follow long tasks step by step.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Scale and Efficiency Numbers<\/h2>\n\n\n\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" width=\"1200\" height=\"444\" src=\"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-1200x444.webp\" alt=\"Nemotron 3 Super Benchmarks\" class=\"wp-image-530890\" srcset=\"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-1200x444.webp 1200w, https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-300x111.webp 300w, https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-250x92.webp 250w, https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-768x284.webp 768w, https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia.webp 1531w\" sizes=\"(max-width: 1200px) 100vw, 1200px\" loading=\"lazy\" \/><\/figure>\n\n\n\n<p><em>Image Source :\u00a0<\/em><a href=\"https:\/\/developer.nvidia.com\/blog\/introducing-nemotron-3-super-an-open-hybrid-mamba-transformer-moe-for-agentic-reasoning\/\">nemotron-3-super@nvidia-blogs<\/a><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Total parameters: 120 billion (120.6B with some reports)<\/li>\n\n\n\n<li>Parameters per forward pass: Active: Approximately 12 billion (12.7B in certain tests).<\/li>\n\n\n\n<li>Context length: To 1 million tokens.<\/li>\n\n\n\n<li><strong>Throughput gains:<\/strong> It runs much faster than older models. It is about 5\u00d7 faster than past Nemotron Super models, 2.2\u00d7 faster than GPT-OSS-120B, and about 7.5\u00d7 faster than Qwen3.5-122B for long tasks.<\/li>\n<\/ul>\n\n\n\n<p>The model performs very well in reasoning, coding, math, and agent tasks. It also works much faster than many other models.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Data Highlights and Training of Nemotron 3 Super<\/h2>\n\n\n\n<p>Super pre-trains similarly to its smaller brother on approximately 25 trillion tokens in 2 stages:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Phase 1 (80 percent): Wide scope and coverage.<\/li>\n\n\n\n<li>Phase 2 (20 percent): Quality data revolving around benchmark accuracy and agency competence.<\/li>\n<\/ul>\n\n\n\n<p>Agent skills are improved during post-training using large reinforcement learning setups.<\/p>\n\n\n\n<p>On Hugging Face, NVIDIA also shared special training data, including synthetic code examples.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Availability and Openness of Nemotron 3 Super<\/h2>\n\n\n\n<p>As NVIDIA believes in open AI, Nemotron 3 Super is open:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Weights (base, post-trained, quantized versions such as FP8, BF16, NVFP4)<\/li>\n\n\n\n<li>Training datasets and recipes<\/li>\n\n\n\n<li>Availably on Hugging (e.g. nvidia\/NVIDIA-Nemotron-3-Super-120B-A12B-FP8)<\/li>\n<\/ul>\n\n\n\n<p>Also supported on NVIDIA inference platforms, Together, Perplexity, OpenRouter, and plenty of cloud providers (Google Vertex AI, Oracle, soon AWS Bedrock, Azure, etc.).<\/p>\n\n\n\n<p>Developers can fine-tune it, compress it, or run it on their own systems. This is helpful for private or secure business use.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Conclusion<\/h2>\n\n\n\n<p>Nemotron-3 Super shows a shift toward smarter and more efficient <a href=\"https:\/\/webkul.com\/artificial-intelligence\/\">Artifical Intelligence<\/a> models, instead of just making bigger ones. The next model, Nemotron-3 Ultra, will be even larger.<\/p>\n\n\n\n<p>It combines Mamba\u2019s speed, smart routing from LatentMoE, and a very large context. This helps AI agents work well on real hardware, even on one GPU or small clusters.<\/p>\n\n\n\n<p>You can try this model now if you build AI agents for coding, security checks, research help, or managing many agents.<\/p>\n\n\n\n<p>It gives strong results while using very little power. This means people can build powerful AI without huge GPU farms.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p><em>Want to Build AI-powered solutions visit\u00a0<a href=\"https:\/\/webkul.com\/generative-ai-services-and-solutions\/\">Webkul<\/a>!<\/em><\/p>\n<\/blockquote>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>NVIDIA recently announced Nemotron 3 Super. It is an open-source AI model that helps AI agents work better and faster. It was released on March 11, 2026. The model uses a new design that helps it reason better while using much less computing power, even for long tasks and multi-agent systems. Why Agentic Reasoning Needs <a href=\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/\">[&#8230;]<\/a><\/p>\n","protected":false},"author":724,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[13702],"tags":[13571,7240],"class_list":["post-530701","post","type-post","status-publish","format-standard","hentry","category-machine-learning","tag-artificial-intelligence","tag-machine-learning"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v24.5 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Nemotron 3 Super: Mamba + MoE Magic - Webkul Blog<\/title>\n<meta name=\"description\" content=\"Explore NVIDIA Nemotron 3 Super, an open-source AI model built for faster agentic reasoning and long-context tasks.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Nemotron 3 Super: Mamba + MoE Magic - Webkul Blog\" \/>\n<meta property=\"og:description\" content=\"Explore NVIDIA Nemotron 3 Super, an open-source AI model built for faster agentic reasoning and long-context tasks.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/\" \/>\n<meta property=\"og:site_name\" content=\"Webkul Blog\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/webkul\/\" \/>\n<meta property=\"article:published_time\" content=\"2026-03-16T11:05:32+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2026-03-16T11:05:41+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-1200x444.webp\" \/>\n<meta name=\"author\" content=\"Prashant Saini\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@webkul\" \/>\n<meta name=\"twitter:site\" content=\"@webkul\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Prashant Saini\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"4 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/\"},\"author\":{\"name\":\"Prashant Saini\",\"@id\":\"https:\/\/webkul.com\/blog\/#\/schema\/person\/53a57eff87fe1f3e9e69c165efdabdc4\"},\"headline\":\"Nemotron 3 Super: Mamba + MoE Magic\",\"datePublished\":\"2026-03-16T11:05:32+00:00\",\"dateModified\":\"2026-03-16T11:05:41+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/\"},\"wordCount\":722,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/webkul.com\/blog\/#organization\"},\"image\":{\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-1200x444.webp\",\"keywords\":[\"Artificial Intelligence\",\"machine learning\"],\"articleSection\":[\"machine learning\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/\",\"url\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/\",\"name\":\"Nemotron 3 Super: Mamba + MoE Magic - Webkul Blog\",\"isPartOf\":{\"@id\":\"https:\/\/webkul.com\/blog\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-1200x444.webp\",\"datePublished\":\"2026-03-16T11:05:32+00:00\",\"dateModified\":\"2026-03-16T11:05:41+00:00\",\"description\":\"Explore NVIDIA Nemotron 3 Super, an open-source AI model built for faster agentic reasoning and long-context tasks.\",\"breadcrumb\":{\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#primaryimage\",\"url\":\"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia.webp\",\"contentUrl\":\"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia.webp\",\"width\":1531,\"height\":566},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/webkul.com\/blog\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Nemotron 3 Super: Mamba + MoE Magic\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/webkul.com\/blog\/#website\",\"url\":\"https:\/\/webkul.com\/blog\/\",\"name\":\"Webkul Blog\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/webkul.com\/blog\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/webkul.com\/blog\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/webkul.com\/blog\/#organization\",\"name\":\"WebKul Software Private Limited\",\"url\":\"https:\/\/webkul.com\/blog\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/webkul.com\/blog\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2021\/08\/webkul-logo-accent-sq.png\",\"contentUrl\":\"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2021\/08\/webkul-logo-accent-sq.png\",\"width\":380,\"height\":380,\"caption\":\"WebKul Software Private Limited\"},\"image\":{\"@id\":\"https:\/\/webkul.com\/blog\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/webkul\/\",\"https:\/\/x.com\/webkul\",\"https:\/\/www.instagram.com\/webkul\/\",\"https:\/\/www.linkedin.com\/company\/webkul\",\"https:\/\/www.youtube.com\/user\/webkul\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/webkul.com\/blog\/#\/schema\/person\/53a57eff87fe1f3e9e69c165efdabdc4\",\"name\":\"Prashant Saini\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/webkul.com\/blog\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/90bd6382a7aa9ee0d5835bfaab3a739f91c37833f8e0d7cad51cd6a52b4914f0?s=96&d=https%3A%2F%2Fcdnblog.webkul.com%2Fblog%2Fwp-content%2Fuploads%2F2019%2F10%2Fmike.png&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/90bd6382a7aa9ee0d5835bfaab3a739f91c37833f8e0d7cad51cd6a52b4914f0?s=96&d=https%3A%2F%2Fcdnblog.webkul.com%2Fblog%2Fwp-content%2Fuploads%2F2019%2F10%2Fmike.png&r=g\",\"caption\":\"Prashant Saini\"},\"description\":\"Prashant, a passionate Machine Learning and AI enthusiast, specialized in building intelligent solutions using Python and Generative AI technologies.\",\"url\":\"https:\/\/webkul.com\/blog\/author\/prashant-ml322\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Nemotron 3 Super: Mamba + MoE Magic - Webkul Blog","description":"Explore NVIDIA Nemotron 3 Super, an open-source AI model built for faster agentic reasoning and long-context tasks.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/","og_locale":"en_US","og_type":"article","og_title":"Nemotron 3 Super: Mamba + MoE Magic - Webkul Blog","og_description":"Explore NVIDIA Nemotron 3 Super, an open-source AI model built for faster agentic reasoning and long-context tasks.","og_url":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/","og_site_name":"Webkul Blog","article_publisher":"https:\/\/www.facebook.com\/webkul\/","article_published_time":"2026-03-16T11:05:32+00:00","article_modified_time":"2026-03-16T11:05:41+00:00","og_image":[{"url":"https:\/\/webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-1200x444.webp","type":"","width":"","height":""}],"author":"Prashant Saini","twitter_card":"summary_large_image","twitter_creator":"@webkul","twitter_site":"@webkul","twitter_misc":{"Written by":"Prashant Saini","Est. reading time":"4 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#article","isPartOf":{"@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/"},"author":{"name":"Prashant Saini","@id":"https:\/\/webkul.com\/blog\/#\/schema\/person\/53a57eff87fe1f3e9e69c165efdabdc4"},"headline":"Nemotron 3 Super: Mamba + MoE Magic","datePublished":"2026-03-16T11:05:32+00:00","dateModified":"2026-03-16T11:05:41+00:00","mainEntityOfPage":{"@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/"},"wordCount":722,"commentCount":0,"publisher":{"@id":"https:\/\/webkul.com\/blog\/#organization"},"image":{"@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#primaryimage"},"thumbnailUrl":"https:\/\/webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-1200x444.webp","keywords":["Artificial Intelligence","machine learning"],"articleSection":["machine learning"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/","url":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/","name":"Nemotron 3 Super: Mamba + MoE Magic - Webkul Blog","isPartOf":{"@id":"https:\/\/webkul.com\/blog\/#website"},"primaryImageOfPage":{"@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#primaryimage"},"image":{"@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#primaryimage"},"thumbnailUrl":"https:\/\/webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia-1200x444.webp","datePublished":"2026-03-16T11:05:32+00:00","dateModified":"2026-03-16T11:05:41+00:00","description":"Explore NVIDIA Nemotron 3 Super, an open-source AI model built for faster agentic reasoning and long-context tasks.","breadcrumb":{"@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#primaryimage","url":"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia.webp","contentUrl":"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2026\/03\/nvidia.webp","width":1531,"height":566},{"@type":"BreadcrumbList","@id":"https:\/\/webkul.com\/blog\/nemotron-3-super-mamba-moe-magic\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/webkul.com\/blog\/"},{"@type":"ListItem","position":2,"name":"Nemotron 3 Super: Mamba + MoE Magic"}]},{"@type":"WebSite","@id":"https:\/\/webkul.com\/blog\/#website","url":"https:\/\/webkul.com\/blog\/","name":"Webkul Blog","description":"","publisher":{"@id":"https:\/\/webkul.com\/blog\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/webkul.com\/blog\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/webkul.com\/blog\/#organization","name":"WebKul Software Private Limited","url":"https:\/\/webkul.com\/blog\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/webkul.com\/blog\/#\/schema\/logo\/image\/","url":"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2021\/08\/webkul-logo-accent-sq.png","contentUrl":"https:\/\/cdnblog.webkul.com\/blog\/wp-content\/uploads\/2021\/08\/webkul-logo-accent-sq.png","width":380,"height":380,"caption":"WebKul Software Private Limited"},"image":{"@id":"https:\/\/webkul.com\/blog\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/webkul\/","https:\/\/x.com\/webkul","https:\/\/www.instagram.com\/webkul\/","https:\/\/www.linkedin.com\/company\/webkul","https:\/\/www.youtube.com\/user\/webkul\/"]},{"@type":"Person","@id":"https:\/\/webkul.com\/blog\/#\/schema\/person\/53a57eff87fe1f3e9e69c165efdabdc4","name":"Prashant Saini","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/webkul.com\/blog\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/90bd6382a7aa9ee0d5835bfaab3a739f91c37833f8e0d7cad51cd6a52b4914f0?s=96&d=https%3A%2F%2Fcdnblog.webkul.com%2Fblog%2Fwp-content%2Fuploads%2F2019%2F10%2Fmike.png&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/90bd6382a7aa9ee0d5835bfaab3a739f91c37833f8e0d7cad51cd6a52b4914f0?s=96&d=https%3A%2F%2Fcdnblog.webkul.com%2Fblog%2Fwp-content%2Fuploads%2F2019%2F10%2Fmike.png&r=g","caption":"Prashant Saini"},"description":"Prashant, a passionate Machine Learning and AI enthusiast, specialized in building intelligent solutions using Python and Generative AI technologies.","url":"https:\/\/webkul.com\/blog\/author\/prashant-ml322\/"}]}},"amp_enabled":true,"_links":{"self":[{"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/posts\/530701","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/users\/724"}],"replies":[{"embeddable":true,"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/comments?post=530701"}],"version-history":[{"count":2,"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/posts\/530701\/revisions"}],"predecessor-version":[{"id":530896,"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/posts\/530701\/revisions\/530896"}],"wp:attachment":[{"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/media?parent=530701"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/categories?post=530701"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/webkul.com\/blog\/wp-json\/wp\/v2\/tags?post=530701"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}