<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem Core: Super Jarvis</title>
    <description>The latest articles on Forem Core by Super Jarvis (@super_jarvis_76aa3fc6035d).</description>
    <link>https://core.forem.com/super_jarvis_76aa3fc6035d</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3890917%2F7334052a-5af2-49af-b96a-5f2e69309689.png</url>
      <title>Forem Core: Super Jarvis</title>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://core.forem.com/feed/super_jarvis_76aa3fc6035d"/>
    <language>en</language>
    <item>
      <title>DeepSeek V4 vs Other Models: When Pro or Flash Makes Sense</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:29:55 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-vs-other-models-when-pro-or-flash-makes-sense-5hc</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-vs-other-models-when-pro-or-flash-makes-sense-5hc</guid>
      <description>&lt;p&gt;DeepSeek V4 is best evaluated as a two-model family rather than one model.&lt;/p&gt;

&lt;p&gt;DeepSeek V4 Pro is the flagship path. DeepSeek V4 Flash is the efficient path. Both list 1M context in the current DeepSeek API pricing table.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faj6gn1f0100zhcdljl7m.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faj6gn1f0100zhcdljl7m.jpg" alt="DeepSeek V4 routing comparison dashboard" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;A comparison is only useful when it turns into a routing rule: default to the cheaper reliable path, then escalate when quality risk increases.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  V4 Pro vs V4 Flash
&lt;/h2&gt;

&lt;p&gt;Choose Pro when:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The task needs the best available DeepSeek V4 benchmark ceiling.&lt;/li&gt;
&lt;li&gt;The prompt involves code repair, planning, math, or multi-step tools.&lt;/li&gt;
&lt;li&gt;A wrong answer is more expensive than a slower or pricier answer.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Choose Flash when:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The task is high-volume.&lt;/li&gt;
&lt;li&gt;The output can be checked, retried, or escalated.&lt;/li&gt;
&lt;li&gt;You need 1M context but want lower input and output token costs.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Comparing to other model families
&lt;/h2&gt;

&lt;p&gt;Against other frontier models, DeepSeek V4 Pro should be tested on your hardest real workflows: coding, long-context reasoning, and agentic tasks.&lt;/p&gt;

&lt;p&gt;Against efficient models, DeepSeek V4 Flash is the more natural comparison because it keeps 1M context while using lower per-token prices.&lt;/p&gt;

&lt;h2&gt;
  
  
  Best routing pattern
&lt;/h2&gt;

&lt;p&gt;A practical routing setup is:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Start with Flash for cheap comprehension and summaries.&lt;/li&gt;
&lt;li&gt;Escalate to Pro when the task is complex or user-visible.&lt;/li&gt;
&lt;li&gt;Add web search only when freshness matters.&lt;/li&gt;
&lt;li&gt;Add Thinking only when the task benefits from deeper reasoning.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This keeps cost predictable while preserving quality for hard prompts.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-vs-other-models?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Technical Report: Architecture, Training, and Benchmarks</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:29:08 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-technical-report-architecture-training-and-benchmarks-k03</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-technical-report-architecture-training-and-benchmarks-k03</guid>
      <description>&lt;p&gt;The DeepSeek V4 technical report describes a preview V4 family with two Mixture-of-Experts language models:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;DeepSeek V4 Pro&lt;/strong&gt;: 1.6T total parameters, 49B activated parameters, 1M context.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;DeepSeek V4 Flash&lt;/strong&gt;: 284B total parameters, 13B activated parameters, 1M context.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Primary sources:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro" rel="noopener noreferrer"&gt;DeepSeek-V4-Pro model card&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main/DeepSeek_V4.pdf" rel="noopener noreferrer"&gt;DeepSeek_V4.pdf&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://api-docs.deepseek.com/quick_start/pricing/" rel="noopener noreferrer"&gt;DeepSeek API pricing&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What the technical report focuses on
&lt;/h2&gt;

&lt;p&gt;The report frames DeepSeek V4 around efficient long-context intelligence. The headline product implication is simple: both V4 Pro and V4 Flash expose a 1M-token context window, but they target different cost and capability envelopes.&lt;/p&gt;

&lt;p&gt;Pro is the higher-capacity model for hard reasoning, coding, and agentic workflows. Flash is the lower-cost model for high-volume chat, summarization, routing, and everyday product paths.&lt;/p&gt;

&lt;h2&gt;
  
  
  Architecture notes
&lt;/h2&gt;

&lt;p&gt;The report highlights several architecture and optimization upgrades:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Hybrid attention for long-context efficiency.&lt;/li&gt;
&lt;li&gt;Manifold-Constrained Hyper-Connections for stronger signal propagation.&lt;/li&gt;
&lt;li&gt;Muon optimizer for training stability and convergence.&lt;/li&gt;
&lt;li&gt;MoE scaling with separate Pro and Flash model sizes.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F60nybr3t9rb84icrqxq5.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F60nybr3t9rb84icrqxq5.jpg" alt="DeepSeek V4 report layers and evidence map" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Use the architecture section to decide what to measure, not as a substitute for measuring your own prompts.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;For builders, the practical question is not just which model has the larger parameter count. The question is where longer context, cache behavior, and reasoning effort change the cost-quality curve.&lt;/p&gt;

&lt;h2&gt;
  
  
  Training and post-training
&lt;/h2&gt;

&lt;p&gt;DeepSeek says the V4 models are pre-trained on more than 32T tokens and then post-trained with a multi-stage process. The release materials describe domain-specific expert cultivation followed by model consolidation.&lt;/p&gt;

&lt;p&gt;That matters for product evaluation because one benchmark score is not enough. You should test domain tasks directly: code repair, long document synthesis, tool-use workflows, structured extraction, and high-volume support chat.&lt;/p&gt;

&lt;h2&gt;
  
  
  Reasoning modes
&lt;/h2&gt;

&lt;p&gt;The technical report and model card describe non-thinking, thinking, and max-thinking styles. In practice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Use non-thinking mode for low-risk, fast, low-cost responses.&lt;/li&gt;
&lt;li&gt;Use thinking mode for math, coding, planning, and multi-step reasoning.&lt;/li&gt;
&lt;li&gt;Use max-style reasoning only when the added latency and cost are justified.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The current DeepSeek API pricing page lists &lt;code&gt;deepseek-v4-flash&lt;/code&gt; and &lt;code&gt;deepseek-v4-pro&lt;/code&gt; as the V4 model IDs.&lt;/p&gt;

&lt;h2&gt;
  
  
  Benchmark signals
&lt;/h2&gt;

&lt;p&gt;The release materials include benchmark snapshots across knowledge, coding, long-context, and agentic tasks. The site tracks a few practical anchor scores:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;MMLU-Pro&lt;/th&gt;
&lt;th&gt;LiveCodeBench&lt;/th&gt;
&lt;th&gt;SWE Verified&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Flash Max&lt;/td&gt;
&lt;td&gt;86.2&lt;/td&gt;
&lt;td&gt;91.6&lt;/td&gt;
&lt;td&gt;79.0&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Pro Max&lt;/td&gt;
&lt;td&gt;87.5&lt;/td&gt;
&lt;td&gt;93.5&lt;/td&gt;
&lt;td&gt;80.6&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Treat these as routing hints, not final product truth. If your application depends on code changes, retrieval quality, or tool calls, build an eval set from your own traffic and compare Flash against Pro with the same prompts.&lt;/p&gt;

&lt;h2&gt;
  
  
  Implementation checklist
&lt;/h2&gt;

&lt;p&gt;Before adopting DeepSeek V4 in production, verify:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Which workflows need Pro instead of Flash.&lt;/li&gt;
&lt;li&gt;Whether Thinking improves your specific task enough to justify the cost.&lt;/li&gt;
&lt;li&gt;How much prompt caching reduces repeated-context cost.&lt;/li&gt;
&lt;li&gt;Whether your longest real documents fit cleanly inside the 1M context window.&lt;/li&gt;
&lt;li&gt;Whether tool-use and JSON outputs are stable enough for your product contracts.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The technical report explains the direction. Your own evals should decide routing, retry behavior, and credit pricing.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-technical-report?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Size: Parameters, Active Parameters, and Context</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:28:21 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-size-parameters-active-parameters-and-context-4n9b</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-size-parameters-active-parameters-and-context-4n9b</guid>
      <description>&lt;p&gt;DeepSeek V4 size is easiest to understand by separating total parameters, active parameters, and context length.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F40r1mvbvy1kjixe9be51.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F40r1mvbvy1kjixe9be51.jpg" alt="DeepSeek V4 model size and context illustration" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The useful distinction is total capacity versus active inference cost: MoE scale lets a model be large without activating every parameter for every token.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Official model sizes
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;Total parameters&lt;/th&gt;
&lt;th&gt;Active parameters&lt;/th&gt;
&lt;th&gt;Context&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Flash&lt;/td&gt;
&lt;td&gt;284B&lt;/td&gt;
&lt;td&gt;13B&lt;/td&gt;
&lt;td&gt;1M tokens&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Pro&lt;/td&gt;
&lt;td&gt;1.6T&lt;/td&gt;
&lt;td&gt;49B&lt;/td&gt;
&lt;td&gt;1M tokens&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Sources: &lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro" rel="noopener noreferrer"&gt;DeepSeek-V4-Pro model card&lt;/a&gt; and &lt;a href="https://api-docs.deepseek.com/quick_start/pricing/" rel="noopener noreferrer"&gt;DeepSeek API pricing&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What active parameters mean
&lt;/h2&gt;

&lt;p&gt;DeepSeek V4 is an MoE family, so total parameters and active parameters are different. Total parameters describe the full model capacity. Active parameters describe the approximate amount used per token during inference.&lt;/p&gt;

&lt;p&gt;This is why Flash can be much cheaper while still remaining useful: it has fewer active parameters and lower token prices.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why 1M context matters
&lt;/h2&gt;

&lt;p&gt;A 1M context window changes product design. Instead of sending only the last few messages, you can include large documents, long project histories, logs, or source files. The tradeoff is cost and latency, so context should still be curated rather than dumped blindly.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-size?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Price: Pro vs Flash API Costs</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:27:35 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-price-pro-vs-flash-api-costs-m34</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-price-pro-vs-flash-api-costs-m34</guid>
      <description>&lt;p&gt;DeepSeek V4 pricing is split across two API models: &lt;code&gt;deepseek-v4-pro&lt;/code&gt; and &lt;code&gt;deepseek-v4-flash&lt;/code&gt;.&lt;/p&gt;

&lt;p&gt;The official pricing page lists separate rates for cache-hit input, cache-miss input, and output tokens. That matters because repeated system prompts, reused context, and stable templates can make cache-hit pricing materially cheaper.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F82m98lkz6utm5z7icm4f.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F82m98lkz6utm5z7icm4f.jpg" alt="DeepSeek V4 Pro and Flash pricing lanes" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Think of Flash and Pro as two pricing lanes: Flash handles volume, while Pro is reserved for prompts where failure cost is higher.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Official API prices
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;Cache-hit input&lt;/th&gt;
&lt;th&gt;Cache-miss input&lt;/th&gt;
&lt;th&gt;Output&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Flash&lt;/td&gt;
&lt;td&gt;$0.028 / 1M tokens&lt;/td&gt;
&lt;td&gt;$0.14 / 1M tokens&lt;/td&gt;
&lt;td&gt;$0.28 / 1M tokens&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Pro&lt;/td&gt;
&lt;td&gt;$0.145 / 1M tokens&lt;/td&gt;
&lt;td&gt;$1.74 / 1M tokens&lt;/td&gt;
&lt;td&gt;$3.48 / 1M tokens&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Source: &lt;a href="https://api-docs.deepseek.com/quick_start/pricing/" rel="noopener noreferrer"&gt;DeepSeek API pricing&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to choose
&lt;/h2&gt;

&lt;p&gt;Use DeepSeek V4 Flash when the workload is high-volume: chat, summaries, extraction, classification, routing, and first-pass analysis.&lt;/p&gt;

&lt;p&gt;Use DeepSeek V4 Pro when the task has a higher failure cost: difficult code repair, long reasoning, advanced math, agent planning, or final answer synthesis after cheaper models have prepared context.&lt;/p&gt;

&lt;h2&gt;
  
  
  Credit mapping on this site
&lt;/h2&gt;

&lt;p&gt;This site uses a simple credit layer above the official API:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Flash chat: 1 credit&lt;/li&gt;
&lt;li&gt;Pro chat: 4 credits&lt;/li&gt;
&lt;li&gt;Thinking: +1 credit&lt;/li&gt;
&lt;li&gt;Web search: +2 credits&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is not DeepSeek's official billing model. It is a product-level abstraction so users can compare Flash, Pro, Thinking, and web search in one interface.&lt;/p&gt;

&lt;h2&gt;
  
  
  Practical cost advice
&lt;/h2&gt;

&lt;p&gt;Keep reusable instructions stable so prompt caching can work. Route cheap, repetitive prompts to Flash. Escalate to Pro only when the answer needs the stronger reasoning ceiling.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-price?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Paper: What Builders Should Notice</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:26:48 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-paper-what-builders-should-notice-3jac</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-paper-what-builders-should-notice-3jac</guid>
      <description>&lt;p&gt;The DeepSeek V4 paper and model card describe the V4 family as MoE language models trained with MLA and DeepSeekSparse attention.&lt;/p&gt;

&lt;p&gt;Primary sources:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro" rel="noopener noreferrer"&gt;DeepSeek-V4-Pro model card&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main/DeepSeek_V4.pdf" rel="noopener noreferrer"&gt;DeepSeek_V4.pdf&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqm9bz98fq257xajf3enf.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqm9bz98fq257xajf3enf.jpg" alt="DeepSeek V4 paper reading workspace" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Read the paper as a product-routing document: architecture details matter most when they change latency, cost, context, or reliability.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Builder takeaways
&lt;/h2&gt;

&lt;p&gt;The release has two important product implications.&lt;/p&gt;

&lt;p&gt;First, the model family splits capacity. Pro is much larger and targets stronger reasoning. Flash is smaller and cheaper while still exposing a 1M context window.&lt;/p&gt;

&lt;p&gt;Second, the API pricing encourages cache-aware prompt design. Reused input can be cheaper than fresh cache-miss input, so teams should stabilize system prompts and repeated context templates.&lt;/p&gt;

&lt;h2&gt;
  
  
  What to test after reading
&lt;/h2&gt;

&lt;p&gt;After reading the paper, build a task set that reflects your product:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;long context retrieval and synthesis&lt;/li&gt;
&lt;li&gt;code repair and code review&lt;/li&gt;
&lt;li&gt;multi-step planning&lt;/li&gt;
&lt;li&gt;factual answers with web search&lt;/li&gt;
&lt;li&gt;structured JSON outputs&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then compare Flash and Pro with the same prompts. The paper explains architecture direction, but your eval decides routing.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-paper?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 API: Model IDs, Base URL, Thinking, and Tools</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:26:02 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-api-model-ids-base-url-thinking-and-tools-3m5k</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-api-model-ids-base-url-thinking-and-tools-3m5k</guid>
      <description>&lt;p&gt;DeepSeek V4 is exposed through the DeepSeek OpenAI-compatible API. The current pricing page lists two V4 model IDs:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;code&gt;deepseek-v4-pro&lt;/code&gt;&lt;/li&gt;
&lt;li&gt;&lt;code&gt;deepseek-v4-flash&lt;/code&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The base URL is:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;https://api.deepseek.com
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Source: &lt;a href="https://api-docs.deepseek.com/quick_start/pricing/" rel="noopener noreferrer"&gt;DeepSeek API pricing&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fibwaw9mwfta9de11l3t9.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fibwaw9mwfta9de11l3t9.jpg" alt="DeepSeek V4 API request pipeline" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;API integration is mostly about choosing the right model ID, keeping the request shape compatible, and deciding when tools or Thinking should be enabled.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Minimal request shape
&lt;/h2&gt;

&lt;p&gt;Use the chat completions API with one of the V4 model IDs:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight json"&gt;&lt;code&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"model"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"deepseek-v4-flash"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"messages"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="w"&gt;
    &lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
      &lt;/span&gt;&lt;span class="nl"&gt;"role"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"user"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
      &lt;/span&gt;&lt;span class="nl"&gt;"content"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"Explain DeepSeek V4 Flash pricing."&lt;/span&gt;&lt;span class="w"&gt;
    &lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Thinking mode
&lt;/h2&gt;

&lt;p&gt;DeepSeek documents Thinking as a request option with enabled or disabled mode, plus reasoning effort. Use Thinking when you want the model to spend more reasoning budget on difficult tasks.&lt;/p&gt;

&lt;p&gt;In product terms:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Disable Thinking for fast answers and low-cost paths.&lt;/li&gt;
&lt;li&gt;Enable Thinking for code repair, planning, math, and long analysis.&lt;/li&gt;
&lt;li&gt;Use Pro when the answer quality ceiling matters more than cost.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Tools and web search
&lt;/h2&gt;

&lt;p&gt;DeepSeek V4 can be used behind a tool-enabled chat route. On this site, web search is implemented as a server-side &lt;code&gt;search_web&lt;/code&gt; tool and then passed into the model response. That means web search depends on the site's search provider configuration, not only DeepSeek itself.&lt;/p&gt;

&lt;h2&gt;
  
  
  Image upload
&lt;/h2&gt;

&lt;p&gt;The site supports image attachment upload and passes public image references into chat. The current V4 API documentation primarily describes text, Thinking, tools, JSON, and FIM surfaces, so direct image understanding should be verified in your runtime before promising vision behavior.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-api?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Benchmark: Pro and Flash Scores</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:25:04 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-benchmark-pro-and-flash-scores-61f</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-benchmark-pro-and-flash-scores-61f</guid>
      <description>&lt;p&gt;The DeepSeek V4 release materials include benchmark rows for DeepSeek V4 Flash and DeepSeek V4 Pro in Max mode.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F50vvm3yd07bcvk5qgckn.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F50vvm3yd07bcvk5qgckn.jpg" alt="DeepSeek V4 benchmark dashboard" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Benchmarks are useful as a first routing signal, but production defaults should still be decided with prompts from your own workload.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Official snapshot
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;MMLU-Pro&lt;/th&gt;
&lt;th&gt;LiveCodeBench&lt;/th&gt;
&lt;th&gt;SWE Verified&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Flash&lt;/td&gt;
&lt;td&gt;86.2&lt;/td&gt;
&lt;td&gt;91.6&lt;/td&gt;
&lt;td&gt;79.0&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Pro&lt;/td&gt;
&lt;td&gt;87.5&lt;/td&gt;
&lt;td&gt;93.5&lt;/td&gt;
&lt;td&gt;80.6&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Sources: &lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro" rel="noopener noreferrer"&gt;DeepSeek-V4-Pro model card&lt;/a&gt; and &lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main/DeepSeek_V4.pdf" rel="noopener noreferrer"&gt;DeepSeek_V4.pdf&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What the numbers suggest
&lt;/h2&gt;

&lt;p&gt;Pro leads the snapshot, especially where reasoning and coding ceilings matter. Flash is close enough that it can be the default for many high-volume workflows, especially when the task can tolerate a second pass or escalation.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to evaluate in production
&lt;/h2&gt;

&lt;p&gt;Do not ship on public benchmarks alone. Build a small internal eval set with your real prompts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;20 frequent user requests&lt;/li&gt;
&lt;li&gt;20 difficult edge cases&lt;/li&gt;
&lt;li&gt;20 code or reasoning tasks&lt;/li&gt;
&lt;li&gt;10 long-context tasks&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Run Flash first, Pro second, then compare correctness, latency, and cost. The best default is usually workload-specific.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-benchmark?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 vs Other Models: When Pro or Flash Makes Sense</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:19:15 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-vs-other-models-when-pro-or-flash-makes-sense-24of</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-vs-other-models-when-pro-or-flash-makes-sense-24of</guid>
      <description>&lt;p&gt;DeepSeek V4 is best evaluated as a two-model family rather than one model.&lt;/p&gt;

&lt;p&gt;DeepSeek V4 Pro is the flagship path. DeepSeek V4 Flash is the efficient path. Both list 1M context in the current DeepSeek API pricing table.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faj6gn1f0100zhcdljl7m.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faj6gn1f0100zhcdljl7m.jpg" alt="DeepSeek V4 routing comparison dashboard" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;A comparison is only useful when it turns into a routing rule: default to the cheaper reliable path, then escalate when quality risk increases.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  V4 Pro vs V4 Flash
&lt;/h2&gt;

&lt;p&gt;Choose Pro when:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The task needs the best available DeepSeek V4 benchmark ceiling.&lt;/li&gt;
&lt;li&gt;The prompt involves code repair, planning, math, or multi-step tools.&lt;/li&gt;
&lt;li&gt;A wrong answer is more expensive than a slower or pricier answer.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Choose Flash when:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The task is high-volume.&lt;/li&gt;
&lt;li&gt;The output can be checked, retried, or escalated.&lt;/li&gt;
&lt;li&gt;You need 1M context but want lower input and output token costs.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Comparing to other model families
&lt;/h2&gt;

&lt;p&gt;Against other frontier models, DeepSeek V4 Pro should be tested on your hardest real workflows: coding, long-context reasoning, and agentic tasks.&lt;/p&gt;

&lt;p&gt;Against efficient models, DeepSeek V4 Flash is the more natural comparison because it keeps 1M context while using lower per-token prices.&lt;/p&gt;

&lt;h2&gt;
  
  
  Best routing pattern
&lt;/h2&gt;

&lt;p&gt;A practical routing setup is:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Start with Flash for cheap comprehension and summaries.&lt;/li&gt;
&lt;li&gt;Escalate to Pro when the task is complex or user-visible.&lt;/li&gt;
&lt;li&gt;Add web search only when freshness matters.&lt;/li&gt;
&lt;li&gt;Add Thinking only when the task benefits from deeper reasoning.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This keeps cost predictable while preserving quality for hard prompts.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-vs-other-models?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Size: Parameters, Active Parameters, and Context</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:18:28 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-size-parameters-active-parameters-and-context-hl3</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-size-parameters-active-parameters-and-context-hl3</guid>
      <description>&lt;p&gt;DeepSeek V4 size is easiest to understand by separating total parameters, active parameters, and context length.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F40r1mvbvy1kjixe9be51.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F40r1mvbvy1kjixe9be51.jpg" alt="DeepSeek V4 model size and context illustration" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The useful distinction is total capacity versus active inference cost: MoE scale lets a model be large without activating every parameter for every token.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Official model sizes
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;Total parameters&lt;/th&gt;
&lt;th&gt;Active parameters&lt;/th&gt;
&lt;th&gt;Context&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Flash&lt;/td&gt;
&lt;td&gt;284B&lt;/td&gt;
&lt;td&gt;13B&lt;/td&gt;
&lt;td&gt;1M tokens&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Pro&lt;/td&gt;
&lt;td&gt;1.6T&lt;/td&gt;
&lt;td&gt;49B&lt;/td&gt;
&lt;td&gt;1M tokens&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Sources: &lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro" rel="noopener noreferrer"&gt;DeepSeek-V4-Pro model card&lt;/a&gt; and &lt;a href="https://api-docs.deepseek.com/quick_start/pricing/" rel="noopener noreferrer"&gt;DeepSeek API pricing&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What active parameters mean
&lt;/h2&gt;

&lt;p&gt;DeepSeek V4 is an MoE family, so total parameters and active parameters are different. Total parameters describe the full model capacity. Active parameters describe the approximate amount used per token during inference.&lt;/p&gt;

&lt;p&gt;This is why Flash can be much cheaper while still remaining useful: it has fewer active parameters and lower token prices.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why 1M context matters
&lt;/h2&gt;

&lt;p&gt;A 1M context window changes product design. Instead of sending only the last few messages, you can include large documents, long project histories, logs, or source files. The tradeoff is cost and latency, so context should still be curated rather than dumped blindly.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-size?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Price: Pro vs Flash API Costs</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:17:42 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-price-pro-vs-flash-api-costs-2jkn</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-price-pro-vs-flash-api-costs-2jkn</guid>
      <description>&lt;p&gt;DeepSeek V4 pricing is split across two API models: &lt;code&gt;deepseek-v4-pro&lt;/code&gt; and &lt;code&gt;deepseek-v4-flash&lt;/code&gt;.&lt;/p&gt;

&lt;p&gt;The official pricing page lists separate rates for cache-hit input, cache-miss input, and output tokens. That matters because repeated system prompts, reused context, and stable templates can make cache-hit pricing materially cheaper.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F82m98lkz6utm5z7icm4f.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F82m98lkz6utm5z7icm4f.jpg" alt="DeepSeek V4 Pro and Flash pricing lanes" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Think of Flash and Pro as two pricing lanes: Flash handles volume, while Pro is reserved for prompts where failure cost is higher.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Official API prices
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;Cache-hit input&lt;/th&gt;
&lt;th&gt;Cache-miss input&lt;/th&gt;
&lt;th&gt;Output&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Flash&lt;/td&gt;
&lt;td&gt;$0.028 / 1M tokens&lt;/td&gt;
&lt;td&gt;$0.14 / 1M tokens&lt;/td&gt;
&lt;td&gt;$0.28 / 1M tokens&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Pro&lt;/td&gt;
&lt;td&gt;$0.145 / 1M tokens&lt;/td&gt;
&lt;td&gt;$1.74 / 1M tokens&lt;/td&gt;
&lt;td&gt;$3.48 / 1M tokens&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Source: &lt;a href="https://api-docs.deepseek.com/quick_start/pricing/" rel="noopener noreferrer"&gt;DeepSeek API pricing&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to choose
&lt;/h2&gt;

&lt;p&gt;Use DeepSeek V4 Flash when the workload is high-volume: chat, summaries, extraction, classification, routing, and first-pass analysis.&lt;/p&gt;

&lt;p&gt;Use DeepSeek V4 Pro when the task has a higher failure cost: difficult code repair, long reasoning, advanced math, agent planning, or final answer synthesis after cheaper models have prepared context.&lt;/p&gt;

&lt;h2&gt;
  
  
  Credit mapping on this site
&lt;/h2&gt;

&lt;p&gt;This site uses a simple credit layer above the official API:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Flash chat: 1 credit&lt;/li&gt;
&lt;li&gt;Pro chat: 4 credits&lt;/li&gt;
&lt;li&gt;Thinking: +1 credit&lt;/li&gt;
&lt;li&gt;Web search: +2 credits&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is not DeepSeek's official billing model. It is a product-level abstraction so users can compare Flash, Pro, Thinking, and web search in one interface.&lt;/p&gt;

&lt;h2&gt;
  
  
  Practical cost advice
&lt;/h2&gt;

&lt;p&gt;Keep reusable instructions stable so prompt caching can work. Route cheap, repetitive prompts to Flash. Escalate to Pro only when the answer needs the stronger reasoning ceiling.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-price?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Technical Report: Architecture, Training, and Benchmarks</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:14:45 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-technical-report-architecture-training-and-benchmarks-4h1b</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-technical-report-architecture-training-and-benchmarks-4h1b</guid>
      <description>&lt;p&gt;The DeepSeek V4 technical report describes a preview V4 family with two Mixture-of-Experts language models:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;DeepSeek V4 Pro&lt;/strong&gt;: 1.6T total parameters, 49B activated parameters, 1M context.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;DeepSeek V4 Flash&lt;/strong&gt;: 284B total parameters, 13B activated parameters, 1M context.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Primary sources:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro" rel="noopener noreferrer"&gt;DeepSeek-V4-Pro model card&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main/DeepSeek_V4.pdf" rel="noopener noreferrer"&gt;DeepSeek_V4.pdf&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://api-docs.deepseek.com/quick_start/pricing/" rel="noopener noreferrer"&gt;DeepSeek API pricing&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What the technical report focuses on
&lt;/h2&gt;

&lt;p&gt;The report frames DeepSeek V4 around efficient long-context intelligence. The headline product implication is simple: both V4 Pro and V4 Flash expose a 1M-token context window, but they target different cost and capability envelopes.&lt;/p&gt;

&lt;p&gt;Pro is the higher-capacity model for hard reasoning, coding, and agentic workflows. Flash is the lower-cost model for high-volume chat, summarization, routing, and everyday product paths.&lt;/p&gt;

&lt;h2&gt;
  
  
  Architecture notes
&lt;/h2&gt;

&lt;p&gt;The report highlights several architecture and optimization upgrades:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Hybrid attention for long-context efficiency.&lt;/li&gt;
&lt;li&gt;Manifold-Constrained Hyper-Connections for stronger signal propagation.&lt;/li&gt;
&lt;li&gt;Muon optimizer for training stability and convergence.&lt;/li&gt;
&lt;li&gt;MoE scaling with separate Pro and Flash model sizes.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F60nybr3t9rb84icrqxq5.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F60nybr3t9rb84icrqxq5.jpg" alt="DeepSeek V4 report layers and evidence map" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Use the architecture section to decide what to measure, not as a substitute for measuring your own prompts.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;For builders, the practical question is not just which model has the larger parameter count. The question is where longer context, cache behavior, and reasoning effort change the cost-quality curve.&lt;/p&gt;

&lt;h2&gt;
  
  
  Training and post-training
&lt;/h2&gt;

&lt;p&gt;DeepSeek says the V4 models are pre-trained on more than 32T tokens and then post-trained with a multi-stage process. The release materials describe domain-specific expert cultivation followed by model consolidation.&lt;/p&gt;

&lt;p&gt;That matters for product evaluation because one benchmark score is not enough. You should test domain tasks directly: code repair, long document synthesis, tool-use workflows, structured extraction, and high-volume support chat.&lt;/p&gt;

&lt;h2&gt;
  
  
  Reasoning modes
&lt;/h2&gt;

&lt;p&gt;The technical report and model card describe non-thinking, thinking, and max-thinking styles. In practice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Use non-thinking mode for low-risk, fast, low-cost responses.&lt;/li&gt;
&lt;li&gt;Use thinking mode for math, coding, planning, and multi-step reasoning.&lt;/li&gt;
&lt;li&gt;Use max-style reasoning only when the added latency and cost are justified.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The current DeepSeek API pricing page lists &lt;code&gt;deepseek-v4-flash&lt;/code&gt; and &lt;code&gt;deepseek-v4-pro&lt;/code&gt; as the V4 model IDs.&lt;/p&gt;

&lt;h2&gt;
  
  
  Benchmark signals
&lt;/h2&gt;

&lt;p&gt;The release materials include benchmark snapshots across knowledge, coding, long-context, and agentic tasks. The site tracks a few practical anchor scores:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;MMLU-Pro&lt;/th&gt;
&lt;th&gt;LiveCodeBench&lt;/th&gt;
&lt;th&gt;SWE Verified&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Flash Max&lt;/td&gt;
&lt;td&gt;86.2&lt;/td&gt;
&lt;td&gt;91.6&lt;/td&gt;
&lt;td&gt;79.0&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DeepSeek V4 Pro Max&lt;/td&gt;
&lt;td&gt;87.5&lt;/td&gt;
&lt;td&gt;93.5&lt;/td&gt;
&lt;td&gt;80.6&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Treat these as routing hints, not final product truth. If your application depends on code changes, retrieval quality, or tool calls, build an eval set from your own traffic and compare Flash against Pro with the same prompts.&lt;/p&gt;

&lt;h2&gt;
  
  
  Implementation checklist
&lt;/h2&gt;

&lt;p&gt;Before adopting DeepSeek V4 in production, verify:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Which workflows need Pro instead of Flash.&lt;/li&gt;
&lt;li&gt;Whether Thinking improves your specific task enough to justify the cost.&lt;/li&gt;
&lt;li&gt;How much prompt caching reduces repeated-context cost.&lt;/li&gt;
&lt;li&gt;Whether your longest real documents fit cleanly inside the 1M context window.&lt;/li&gt;
&lt;li&gt;Whether tool-use and JSON outputs are stable enough for your product contracts.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The technical report explains the direction. Your own evals should decide routing, retry behavior, and credit pricing.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-technical-report?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>DeepSeek V4 Paper: What Builders Should Notice</title>
      <dc:creator>Super Jarvis</dc:creator>
      <pubDate>Tue, 28 Apr 2026 17:14:19 +0000</pubDate>
      <link>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-paper-what-builders-should-notice-4gk1</link>
      <guid>https://core.forem.com/super_jarvis_76aa3fc6035d/deepseek-v4-paper-what-builders-should-notice-4gk1</guid>
      <description>&lt;p&gt;The DeepSeek V4 paper and model card describe the V4 family as MoE language models trained with MLA and DeepSeekSparse attention.&lt;/p&gt;

&lt;p&gt;Primary sources:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro" rel="noopener noreferrer"&gt;DeepSeek-V4-Pro model card&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main/DeepSeek_V4.pdf" rel="noopener noreferrer"&gt;DeepSeek_V4.pdf&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqm9bz98fq257xajf3enf.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqm9bz98fq257xajf3enf.jpg" alt="DeepSeek V4 paper reading workspace" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Read the paper as a product-routing document: architecture details matter most when they change latency, cost, context, or reliability.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Builder takeaways
&lt;/h2&gt;

&lt;p&gt;The release has two important product implications.&lt;/p&gt;

&lt;p&gt;First, the model family splits capacity. Pro is much larger and targets stronger reasoning. Flash is smaller and cheaper while still exposing a 1M context window.&lt;/p&gt;

&lt;p&gt;Second, the API pricing encourages cache-aware prompt design. Reused input can be cheaper than fresh cache-miss input, so teams should stabilize system prompts and repeated context templates.&lt;/p&gt;

&lt;h2&gt;
  
  
  What to test after reading
&lt;/h2&gt;

&lt;p&gt;After reading the paper, build a task set that reflects your product:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;long context retrieval and synthesis&lt;/li&gt;
&lt;li&gt;code repair and code review&lt;/li&gt;
&lt;li&gt;multi-step planning&lt;/li&gt;
&lt;li&gt;factual answers with web search&lt;/li&gt;
&lt;li&gt;structured JSON outputs&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then compare Flash and Pro with the same prompts. The paper explains architecture direction, but your eval decides routing.&lt;/p&gt;




&lt;p&gt;Source article: &lt;a href="https://deepseekv4.space/blog/deepseek-v4-paper?utm_source=devto&amp;amp;utm_medium=syndication&amp;amp;utm_campaign=blog-en" rel="noopener noreferrer"&gt;Read the original post&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Homepage: &lt;a href="https://deepseekv4.space/" rel="noopener noreferrer"&gt;Visit the site&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Model pages:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-pro" rel="noopener noreferrer"&gt;DeepSeek V4 Pro&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://deepseekv4.space/deepseek-v4-flash" rel="noopener noreferrer"&gt;DeepSeek V4 Flash&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
  </channel>
</rss>
