reddit-playbooks

r/LLMDevs

MODERATEplaybookView on Reddit ↗

A space for Enthusiasts, Developers and Researchers to discuss LLMs and their applications.

Subscribers
140K
Posts/day
22.1
Age
3.1y
Top week
137
Top month
162
Top year
2,492

Reddit Community Analysis: r/LLMDevs

1. Data Sources & Methodology

  • 310 unique posts after deduplication across 4 time periods (all-time, year, month, week), 4 pages each (16 raw JSON files)
  • Date collected: April 10, 2026
  • Subreddit subscribers: 141,138
  • Score range: 1 to 4,844
  • Median score: ~15 (dataset skews heavy with mid/long-tail since "week" captures many low-score posts)
  • Top 10 threshold: 1,192
  • Top 25 threshold: 491
  • Top 50 threshold: 213
  • Top 100 threshold: ~86
PeriodPostsScore RangeNotes
All-time~100168-4,844DeepSeek-era memes + canonical RAG writeups
Year~100108-2,493RAG guides, vibe coding memes, tool landscapes
Month2 dedicated-Single "Free Model List" crossover
Week~1001-182Mostly tool launches, help posts, low-score grind

This is a content strategy guide for distributing through r/LLMDevs, not a sociological study.

Cross-subreddit calibration: r/LLMDevs peaks at ~4,844 vs. r/LocalLLaMA's ~6,875, r/ClaudeAI's ~8,084, r/MachineLearning (similar tier), and r/AI_Agents (~2,500). With 141K subscribers (about 1/5 of r/LocalLLaMA, 1/4 of r/ClaudeAI), r/LLMDevs is a mid-tier technical AI sub. A realistic ceiling for a well-executed text-based builder writeup is ~900. A meme or vibe-coding snark post can hit 1,000-2,500. The viral top (4,000+) is reserved for image memes with universal developer appeal. Tool launches almost never clear 500.


2. Subreddit Character

r/LLMDevs is an uneasy hybrid: a technical forum for production LLM engineers that got overrun by meme accounts and beginner tourist traffic, with a small core of credible builders keeping the lights on with massive RAG/production writeups. It is roughly what r/MachineLearning would look like if you let Twitter in.

The community contains two distinct populations that share a feed:

  1. The working engineers – Mostly anonymous. They post long, comment-heavy text writeups about production RAG, multi-agent systems, fine-tuning, enterprise deployments, and framework comparisons. They upvote technical substance, downvote hand-waving, and politely destroy grifters in comments. Posts like "Building RAG systems at enterprise scale (20K+ docs)" (805, 1.0 ratio, 145 comments) and "I have written the same AI agent in 9 different python frameworks" (211, 0.99 ratio) are their canon.
  2. The meme/news crowd – Image-format screenshot memes about DeepSeek, vibe coders, ChatGPT users, and "big AI lab hypocrisy." These dominate the top of the leaderboard by raw score but contribute little to the technical discussion. "Soo Truee!" (4,844, 71 comments), "Everything is a wrapper" (1,192), "It's a free real estate from so called 'vibe coders'" (2,493).

Product launches are tolerated, not welcomed. Rule 5 ("No commercial self-promotion") explicitly requires FOSS licensing OR (source-available AND prior mod approval AND clear disclaimer). Rule 10 ("Be genuine: No disguised advertising") threatens permanent bans for astroturfing, sockpuppeting, and "pretending to be requests for help, feedback, or discussion but actually serving as marketing." This is the single most important rule to understand. The community actively calls out suspected promotion in comments, and low-ratio tool launches cluster in the 0.78-0.89 range ("We beat Google Deepmind but got killed by a chinese lab" 0.80, "CEO of Klarna claiming they are replacing Jira with a vibe coded app" 0.78, "Mythos is Opus 4.7" 0.78). Tool-launch flair average is depressed: the top-scoring "Tools" post peaks around 633 ("I accidentally built a vector database using video compression"), while the average Tools post lives in the 50-200 range.

Humor works, but specifically dev-insider humor. Not Chad memes. Not drake memes. The humor that lands is schadenfreude aimed at: (a) vibe coders shipping broken apps, (b) OpenAI's hypocrisy, (c) "everything is a wrapper" nihilism, (d) developers thanking their LLMs out of AGI-safety paranoia. The top-scoring post in the dataset ("Soo Truee!", 4,844) is a meme image. The top text post is long-form technical writing ("I built RAG for a rocket research company: 125K docs", 955).

Technical level is medium-high but uneven. Core readers understand RAG pipelines, embeddings, chunking strategies, vector DBs, context windows, quantization, LoRA, GRPO, and agent frameworks. But the tail includes lots of "Where do I start?" posts and "Best laptop for LLM?" help-wanted posts that cluster at single-digit scores.

Core cultural values, ranked by intensity:

  1. Skepticism of AI hype – The audience respects builders who admit failure ("It feels like most AI projects at work are failing and nobody talks about it" – 614, 0.97 ratio), who deconstruct framework bloat ("Not using Langchain ever !!!" – 184, 0.97), and who benchmark honestly ("LLM accuracy drops by 40% when increasing from single-turn to multi-turn" – 87, 0.97). Hype-laden marketing language gets flagged and downvoted.
  2. Anti-vibe-coding (or at least, wary of it) – Vibe coding is the single most polarizing topic. Mockery of vibe coders hits consistently high: "Vibe coding from a computer scientist's lens" (1,207, 0.94), "free real estate from so called 'vibe coders'" (2,493, 0.98), "Doctor vibe coding app under £75" (1,439, 0.94, 366 comments). But nuanced defenses also work if authored by credible devs: "The power of coding LLM in the hands of a 20+y experienced dev" (747, 0.93), "AI won't replace devs — but devs who master AI will replace the rest" (216, 0.82 – friction).
  3. DeepSeek/Qwen fandom – Shared with r/LocalLLaMA. "deepseek is a side project" (2,621), "DeepSeek R1 671B running on M2 Ultras" (2,311), "Prompted Deepseek R1 to choose a number 1-100, it thought for 96 seconds" (759), "Qwen is insane (testing real-time personal trainer)" (189), "Train your own Reasoning model like DeepSeek-R1 locally" (277).
  4. RAG as spiritual topic – RAG writeups are the most valuable text-format archetype. Long, detailed, enterprise-grounded RAG posts (especially Low_Acanthisitta7686's series) cluster in the 200-955 range with 40-165 comments each. RAG is also the topic most often mocked ("If RAG is dead, what will replace it?" – 160, 0.87 friction).
  5. Framework fatigue – LangChain hatred is a community-binding ritual ("Not using Langchain ever !!!" 184). Side-by-side framework comparisons perform well. Commenters prefer "just use the OpenAI SDK + a while loop" minimalism.
  6. Production > demos – "Production" and "enterprise" signal credibility. "200K+ documents", "60+ concurrent users", "6 months in production" – all hit. "Built a POC in a weekend" does not.

Enforcement mechanisms: 10 explicit rules, with the three most active being Rule 2 (Ask ethically – disclose purpose of surveys), Rule 5 (No commercial self-promotion – must be FOSS), and Rule 10 (No disguised advertising – permanent ban for astroturfing). No posting format requirement. No karma threshold. No mandatory PCP/Problem-Comparison-Pricing structure. Flair is optional but most top posts have one. Mods appear to enforce Rule 10 aggressively against thinly veiled promotion; several mid-tier posts show the telltale "controversial" ratio signature (0.78-0.87) of community-called-out promotion that wasn't removed.

How this sub differs from similar subs: On r/LocalLLaMA, you share something you can run on hardware you own. On r/MachineLearning, you cite papers and methods. On r/ClaudeAI, you tell a building-with-Claude story. On r/AI_Agents, you show the agent workflow diagram. On r/LLMDevs, you either (a) write 2,000+ words of production war stories, (b) drop a developer-insider meme, or (c) post a long framework/RAG comparison. Tool launches go here only as a last resort, and they mostly underperform.


3. The All-Time Leaderboard

Dataset median: ~15. Top-25 threshold: 491. Top-10 threshold: 1,192.

RankScoreFlairRatioCommentsFormatTitle
14,844Discussion0.9971IMAGESoo Truee!
22,621News1.0087IMAGEdeepseek is a side project
32,493Discussion0.98128IMAGEIt's a free real estate from so called "vibe coders"
42,311Discussion1.00111VIDEODeepSeek R1 671B (404GB) running on Apple M2 (2 M2 Ultras) flawlessly
51,761Discussion0.9983GALLERYOn to the next one 🤣
61,653News0.9951IMAGEState of OpenAI & Microsoft: Yesterday vs Today
71,439Discussion0.94366IMAGEDoctor vibe coding app under £75 alone in 5 days
81,207Discussion0.94147IMAGEVibe coding from a computer scientist's lens
91,192Discussion0.95125IMAGEEverything is a wrapper
10955Discussion0.98151TEXTI built RAG for a rocket research company: 125K docs, vision models for diagrams
11879Resource0.9959TEXTHow was DeepSeek-R1 built; For dummies
12870Resource0.95165IMAGEif people understood how good local LLMs are getting
13866Discussion0.98123TEXTI Built RAG Systems for Enterprises (20K+ Docs). Learning path
14805Discussion1.00145TEXTBuilding RAG systems at enterprise scale (20K+ docs)
15803Discussion0.96163IMAGEI let 24 AI models trade to see if they can manage risk
16759Discussion0.99136GALLERYPrompted Deepseek R1 to choose number 1-100, thought for 96 seconds
17747Discussion0.93172TEXTThe power of coding LLM in the hands of a 20+y experienced dev
18726Discussion0.9888TEXTWhy the heck is LLM observation and management tools so expensive?
19686Discussion0.9949IMAGEScary smart
20645Discussion0.87261IMAGEIt's DeepSee again (friction)
21633Tools0.9780TEXTI accidentally built a vector database using video compression
22614Discussion0.97148TEXTIt feels like most AI projects at work are failing
23559Tools0.9736IMAGENext generation of developers
24559Discussion0.9911IMAGELike fr 😅
25491Resource0.9941LINKI built Open Source Deep Research - here's how it works

Notable observations:

  • 9 of the top 10 are IMAGE/VIDEO/GALLERY meme format. Only 1 text post cracks the top 10 (rocket RAG).
  • Low_Acanthisitta7686 (Raj) owns 4 of the top 25 with enterprise RAG writeups. This is the single most prolific credible author.
  • All top 25 have ratios 0.87-1.00, meaning the high-scoring content is well-received even when contentious. Friction only appears in rank 20 ("It's DeepSee again", 0.87) and rank 7 ("Doctor vibe coding", 0.94) where the vibe-coding topic splits the community.
  • The top 3 highest-comment-count posts in the top 25 are discussion-provoking memes with a side of controversy: "Doctor vibe coding" (366 comments on 1,439 score – 0.254 C/U), "It's DeepSee again" (261 on 645 – 0.405 C/U), and "20+y dev coding LLM" (172 on 747 – 0.230 C/U).

4. Content Type Dominance at Scale

FlairCount Top 25Count Top 50Count AllAvg Score (All)Avg RatioBest Post (Title / Score)
Discussion1727~140~1300.93Soo Truee! / 4,844
Resource411~60~900.95How was DeepSeek-R1 built / 879
News25~20~1550.94deepseek is a side project / 2,621
Tools25~28~700.91I accidentally built a vector DB using video compression / 633
Great Discussion 💭03~10~950.86Are LLMs Models Collapsing? / 408
Great Resource 🚀03~18~550.91AI Coding Agent Dev Tools Landscape 2026 / 403
Help Wanted01~15~300.94Building an opensource Living Context Engine / 322
(no flair)00~3~1000.97I built this website to compare LLMs across benchmarks / 159

The surprising finding: "Great Discussion 💭" has an average ratio of 0.86 – the lowest of any flair. This flair appears to be applied to posts the community doesn't actually love. "AI won't replace devs — but devs who master AI will replace the rest" (216, 0.82), "Are LLMs Models Collapsing?" (408, 0.85), "The more I learn about LLMs, I get genuinely upset at how most use AI" (255, 0.83). The 💭 emoji flair is effectively a signal that a post is provocative and will split the room – treat it as a controversial flag, not a quality flag.

Resource vs Discussion split for technical writeups: Credible long-form technical posts actually perform better with the "Discussion" flair than with "Resource" at the top end. The top 4 text posts in the dataset (rocket RAG 955, enterprise RAG 866, enterprise RAG 805, 20+y dev 747) are all Discussion-flaired. Resource flair caps out around 879 (DeepSeek-R1 explainer). The implicit logic: "Resource" signals "here's a link to something else"; "Discussion" signals "here's my original work and my take on it, argue with me."


5. Content Archetypes That Work

Archetype 1: The Dev-Insider Meme (Ceiling: 4,844; Floor: ~400)

Examples:

  • "Soo Truee!" (4,844, 71 comments, Discussion, IMAGE)
  • "deepseek is a side project" (2,621, News, IMAGE)
  • "It's a free real estate from so called 'vibe coders'" (2,493, 128 comments)
  • "State of OpenAI & Microsoft: Yesterday vs Today" (1,653)
  • "Everything is a wrapper" (1,192)
  • "Scary smart" (686)
  • "Next generation of developers" (559)
  • "Like fr 😅" (559, only 11 comments – pure passive upvotes)

The pattern: Static screenshot (often a tweet, a meme template, or a chart). Self-explanatory within 2 seconds. Built around a shared developer grievance or a Chinese-model W. Low comment counts relative to score (C/U 0.01-0.08) – people upvote without stopping to discuss. Titles are short, often punctuation-free, sometimes lowercase.

Why it matters for distribution: This archetype has the highest score ceiling on the sub (top 9 of 10 posts), but it is useless for product distribution. No one clicks through a meme to your product. The meme archetype is a presence-builder: post memes under a username you want people to recognize, so when you later drop a tool launch they don't instantly flag you as a spam account.

Archetype 2: The Vibe Coding Mockery (Ceiling: ~2,500; Floor: ~300)

Examples:

  • "It's a free real estate from so called 'vibe coders'" (2,493, 128 comments)
  • "Doctor vibe coding app under £75 alone in 5 days" (1,439, 366 comments – highest C/U in top 10)
  • "Vibe coding from a computer scientist's lens" (1,207)
  • "In the Era of Vibe Coding Fundamentals are Still important!" (302)
  • "vibe coding:" (358, video)

The pattern: Screenshot or title mocking someone who shipped a broken/dangerous app with AI assistance, especially if they had no engineering background. The healthcare-data "Doctor vibe coding" post (0.94 ratio, 366 comments) is the template: show an AI-generated thing, ask "why is this concerning?", let the community pile on in comments.

Why it matters for distribution: This is the highest-engagement archetype (366-128 comments per post). If your tool addresses something vibe coders commonly get wrong (security, error handling, testing, architecture), you can frame your launch as "I built this because I kept seeing vibe coders do X." The framing earns pre-baked goodwill from the senior-dev core of the sub.

Archetype 3: The Enterprise RAG War Story (Ceiling: ~955; Floor: ~200)

Examples:

  • "I built RAG for a rocket research company: 125K docs" (955, 151 comments, TEXT) – Low_Acanthisitta7686
  • "I Built RAG Systems for Enterprises (20K+ Docs). Learning path" (866, 123 comments)
  • "Building RAG systems at enterprise scale (20K+ docs)" (805, 145 comments)
  • "I built RAG for 10K+ NASA docs (1950s–present) in 2 weeks" (284, 50 comments)
  • "Multi-modal RAG at scale: Processing 200K+ documents (pharma/finance/aerospace)" (218, 40 comments)
  • "7 months of Qwen in production enterprise" (243, 81 comments)
  • "EpsteinFiles-RAG: Building a RAG Pipeline on 2M+ Pages" (216, 33 comments)

The pattern: 2,000-5,000 word first-person TEXT post. Leads with a specific domain (pharma, aerospace, legal, finance) and a specific scale (20K / 125K / 200K docs). Mentions hard constraints: air-gapped, compliance, on-prem, concurrent users, H100 budget. Catalogs what failed before what worked. Names specific tools (Qwen, Docling, Qdrant, vLLM, LangGraph). Includes at least one "here's the weird hacky thing that actually worked" passage. Ends with offering to answer questions. Often has a "I used Claude for grammar/formatting polish" disclaimer at the bottom (transparency norm).

Why it matters for distribution: This is the single best archetype for credibility-building on the sub. Score ceiling is lower than memes but average comment count is higher (40-165), engagement is substantive, and readers become followers. Low_Acanthisitta7686 has built a personal brand as "the RAG guy" across ~6 posts. If you have genuine production experience, this is the highest-ROI archetype for turning readers into users/clients/followers.

Archetype 4: The Framework Slaughter (Ceiling: ~210; Floor: ~85)

Examples:

  • "I have written the same AI agent in 9 different python frameworks, here are my impressions" (211, 52 comments, 0.99)
  • "Not using Langchain ever !!!" (184, 60 comments, 0.97)
  • "What is currently the best production ready LLM framework?" (144, 57 comments)
  • "After months on Cursor, I just switched back to VS Code" (88, 28 comments)
  • "I realized why multi-agent LLM fails after building one" (156, 50 comments)

The pattern: A single developer does real work across N frameworks/tools, then posts honest comparative impressions. Ratios are near-perfect (0.97-0.99) because the format is inherently fair. Comments pile in with people adding their own takes.

Why it matters for distribution: If your product competes with a framework, DO NOT post "why my framework is better." Instead, write a neutral N-way comparison that happens to include yours. The community rewards the comparison; your tool becomes visible without the launch posture.

Archetype 5: The Cost/Observability Rant (Ceiling: ~726; Floor: ~50)

Examples:

  • "Why the heck is LLM observation and management tools so expensive?" (726, 88 comments)
  • "How do large AI apps manage LLM costs at scale?" (26, 48 comments)
  • "Are AI eval tools worth it or should we build in house?" (13, 15 comments)

The pattern: A developer names prices publicly (Langfuse $100/mo, Honeyhive, Promptlayer $50 for 100k requests, etc.), calls them "fuck you" expensive, and asks for alternatives. The rant tone actually helps – "Why the heck is..." hit 0.98 ratio and 88 comments.

Why it matters for distribution: If your tool is a cheaper alternative to an expensive incumbent category (observability, eval, prompt management, vector DB), this is your distribution vehicle. Pose as a frustrated dev, list real competitor pricing, ask for help. Only drop your tool in comments (if at all) after the thread is well-established. Posting your tool as the opening post almost always underperforms.

Archetype 6: The "I Built a Weird Thing" Curiosity Post (Ceiling: ~633; Floor: ~50)

Examples:

  • "I accidentally built a vector database using video compression" (633, 80 comments) – encoded docs as QR code video frames, 1.4GB for 10K PDFs
  • "I can't stop 'doomscrolling' Google maps so I built an AI that researches everywhere on Earth" (240, 50 comments)
  • "I trapped an LLM into an art installation and made it question its own existence endlessly" (87)
  • "I built a tiny LLM from scratch that talks like a fish" (66)
  • "I ran Claude Code in a self-learning loop until it translated our entire Python repo to TypeScript" (213, 32 comments)

The pattern: One-line title that describes a genuinely weird or stunt-like use of LLMs. The weirder, the better. Builder personally did something unusual. Often open-source. Comments are a mix of "how did you do this?" and "this is nonsense / brilliant."

Why it matters for distribution: The curiosity hook is the best way to get a tool launch past the sub's skepticism filter. "I built [practical tool]" underperforms; "I accidentally built [weird thing that happens to be useful]" performs. Rule of thumb: if your title doesn't sound like a stunt, rewrite it until it does.

Archetype 7: The Honest Failure Post (Ceiling: ~614; Floor: ~25)

Examples:

  • "It feels like most AI projects at work are failing and nobody talks about it" (614, 148 comments, 0.97)
  • "Our 'AI-first' strategy has turned into 'every team picks their own AI stack' chaos" (15)
  • "We hired 'AI Engineers' before. It didn't go well" (13, 38 comments)

The pattern: Low-ego first-person admission that the current state of AI-at-work is messier than the marketing. Specific details. No product pitch.

Why it matters for distribution: This archetype doesn't directly sell anything, but it's the highest-trust post format on the sub. Use it to establish you're "one of us" before launching anything. A Tuesday honest-failure post followed by a Friday tool launch under the same username outperforms a cold launch by a wide margin.


6. Format Analysis

FormatTop 25Top 50Full DatasetNotes
IMAGE14 (56%)22 (44%)~90 (29%)Dominates top tier – meme-driven
TEXT7 (28%)19 (38%)~160 (52%)The credibility format
GALLERY2 (8%)4 (8%)~10 (3%)Mostly screenshot sequences
VIDEO1 (4%)3 (6%)~18 (6%)Underused, high upside
LINK1 (4%)2 (4%)~12 (4%)Blog posts / GitHub links

Key pattern: IMAGE dominates the very top (top 10: 7 image, 2 gallery/video, 1 text), but TEXT dominates the substantive middle tier (ranks 10-50 are majority TEXT). The bimodal distribution matches the two-population community: memes vs. writeups.

What Format to Use For What

  • Tool/app launches → TEXT with detailed writeup. Image-only tool launches score <100. If you must use an image, pair it with a detailed text explanation in the body.
  • Production war stories → TEXT, 2,000+ words, first-person, specific. This is the archetype 3 playbook.
  • Framework comparisons → TEXT with markdown tables. "I have written the same AI agent in 9 different python frameworks" (211, 0.99) is the template.
  • Benchmarks / metrics → TEXT with tables, or IMAGE of a clean chart ("Which Format is Best for Passing Tables of Data to LLMs?" – 168, 0.97, IMAGE with a single table).
  • Memes / cultural observation → IMAGE. A tweet screenshot or chart meme. Keep titles short.
  • Demos → VIDEO (archetype 6 curiosity posts benefit from video). "Qwen is insane (testing a real-time personal trainer)" (189, VIDEO) and "DeepSeek R1 671B on Apple M2" (2,311, VIDEO) show video can clear the normal tool-launch ceiling when it shows something visibly impressive.
  • Dev tool/landscape maps → IMAGE with legend ("AI Coding Agent Dev Tools Landscape 2026" – 403). These are evergreen and get reposted by the same author over time.

What Makes a Good Demo Video on r/LLMDevs

VIDEO is underused (6% of dataset) but has high upside when it shows something technically impossible-looking. From analyzing top video posts:

  1. Show the absurd thing working – "671B model on an M2 Ultra", "real-time personal trainer", "drawio live editor". The video's job is to answer "wait, this actually works?"
  2. No narration required – Top videos are screen recordings with on-screen text, not talking heads. Viewers are muted scrolling.
  3. Under 60 seconds – Assume 3-5 seconds of attention. Front-load the hook.
  4. Show the input → output gap – Cut from "this is what I typed" to "this is what it produced" quickly.
  5. End with a title card or URL if open source. Low_Acanthisitta7686's style of putting the GitHub link in a pinned comment also works.

7. Flair/Category Strategy

Raw Performance Ranking

  1. Discussion – Highest average score, widest usage, and highest ceiling (4,844). Default choice for 80% of posts.
  2. News – Highest average-per-post at the top tier (~155). Use for genuinely newsworthy external events (model releases, corporate drama, leaks). Do not use News for your own product.
  3. Resource – Safe, ratio-positive, but lower ceiling (~879 peak). Best for guides, curated lists, and technical writeups that link to external content.
  4. Tools – The official "I made a tool" flair. Technically correct but performance is depressed (~70 avg, peak 633). Community skepticism of tool launches is priced in here.
  5. Great Resource 🚀 – Similar to Resource but feels more aspirational; often applied to lists-of-lists and landscape maps. Peaks at 403.
  6. Help Wanted – For questions. 322 peak on a "help me with my open-source project" post that was actually a soft-launch.
  7. Great Discussion 💭 – Counterintuitively the lowest-ratio flair (0.86 avg). Avoid this flair unless you want to mark your post as provocative.

Distribution Utility Ranking (different from raw)

For someone trying to promote a product, distribution utility differs from raw score:

  1. Discussion – Best utility. Gives you credibility-by-association with the top tier. Use for archetypes 3, 4, 5, and 7.
  2. Resource – Second-best utility for launches. Frame your tool as a learning resource, not a product. "Here's how I built X and what I learned" beats "Check out X."
  3. Tools – Paradoxically the worst utility flair despite being the "correct" one. Using Tools flair signals "this is an ad" to the most skeptical readers. Only use if your post is also overtly archetype 6 (weird curiosity) or if you're a known recurring builder.
  4. Help Wanted – Under-rated utility. A genuine "I built this, what am I doing wrong?" post earns more comment engagement than a launch post. You get the tool exposure without the promotional cost.
  5. News – Do NOT use for your own product. Reserve strictly for third-party events.
  6. Great Discussion 💭 – Avoid.

Pricing Model Hierarchy (most to least community-friendly)

Based on Rule 5 (must be FOSS OR source-available + mod approval + disclaimer) and observed reactions:

  1. MIT/Apache-licensed open source – Highest tier. "EpsteinFiles-RAG" explicitly states "Python, MIT Licensed, open source" as a selling point. Unsloth's yoracale posts are in this tier.
  2. Open-core with no paywalled features – Rule 5 explicitly permits this: "The free version must be functionally identical to any other version — no locked features behind a paywall / commercial / 'pro' license." This is strict – if your paid tier has any extra features, you're out.
  3. Free tier with clear paid tier (third-party service) – Tolerated if disclosed upfront. "Free Model List (API Keys)" (182) is a good example of a service comparison that was welcomed.
  4. Source-available + clear disclaimer + mod approval – Rule 5 path for commercial products. Rarely seen in the top 100; requires proactive mod contact.
  5. Closed-source commercial SaaS – Effectively banned by Rule 5. Posts masquerading as discussion will be removed and the user may be permanently banned under Rule 10.

There is no title-prefix tag culture on r/LLMDevs. You will not see [OS], [FREE], [Giveaway], or [Updated] in top posts. Don't use them. The lack of such tagging is itself part of the sub's norms.


8. Title Engineering

Deconstructing the top 10 titles:

  1. "Soo Truee!" (4,844) – Pure meme title. Image does all the work. Punctuation used as vibe signal.
  2. "deepseek is a side project" (2,621) – All-lowercase, deadpan. Plays the "casual throwaway" register that lands on the meme crowd.
  3. "It's a free real estate from so called 'vibe coders'" (2,493) – In-joke reference ("free real estate" meme) + scare quotes as editorial commentary.
  4. "DeepSeek R1 671B parameter model (404GB total) running on Apple M2 (2 M2 Ultras) flawlessly" (2,311) – Technical spec porn. Specific numbers, specific hardware, "flawlessly" as the payoff word.
  5. "On to the next one 🤣" (1,761) – Context-free in-joke title with emoji. Depends entirely on the gallery content.
  6. "State of OpenAI & Microsoft: Yesterday vs Today" (1,653) – Before/after framing + corporate-hypocrisy tease.
  7. "Doctor vibe coding app under £75 alone in 5 days" (1,439) – Specific profession + currency + time. Implied concern about healthcare data.
  8. "Vibe coding from a computer scientist's lens:" (1,207) – Expert credential + topic. Trailing colon implies "here's the take."
  9. "Everything is a wrapper" (1,192) – Absolute statement that nihilists and builders both nod at.
  10. "I built RAG for a rocket research company: 125K docs (1970s-present), vision models for rocket diagrams. Lessons from the technical challenges" (955) – Domain specificity + scale + time span + technical hook + "lessons" payoff.

Title Formulas

Formula 1: "I built [system] for [specific domain] at [specific scale]: [lessons / war stories]"

  • "I built RAG for a rocket research company: 125K docs" (955)
  • "I Built RAG Systems for Enterprises (20K+ Docs). Here's the learning path I wish I had" (866)
  • "Building RAG systems at enterprise scale (20K+ docs): lessons from 10+ enterprise implementations" (805)
  • "I built RAG for 10K+ NASA docs (1950s–present) in 2 weeks" (284)

Formula 2: "[Weird/absurd] + verb + [normal technical goal]"

  • "I accidentally built a vector database using video compression" (633)
  • "I ran Claude Code in a self-learning loop until it successfully translated our entire Python repo to TypeScript" (213)
  • "I trapped an LLM into an art installation and made it question its own existence endlessly" (87)

Formula 3: "Why the heck is [category] so [problem]?"

  • "Why the heck is LLM observation and management tools so expensive?" (726)
  • Implicit in "Not using Langchain ever !!!" (184)
  • "Why we ditched embeddings for knowledge graphs (and why chunking is fundamentally broken)" (195)

Formula 4: "[N] months of [tech] in [setting]: what actually works (and what doesn't)"

  • "7 months of Qwen in production enterprise: what actually works (and what doesn't)" (243)
  • "What I learned running an Always-on AI Agent in production for months (10 lessons)" (24)

Formula 5: The low-case throwaway meme title

  • "deepseek is a side project"
  • "vibe coding:"
  • "every ai app today"
  • "its funny cuz its true"
  • "Like fr 😅"
  • "you do what you gotta do"

This formula is for memes only, never for tool launches or text posts. The lowercase is a deliberate "I didn't try hard" signal.

Formula 6: "I have written / tested / compared [N] [things] – here are my impressions"

  • "I have written the same AI agent in 9 different python frameworks, here are my impressions" (211)
  • "I Built 3 Apps with DeepSeek, OpenAI o1, and Gemini - Here's What Performed Best" (241)
  • "Which Format is Best for Passing Tables of Data to LLMs?" (168)

Title Anti-Patterns (community-specific)

  1. "Check out [tool]" / "[Tool] just launched" – Almost no top-100 post has a generic launch title. Rewrite as an archetype.
  2. Star count / download count in the title – "My open-source project on different RAG techniques just hit 20K stars on GitHub" only scored 86. Growth metrics read as vanity and are ignored.
  3. Rocket emoji / fire emoji abuse – Minimal fire emoji usage in top posts. When emojis appear, it's usually 🤣, 🤨, 💭 – skeptical/comedic, not promotional.
  4. All-caps / sales-y titles – "Free Model List (API Keys)" is fine. "AMAZING NEW FRAMEWORK!!!" does not exist in the top 100.
  5. "Unpopular opinion:" – "Unpopular opinion: prompt engineering is just 'knowing how to talk to your coworker' rebranded" only scored 115 with 0.93 ratio. This framing reads as karma-farming.
  6. Vague questions from beginners – "Where to start from step 0" (3), "Am I not using LLM efficient enough?" (6). These get triaged to single digits. Rule 6 (No low-effort posts) is enforced socially.
  7. Day N of doing X – "Day 10 of showing reality of SaaS AI product" (2), "Day 15..." (1). Accountability-content style is explicitly rejected.

9. Engagement Patterns

Comments-to-upvote ratios by archetype:

ArchetypeAvg C/U RatioInterpretation
Vibe coding mockery0.15-0.25Discussion-heavy, people pile on
Honest failure post0.20-0.30Deep comment threads about shared pain
Enterprise RAG war story0.10-0.20Technical Q&A in comments
Framework slaughter0.20-0.30"Have you tried X?" additions
Cost/observability rant0.08-0.15Alternatives get suggested
Dev-insider meme0.01-0.08Passive upvotes, little discussion
Curiosity stunt post0.10-0.15"How did you do this?" questions
News (third-party)0.02-0.05Pure passive upvotes
Tool launch0.05-0.15Mostly skeptical comments
Help Wanted0.25-0.50People answer, even bad questions

Conditional recommendation:

  • If your goal is VISIBILITY, post a dev-insider meme or a News/leak item. Ceiling 2,500-4,800, passive scrolling, but maximum eyeball count.
  • If your goal is RELATIONSHIPS and discussion, post an enterprise RAG war story, a framework comparison, or an honest failure post. Ceiling is lower (200-950) but comment counts and follower-gain are much higher.
  • If your goal is DIRECT PRODUCT CLICKS, the archetype 6 "weird curiosity" post is the only format that reliably converts. "I accidentally built X" beats "I built X" by a large margin.

Highest-discussion topics (most comments regardless of score)

  1. Vibe coding and whether it's valid – "Doctor vibe coding" (366), "Vibe coding from a computer scientist's lens" (147), "CEO of Klarna claiming they are replacing Jira with a vibe coded app" (93), "power of coding LLM in 20+y dev" (172)
  2. Deepseek / Chinese lab drama – "DeepSee again" (261), "RAG is dead" (113), "deepseek is a side project" (87), "OpenAI calls for bans on DeepSeek" (44)
  3. RAG architecture debates – "Why we ditched embeddings for knowledge graphs" (83), "If RAG is dead, what will replace it?" (113), "Is RAG really necessary for LLM → SQL systems" (100), "50-100 PDFs, what RAG approach?" (95)
  4. LLMs killing critical thinking / mental health – "The more I learn about LLMs, I get genuinely upset" (247), "It feels like most AI projects at work are failing" (148)
  5. Model selection / framework fatigue – "What is currently the best production ready LLM framework?" (57), "9 frameworks impressions" (52), "Not using Langchain ever" (60)

10. What Gets Downvoted

Posts with ratios below 0.85 in the dataset:

TitleScoreRatioWhy it struggled
CEO of Klarna claiming they are replacing Jira with a vibe coded app880.78Perceived as CEO-worship; community distrusts the claim
Mythos is Opus 4.7…830.78Cryptic product-tease title with no substance
We beat Google Deepmind but got killed by a chinese lab790.80Self-aggrandizing framing, suspected promo
AI won't replace devs — but devs who master AI will replace the rest2160.82Clichéd "hustle dev" framing
Found an open-source goldmine!1860.83Vague clickbait title, felt like an ad
The more I learn about LLMs, I get genuinely upset at how most use AI2550.83Moralizing tone, comes across as preachy
I stand by this1880.83Low-context image post, unclear point
Carnegie Mellon just dropped one of the most important AI agent papers1780.83Hyperbolic title ("most important")
Only this LLM books you need2720.89Affiliate-adjacent reading list
How AI is transforming senior engineers into code monkeys1870.90Controversial-by-design framing
Are LLMs Models Collapsing?4080.85AI-slop-feeling writeup
Meta can now predict what your brain is thinking1340.86Clickbait rewrite of a research story
NVIDIA dropped one of The most important AI paper of 20253130.88Hyperbolic title ("most important")
It's DeepSee again6450.87Pro-Deepseek-fatigue sentiment friction

Ratio Tiers

  • Above 0.94 – Universally well-received. Safe for technical writeups, memes, and honest failures.
  • 0.85-0.94 – Net positive but with friction. Often provocative-by-design (controversial opinions, contested framings, viral memes). Expect pile-on comments.
  • Below 0.85 – Controversial or community-hostile. Usually triggers one or more of: perceived promotion, hyperbolic claims, CEO/celebrity worship, moralizing, or suspicion of being karma-farmed/AI-generated.

Anti-Patterns (community-specific)

  1. "Most important [X] of [year]" hyperbole – The community distrusts superlatives. "NVIDIA dropped one of The most important AI paper of 2025" (313, 0.88), "Carnegie Mellon...most important AI agent papers" (178, 0.83).
  2. Suspected promotion flagged as "discussion" – Rule 10 says this earns a permanent ban. Posts that look like soft launches get rated down even if they survive mod removal.
  3. Moralizing about "how people use AI" – The "upset at how people use AI" tone reads as condescending. Technical criticism is welcome; value judgments about users are not.
  4. CEO / founder worship – The Klarna CEO post (0.78) and variants that take corporate PR at face value get punished. The community treats CEO claims as marketing until proven otherwise.
  5. Day N accountability content – "Day 10 of showing reality of SaaS AI product" (2 score). This is treated as spam regardless of content.
  6. Clickbait headlines on research papers – "Meta can now predict what your brain is thinking. read that again." (134, 0.86). The community prefers papers linked directly with neutral framing.
  7. AI-generated post tells – The rocket RAG post (955) defused this by disclosing "I used Claude for grammar/formatting polish." Non-disclosed AI-feeling posts (long, overly structured, perfect grammar, generic insights) cluster in the 0.82-0.89 friction band.

There is no public "hall of shame" or blacklist that I can detect. Enforcement is via: (1) mod removal for Rule 5/10 violations, (2) community downvote friction for the above patterns, (3) commenter callouts ("is this an ad?"). The permanent-ban threat in Rule 10 appears to be taken seriously based on the low volume of obvious astroturf in the top 100.


11. The Distribution Playbook

Phase 1: Pre-launch (Week -4 to 0)

Goal: Establish a posting identity the community recognizes before you ever mention your product.

  1. Read Rule 5 and Rule 10 carefully. If your product is not FOSS-licensed with free version identical to paid, contact mods first or do not post at all. The permanent ban is real.
  2. Post 2-3 low-stakes contribution posts first, spaced across 2-4 weeks. Good options:
    • An honest-failure observation (archetype 7)
    • A framework comparison you actually did (archetype 4)
    • A technical rant about pricing in a category adjacent to yours (archetype 5)
    • A curiosity/stunt post using LLMs (archetype 6)
  3. Comment substantively on other RAG/framework threads. The top comment-earners in the community (Low_Acanthisitta7686, Nir777, yoracale, Arindam_200) are recognized because they show up in comments, not just posts.
  4. Avoid the "Tools" flair for these warmup posts. Use Discussion or Resource.

Phase 2: Launch Day

Optimal launch formula (for FOSS tools):

  • Format: TEXT, 800-2,500 words
  • Flair: Discussion (NOT Tools) if your post is a writeup. Resource if it's a curated list or guide.
  • Title: Use Formula 1 ("I built X for Y at Z scale: lessons") OR Formula 2 ("I accidentally built..."). Avoid generic launch phrasing.
  • Opening paragraph: Who you are, what problem you hit, why existing tools didn't work. Be specific about the problem.
  • Middle: The actual technical content. What failed first. What you tried. What worked. Include at least one "this is the weird hack that actually worked" moment.
  • GitHub link: In the post body, mid-paragraph, not in the title.
  • Closing: Offer to answer questions. Mention your license explicitly if MIT/Apache.
  • Disclosure footer: If you used an LLM to polish grammar, say so ("Note: I used Claude for grammar/formatting polish" is the community-accepted template).

Timing: r/LLMDevs has global audience; US weekday mornings (13:00-17:00 UTC) work well based on the top posts' timestamps (rocket RAG 16:02 UTC, enterprise RAG writeups mostly 13:00-18:00 UTC). Avoid weekends – engagement drops ~40%.

Phase 3: First 24-48 Hours

Comment strategy – pre-written reply templates for the 4-5 most common objections:

  1. "Is this vibe coded?" / "How much of this was written by AI?"

    "The architecture and all technical decisions are mine. I did use [Claude/ChatGPT] for [grammar polish / boilerplate / docs]. The [core algorithm / retrieval logic / X] is hand-written. Happy to walk through any specific file."

  2. "Why not just use LangChain/LlamaIndex/LangGraph?"

    "I did try [framework] first. The specific thing that broke for me was [concrete issue]. For [simple case] LangChain is probably fine, but for [my use case] I ended up needing [specific control]. Here's the commit where I ripped it out: [link]."

  3. "What's your pricing model?" / "Is this going to get paywalled?"

    "MIT licensed, will stay MIT licensed. There is no paid tier planned. If I ever add hosting, the self-hosted version will remain functionally identical — that's a commitment I'll put in the repo."

  4. "Does this actually work at scale?"

    "Honest answer: I've tested it on [X documents / Y concurrent users / Z tokens/sec]. Above that, I haven't. If you hit the wall at larger scale, open an issue and I'll look at it with you."

  5. "How does this compare to [competitor]?"

    "[Competitor] does [X] better. Mine does [Y] better. The main reason to use mine is [specific thing]. If you don't need [that thing], use [competitor] – no hard feelings."

Monitor your ratio in the first 4 hours:

  • Ratio > 0.94 + rising score: You're on track. Keep answering comments promptly.
  • Ratio 0.85-0.94: Friction. Someone accused you of promotion or a technical claim is being challenged. Address the specific challenge in a top-level edit.
  • Ratio < 0.85: Community is hostile. Do NOT edit the post to argue. Engage the top critic respectfully in comments. Consider that the post may need to be deleted and relaunched with different framing.
  • Ratio > 0.94 but score flat after 4 hours: Not controversial but not viral. Content is probably fine; positioning failed. This is the most common outcome for tool launches. Learn and try a different archetype next time.

Phase 4: Ongoing Presence

Repeat cadence: Low_Acanthisitta7686 posts roughly every 2-6 weeks. yoracale posts every 3-8 weeks. Arindam_200 posts every 1-3 weeks. These are the cadences that produce a recognized identity without triggering "spam" callouts.

Building on a hit: If a launch post breaks 400, your next post should be a follow-up with NEW technical content (a deep-dive on one sub-problem from the original), not a v1.1 announcement. Version bump posts underperform.

Stealth distribution tactics:

  1. Answering questions in "Help Wanted" threads – The community has a constant stream of "how do I RAG 50 PDFs" / "which framework should I use" posts. Answer them substantively with a section mentioning your tool only if it's actually relevant. This earns more clicks than a launch post.
  2. Contributing to framework comparison posts – When someone posts "I tried 9 frameworks", comment with an 10th that happens to be yours.
  3. The observability rant inclusion – Posts like "Why is LLM observation so expensive?" list specific competitor pricing. A comment that lists your open-source alternative is welcomed.
  4. Crosslinking your own GitHub in a technical writeup – If your archetype 3 writeup naturally mentions a tool you built, include the link in-line, not as a CTA. Readers click through more when it feels incidental.
  5. Show up in comments of the RAG-guy posts – Low_Acanthisitta7686's comment threads are full of builders. Commenting substantively there (not pitching) gets your username in front of the core technical audience.

Score-Tier Calibration

Tell yourself the truth before posting:

  • Tool launches on r/LLMDevs almost never exceed 500. If you need more visibility, you need a different archetype (meme, RAG writeup, or curiosity post).
  • Enterprise RAG writeups realistically hit 200-950. If you don't have production experience, don't try this archetype – commenters will see through it.
  • Memes can hit 4,000+, but require developer-insider cultural fluency and deliver zero direct product clicks.
  • Framework comparisons hit 150-250 almost regardless of author, if the comparison is honest.
  • Pure help posts cluster at 1-30. The "Help Wanted" flair is for getting comments, not upvotes.

Post-Publication Measurement

  • 4 hours: If score < 20 and ratio < 0.90, the post has failed. Do not delete – a deleted post can look suspicious. Learn and try again in 2-4 weeks.
  • 24 hours: Final score is roughly 70% set. A post at 150 after 24 hours will end at 200-250. A post at 500 after 24 hours will end at 700-900.
  • 7 days: If the post is in a recurring-author pattern (e.g., your third writeup), watch whether the score exceeds the prior post. If flat or declining, consider whether your archetype is saturating.
  • Clickthroughs vs upvotes: Expect roughly 2-5% clickthrough on GitHub links in text posts (much lower than r/LocalLLaMA's ~10%). This community reads and argues more than it clicks.

12. Applying This to Any Project

Quick-reference checklist (read before posting)

  • My project is MIT/Apache or FOSS-licensed with free version functionally identical to any paid version (Rule 5)
  • My post is not disguised advertising; I am not violating Rule 10
  • I am using Discussion or Resource flair, not Tools (unless Tools is the clearest signal)
  • My title is specific and uses Formula 1, 2, 3, 4, or 6 – not a generic launch phrase
  • I have 800+ words of actual technical substance in the body
  • My GitHub/product link is in-body, not in the title
  • I have a pre-written reply template for "is this vibe coded?" and "why not use LangChain?"
  • I disclosed any LLM-assisted writing ("used Claude for grammar polish")
  • I have posted at least one non-promotional contribution in the past 2 weeks under the same username
  • I am launching on a US weekday morning (13:00-17:00 UTC)
  • My title contains zero "most important", zero all-caps, zero star/download counts
  • I have my pricing model explicitly stated (or can answer it in comments within 5 minutes of posting)

Scenario-Based Launch Guides

Scenario 1: Your product is free and open-source (MIT/Apache)

  • Optimal launch formula: Archetype 3 (production war story) OR Archetype 6 (weird curiosity) + Discussion flair
  • Title template: "I built [specific tool] for [specific domain]: [lessons / scale / weird thing that worked]"
  • Key risk: Over-claiming scale or production use when you haven't actually deployed it. The senior engineers in comments will ask follow-up questions that expose the gap.

Scenario 2: Your product uses one-time/lifetime pricing

  • Optimal launch formula: Do NOT launch on r/LLMDevs. Rule 5 explicitly requires FOSS or (source-available + mod approval + disclaimer). Post on r/LocalLLaMA (friendlier to lifetime pricing) or r/macapps (where lifetime pricing is celebrated) instead.
  • Key risk: Permanent ban under Rule 10 if you frame a commercial product as "discussion."

Scenario 3: Your product uses subscription pricing

  • Optimal launch formula: Not on r/LLMDevs. See Scenario 2.
  • If you must engage here: The only safe format is a "cost rant" post (archetype 5) where you enumerate competitor pricing including your own, framed neutrally as a category map. Do not editorialize that your tool is the answer – let readers find it.
  • Key risk: Subscription-priced product launches are the single most common cause of permanent bans on this sub.

Scenario 4: Your product was built with AI (vibe coded)

  • Optimal launch formula: Lead with the AI-assisted-ness as a confession, not a feature. "I used Claude Code to build X. Here's what it got wrong and how I had to intervene." Archetype 3 crossed with archetype 7 (honest failure).
  • Title template: "I ran Claude Code in a loop until it [specific unreasonable task]. Here's what broke." (The "in a loop until it..." post scored 213, 0.92 ratio.)
  • Key risk: Hiding AI assistance and being exposed in comments. Always disclose upfront.

Scenario 5: Your product is a RAG/agent framework competing with LangChain/LlamaIndex

  • Optimal launch formula: Archetype 4 (framework slaughter) with your framework as one option among many. Do a genuine side-by-side with code, not a pitch.
  • Title template: "I wrote the same [task] in [N] frameworks: here's what actually works"
  • Key risk: Readers smell self-promotion if yours comes out suspiciously best. The "9 frameworks" post (211, 0.99) worked because it included "No framework" as the honest best-for-beginners answer.

Cross-Posting Guidance

Use the table below when reusing content across subs:

  • On r/LocalLLaMA: Frame as "runs on hardware you own." Emphasize open weights, quantization, VRAM. GPU rig photos help.
  • On r/ClaudeAI: Frame as "I built this with Claude Code." Emphasize the agentic workflow you used.
  • On r/AI_Agents: Frame as "multi-agent system with [specific orchestration pattern]." Emphasize the diagram.
  • On r/MachineLearning: Frame as methodology + paper citations. Emphasize reproducibility and benchmarks.
  • On r/LLMDevs: Frame as a production war story OR a framework comparison OR a dev-insider observation. Emphasize what broke, what scale, what took 40% of your time. The "this is what nobody warns you about" frame is the highest-trust positioning this community offers.

The same content can generally be reframed 3-4 ways across these subs without issue, provided the title and opening paragraph are rewritten each time and the posts are spaced 3+ days apart.