Reddit Community Analysis: r/LLMDevs
1. Data Sources & Methodology
- 310 unique posts after deduplication across 4 time periods (all-time, year, month, week), 4 pages each (16 raw JSON files)
- Date collected: April 10, 2026
- Subreddit subscribers: 141,138
- Score range: 1 to 4,844
- Median score: ~15 (dataset skews heavy with mid/long-tail since "week" captures many low-score posts)
- Top 10 threshold: 1,192
- Top 25 threshold: 491
- Top 50 threshold: 213
- Top 100 threshold: ~86
| Period | Posts | Score Range | Notes |
|---|---|---|---|
| All-time | ~100 | 168-4,844 | DeepSeek-era memes + canonical RAG writeups |
| Year | ~100 | 108-2,493 | RAG guides, vibe coding memes, tool landscapes |
| Month | 2 dedicated | - | Single "Free Model List" crossover |
| Week | ~100 | 1-182 | Mostly tool launches, help posts, low-score grind |
This is a content strategy guide for distributing through r/LLMDevs, not a sociological study.
Cross-subreddit calibration: r/LLMDevs peaks at ~4,844 vs. r/LocalLLaMA's ~6,875, r/ClaudeAI's ~8,084, r/MachineLearning (similar tier), and r/AI_Agents (~2,500). With 141K subscribers (about 1/5 of r/LocalLLaMA, 1/4 of r/ClaudeAI), r/LLMDevs is a mid-tier technical AI sub. A realistic ceiling for a well-executed text-based builder writeup is ~900. A meme or vibe-coding snark post can hit 1,000-2,500. The viral top (4,000+) is reserved for image memes with universal developer appeal. Tool launches almost never clear 500.
2. Subreddit Character
r/LLMDevs is an uneasy hybrid: a technical forum for production LLM engineers that got overrun by meme accounts and beginner tourist traffic, with a small core of credible builders keeping the lights on with massive RAG/production writeups. It is roughly what r/MachineLearning would look like if you let Twitter in.
The community contains two distinct populations that share a feed:
- The working engineers – Mostly anonymous. They post long, comment-heavy text writeups about production RAG, multi-agent systems, fine-tuning, enterprise deployments, and framework comparisons. They upvote technical substance, downvote hand-waving, and politely destroy grifters in comments. Posts like "Building RAG systems at enterprise scale (20K+ docs)" (805, 1.0 ratio, 145 comments) and "I have written the same AI agent in 9 different python frameworks" (211, 0.99 ratio) are their canon.
- The meme/news crowd – Image-format screenshot memes about DeepSeek, vibe coders, ChatGPT users, and "big AI lab hypocrisy." These dominate the top of the leaderboard by raw score but contribute little to the technical discussion. "Soo Truee!" (4,844, 71 comments), "Everything is a wrapper" (1,192), "It's a free real estate from so called 'vibe coders'" (2,493).
Product launches are tolerated, not welcomed. Rule 5 ("No commercial self-promotion") explicitly requires FOSS licensing OR (source-available AND prior mod approval AND clear disclaimer). Rule 10 ("Be genuine: No disguised advertising") threatens permanent bans for astroturfing, sockpuppeting, and "pretending to be requests for help, feedback, or discussion but actually serving as marketing." This is the single most important rule to understand. The community actively calls out suspected promotion in comments, and low-ratio tool launches cluster in the 0.78-0.89 range ("We beat Google Deepmind but got killed by a chinese lab" 0.80, "CEO of Klarna claiming they are replacing Jira with a vibe coded app" 0.78, "Mythos is Opus 4.7" 0.78). Tool-launch flair average is depressed: the top-scoring "Tools" post peaks around 633 ("I accidentally built a vector database using video compression"), while the average Tools post lives in the 50-200 range.
Humor works, but specifically dev-insider humor. Not Chad memes. Not drake memes. The humor that lands is schadenfreude aimed at: (a) vibe coders shipping broken apps, (b) OpenAI's hypocrisy, (c) "everything is a wrapper" nihilism, (d) developers thanking their LLMs out of AGI-safety paranoia. The top-scoring post in the dataset ("Soo Truee!", 4,844) is a meme image. The top text post is long-form technical writing ("I built RAG for a rocket research company: 125K docs", 955).
Technical level is medium-high but uneven. Core readers understand RAG pipelines, embeddings, chunking strategies, vector DBs, context windows, quantization, LoRA, GRPO, and agent frameworks. But the tail includes lots of "Where do I start?" posts and "Best laptop for LLM?" help-wanted posts that cluster at single-digit scores.
Core cultural values, ranked by intensity:
- Skepticism of AI hype – The audience respects builders who admit failure ("It feels like most AI projects at work are failing and nobody talks about it" – 614, 0.97 ratio), who deconstruct framework bloat ("Not using Langchain ever !!!" – 184, 0.97), and who benchmark honestly ("LLM accuracy drops by 40% when increasing from single-turn to multi-turn" – 87, 0.97). Hype-laden marketing language gets flagged and downvoted.
- Anti-vibe-coding (or at least, wary of it) – Vibe coding is the single most polarizing topic. Mockery of vibe coders hits consistently high: "Vibe coding from a computer scientist's lens" (1,207, 0.94), "free real estate from so called 'vibe coders'" (2,493, 0.98), "Doctor vibe coding app under £75" (1,439, 0.94, 366 comments). But nuanced defenses also work if authored by credible devs: "The power of coding LLM in the hands of a 20+y experienced dev" (747, 0.93), "AI won't replace devs — but devs who master AI will replace the rest" (216, 0.82 – friction).
- DeepSeek/Qwen fandom – Shared with r/LocalLLaMA. "deepseek is a side project" (2,621), "DeepSeek R1 671B running on M2 Ultras" (2,311), "Prompted Deepseek R1 to choose a number 1-100, it thought for 96 seconds" (759), "Qwen is insane (testing real-time personal trainer)" (189), "Train your own Reasoning model like DeepSeek-R1 locally" (277).
- RAG as spiritual topic – RAG writeups are the most valuable text-format archetype. Long, detailed, enterprise-grounded RAG posts (especially Low_Acanthisitta7686's series) cluster in the 200-955 range with 40-165 comments each. RAG is also the topic most often mocked ("If RAG is dead, what will replace it?" – 160, 0.87 friction).
- Framework fatigue – LangChain hatred is a community-binding ritual ("Not using Langchain ever !!!" 184). Side-by-side framework comparisons perform well. Commenters prefer "just use the OpenAI SDK + a while loop" minimalism.
- Production > demos – "Production" and "enterprise" signal credibility. "200K+ documents", "60+ concurrent users", "6 months in production" – all hit. "Built a POC in a weekend" does not.
Enforcement mechanisms: 10 explicit rules, with the three most active being Rule 2 (Ask ethically – disclose purpose of surveys), Rule 5 (No commercial self-promotion – must be FOSS), and Rule 10 (No disguised advertising – permanent ban for astroturfing). No posting format requirement. No karma threshold. No mandatory PCP/Problem-Comparison-Pricing structure. Flair is optional but most top posts have one. Mods appear to enforce Rule 10 aggressively against thinly veiled promotion; several mid-tier posts show the telltale "controversial" ratio signature (0.78-0.87) of community-called-out promotion that wasn't removed.
How this sub differs from similar subs: On r/LocalLLaMA, you share something you can run on hardware you own. On r/MachineLearning, you cite papers and methods. On r/ClaudeAI, you tell a building-with-Claude story. On r/AI_Agents, you show the agent workflow diagram. On r/LLMDevs, you either (a) write 2,000+ words of production war stories, (b) drop a developer-insider meme, or (c) post a long framework/RAG comparison. Tool launches go here only as a last resort, and they mostly underperform.
3. The All-Time Leaderboard
Dataset median: ~15. Top-25 threshold: 491. Top-10 threshold: 1,192.
| Rank | Score | Flair | Ratio | Comments | Format | Title |
|---|---|---|---|---|---|---|
| 1 | 4,844 | Discussion | 0.99 | 71 | IMAGE | Soo Truee! |
| 2 | 2,621 | News | 1.00 | 87 | IMAGE | deepseek is a side project |
| 3 | 2,493 | Discussion | 0.98 | 128 | IMAGE | It's a free real estate from so called "vibe coders" |
| 4 | 2,311 | Discussion | 1.00 | 111 | VIDEO | DeepSeek R1 671B (404GB) running on Apple M2 (2 M2 Ultras) flawlessly |
| 5 | 1,761 | Discussion | 0.99 | 83 | GALLERY | On to the next one 🤣 |
| 6 | 1,653 | News | 0.99 | 51 | IMAGE | State of OpenAI & Microsoft: Yesterday vs Today |
| 7 | 1,439 | Discussion | 0.94 | 366 | IMAGE | Doctor vibe coding app under £75 alone in 5 days |
| 8 | 1,207 | Discussion | 0.94 | 147 | IMAGE | Vibe coding from a computer scientist's lens |
| 9 | 1,192 | Discussion | 0.95 | 125 | IMAGE | Everything is a wrapper |
| 10 | 955 | Discussion | 0.98 | 151 | TEXT | I built RAG for a rocket research company: 125K docs, vision models for diagrams |
| 11 | 879 | Resource | 0.99 | 59 | TEXT | How was DeepSeek-R1 built; For dummies |
| 12 | 870 | Resource | 0.95 | 165 | IMAGE | if people understood how good local LLMs are getting |
| 13 | 866 | Discussion | 0.98 | 123 | TEXT | I Built RAG Systems for Enterprises (20K+ Docs). Learning path |
| 14 | 805 | Discussion | 1.00 | 145 | TEXT | Building RAG systems at enterprise scale (20K+ docs) |
| 15 | 803 | Discussion | 0.96 | 163 | IMAGE | I let 24 AI models trade to see if they can manage risk |
| 16 | 759 | Discussion | 0.99 | 136 | GALLERY | Prompted Deepseek R1 to choose number 1-100, thought for 96 seconds |
| 17 | 747 | Discussion | 0.93 | 172 | TEXT | The power of coding LLM in the hands of a 20+y experienced dev |
| 18 | 726 | Discussion | 0.98 | 88 | TEXT | Why the heck is LLM observation and management tools so expensive? |
| 19 | 686 | Discussion | 0.99 | 49 | IMAGE | Scary smart |
| 20 | 645 | Discussion | 0.87 | 261 | IMAGE | It's DeepSee again (friction) |
| 21 | 633 | Tools | 0.97 | 80 | TEXT | I accidentally built a vector database using video compression |
| 22 | 614 | Discussion | 0.97 | 148 | TEXT | It feels like most AI projects at work are failing |
| 23 | 559 | Tools | 0.97 | 36 | IMAGE | Next generation of developers |
| 24 | 559 | Discussion | 0.99 | 11 | IMAGE | Like fr 😅 |
| 25 | 491 | Resource | 0.99 | 41 | LINK | I built Open Source Deep Research - here's how it works |
Notable observations:
- 9 of the top 10 are IMAGE/VIDEO/GALLERY meme format. Only 1 text post cracks the top 10 (rocket RAG).
- Low_Acanthisitta7686 (Raj) owns 4 of the top 25 with enterprise RAG writeups. This is the single most prolific credible author.
- All top 25 have ratios 0.87-1.00, meaning the high-scoring content is well-received even when contentious. Friction only appears in rank 20 ("It's DeepSee again", 0.87) and rank 7 ("Doctor vibe coding", 0.94) where the vibe-coding topic splits the community.
- The top 3 highest-comment-count posts in the top 25 are discussion-provoking memes with a side of controversy: "Doctor vibe coding" (366 comments on 1,439 score – 0.254 C/U), "It's DeepSee again" (261 on 645 – 0.405 C/U), and "20+y dev coding LLM" (172 on 747 – 0.230 C/U).
4. Content Type Dominance at Scale
| Flair | Count Top 25 | Count Top 50 | Count All | Avg Score (All) | Avg Ratio | Best Post (Title / Score) |
|---|---|---|---|---|---|---|
| Discussion | 17 | 27 | ~140 | ~130 | 0.93 | Soo Truee! / 4,844 |
| Resource | 4 | 11 | ~60 | ~90 | 0.95 | How was DeepSeek-R1 built / 879 |
| News | 2 | 5 | ~20 | ~155 | 0.94 | deepseek is a side project / 2,621 |
| Tools | 2 | 5 | ~28 | ~70 | 0.91 | I accidentally built a vector DB using video compression / 633 |
| Great Discussion 💭 | 0 | 3 | ~10 | ~95 | 0.86 | Are LLMs Models Collapsing? / 408 |
| Great Resource 🚀 | 0 | 3 | ~18 | ~55 | 0.91 | AI Coding Agent Dev Tools Landscape 2026 / 403 |
| Help Wanted | 0 | 1 | ~15 | ~30 | 0.94 | Building an opensource Living Context Engine / 322 |
| (no flair) | 0 | 0 | ~3 | ~100 | 0.97 | I built this website to compare LLMs across benchmarks / 159 |
The surprising finding: "Great Discussion 💭" has an average ratio of 0.86 – the lowest of any flair. This flair appears to be applied to posts the community doesn't actually love. "AI won't replace devs — but devs who master AI will replace the rest" (216, 0.82), "Are LLMs Models Collapsing?" (408, 0.85), "The more I learn about LLMs, I get genuinely upset at how most use AI" (255, 0.83). The 💭 emoji flair is effectively a signal that a post is provocative and will split the room – treat it as a controversial flag, not a quality flag.
Resource vs Discussion split for technical writeups: Credible long-form technical posts actually perform better with the "Discussion" flair than with "Resource" at the top end. The top 4 text posts in the dataset (rocket RAG 955, enterprise RAG 866, enterprise RAG 805, 20+y dev 747) are all Discussion-flaired. Resource flair caps out around 879 (DeepSeek-R1 explainer). The implicit logic: "Resource" signals "here's a link to something else"; "Discussion" signals "here's my original work and my take on it, argue with me."
5. Content Archetypes That Work
Archetype 1: The Dev-Insider Meme (Ceiling: 4,844; Floor: ~400)
Examples:
- "Soo Truee!" (4,844, 71 comments, Discussion, IMAGE)
- "deepseek is a side project" (2,621, News, IMAGE)
- "It's a free real estate from so called 'vibe coders'" (2,493, 128 comments)
- "State of OpenAI & Microsoft: Yesterday vs Today" (1,653)
- "Everything is a wrapper" (1,192)
- "Scary smart" (686)
- "Next generation of developers" (559)
- "Like fr 😅" (559, only 11 comments – pure passive upvotes)
The pattern: Static screenshot (often a tweet, a meme template, or a chart). Self-explanatory within 2 seconds. Built around a shared developer grievance or a Chinese-model W. Low comment counts relative to score (C/U 0.01-0.08) – people upvote without stopping to discuss. Titles are short, often punctuation-free, sometimes lowercase.
Why it matters for distribution: This archetype has the highest score ceiling on the sub (top 9 of 10 posts), but it is useless for product distribution. No one clicks through a meme to your product. The meme archetype is a presence-builder: post memes under a username you want people to recognize, so when you later drop a tool launch they don't instantly flag you as a spam account.
Archetype 2: The Vibe Coding Mockery (Ceiling: ~2,500; Floor: ~300)
Examples:
- "It's a free real estate from so called 'vibe coders'" (2,493, 128 comments)
- "Doctor vibe coding app under £75 alone in 5 days" (1,439, 366 comments – highest C/U in top 10)
- "Vibe coding from a computer scientist's lens" (1,207)
- "In the Era of Vibe Coding Fundamentals are Still important!" (302)
- "vibe coding:" (358, video)
The pattern: Screenshot or title mocking someone who shipped a broken/dangerous app with AI assistance, especially if they had no engineering background. The healthcare-data "Doctor vibe coding" post (0.94 ratio, 366 comments) is the template: show an AI-generated thing, ask "why is this concerning?", let the community pile on in comments.
Why it matters for distribution: This is the highest-engagement archetype (366-128 comments per post). If your tool addresses something vibe coders commonly get wrong (security, error handling, testing, architecture), you can frame your launch as "I built this because I kept seeing vibe coders do X." The framing earns pre-baked goodwill from the senior-dev core of the sub.
Archetype 3: The Enterprise RAG War Story (Ceiling: ~955; Floor: ~200)
Examples:
- "I built RAG for a rocket research company: 125K docs" (955, 151 comments, TEXT) – Low_Acanthisitta7686
- "I Built RAG Systems for Enterprises (20K+ Docs). Learning path" (866, 123 comments)
- "Building RAG systems at enterprise scale (20K+ docs)" (805, 145 comments)
- "I built RAG for 10K+ NASA docs (1950s–present) in 2 weeks" (284, 50 comments)
- "Multi-modal RAG at scale: Processing 200K+ documents (pharma/finance/aerospace)" (218, 40 comments)
- "7 months of Qwen in production enterprise" (243, 81 comments)
- "EpsteinFiles-RAG: Building a RAG Pipeline on 2M+ Pages" (216, 33 comments)
The pattern: 2,000-5,000 word first-person TEXT post. Leads with a specific domain (pharma, aerospace, legal, finance) and a specific scale (20K / 125K / 200K docs). Mentions hard constraints: air-gapped, compliance, on-prem, concurrent users, H100 budget. Catalogs what failed before what worked. Names specific tools (Qwen, Docling, Qdrant, vLLM, LangGraph). Includes at least one "here's the weird hacky thing that actually worked" passage. Ends with offering to answer questions. Often has a "I used Claude for grammar/formatting polish" disclaimer at the bottom (transparency norm).
Why it matters for distribution: This is the single best archetype for credibility-building on the sub. Score ceiling is lower than memes but average comment count is higher (40-165), engagement is substantive, and readers become followers. Low_Acanthisitta7686 has built a personal brand as "the RAG guy" across ~6 posts. If you have genuine production experience, this is the highest-ROI archetype for turning readers into users/clients/followers.
Archetype 4: The Framework Slaughter (Ceiling: ~210; Floor: ~85)
Examples:
- "I have written the same AI agent in 9 different python frameworks, here are my impressions" (211, 52 comments, 0.99)
- "Not using Langchain ever !!!" (184, 60 comments, 0.97)
- "What is currently the best production ready LLM framework?" (144, 57 comments)
- "After months on Cursor, I just switched back to VS Code" (88, 28 comments)
- "I realized why multi-agent LLM fails after building one" (156, 50 comments)
The pattern: A single developer does real work across N frameworks/tools, then posts honest comparative impressions. Ratios are near-perfect (0.97-0.99) because the format is inherently fair. Comments pile in with people adding their own takes.
Why it matters for distribution: If your product competes with a framework, DO NOT post "why my framework is better." Instead, write a neutral N-way comparison that happens to include yours. The community rewards the comparison; your tool becomes visible without the launch posture.
Archetype 5: The Cost/Observability Rant (Ceiling: ~726; Floor: ~50)
Examples:
- "Why the heck is LLM observation and management tools so expensive?" (726, 88 comments)
- "How do large AI apps manage LLM costs at scale?" (26, 48 comments)
- "Are AI eval tools worth it or should we build in house?" (13, 15 comments)
The pattern: A developer names prices publicly (Langfuse $100/mo, Honeyhive, Promptlayer $50 for 100k requests, etc.), calls them "fuck you" expensive, and asks for alternatives. The rant tone actually helps – "Why the heck is..." hit 0.98 ratio and 88 comments.
Why it matters for distribution: If your tool is a cheaper alternative to an expensive incumbent category (observability, eval, prompt management, vector DB), this is your distribution vehicle. Pose as a frustrated dev, list real competitor pricing, ask for help. Only drop your tool in comments (if at all) after the thread is well-established. Posting your tool as the opening post almost always underperforms.
Archetype 6: The "I Built a Weird Thing" Curiosity Post (Ceiling: ~633; Floor: ~50)
Examples:
- "I accidentally built a vector database using video compression" (633, 80 comments) – encoded docs as QR code video frames, 1.4GB for 10K PDFs
- "I can't stop 'doomscrolling' Google maps so I built an AI that researches everywhere on Earth" (240, 50 comments)
- "I trapped an LLM into an art installation and made it question its own existence endlessly" (87)
- "I built a tiny LLM from scratch that talks like a fish" (66)
- "I ran Claude Code in a self-learning loop until it translated our entire Python repo to TypeScript" (213, 32 comments)
The pattern: One-line title that describes a genuinely weird or stunt-like use of LLMs. The weirder, the better. Builder personally did something unusual. Often open-source. Comments are a mix of "how did you do this?" and "this is nonsense / brilliant."
Why it matters for distribution: The curiosity hook is the best way to get a tool launch past the sub's skepticism filter. "I built [practical tool]" underperforms; "I accidentally built [weird thing that happens to be useful]" performs. Rule of thumb: if your title doesn't sound like a stunt, rewrite it until it does.
Archetype 7: The Honest Failure Post (Ceiling: ~614; Floor: ~25)
Examples:
- "It feels like most AI projects at work are failing and nobody talks about it" (614, 148 comments, 0.97)
- "Our 'AI-first' strategy has turned into 'every team picks their own AI stack' chaos" (15)
- "We hired 'AI Engineers' before. It didn't go well" (13, 38 comments)
The pattern: Low-ego first-person admission that the current state of AI-at-work is messier than the marketing. Specific details. No product pitch.
Why it matters for distribution: This archetype doesn't directly sell anything, but it's the highest-trust post format on the sub. Use it to establish you're "one of us" before launching anything. A Tuesday honest-failure post followed by a Friday tool launch under the same username outperforms a cold launch by a wide margin.
6. Format Analysis
| Format | Top 25 | Top 50 | Full Dataset | Notes |
|---|---|---|---|---|
| IMAGE | 14 (56%) | 22 (44%) | ~90 (29%) | Dominates top tier – meme-driven |
| TEXT | 7 (28%) | 19 (38%) | ~160 (52%) | The credibility format |
| GALLERY | 2 (8%) | 4 (8%) | ~10 (3%) | Mostly screenshot sequences |
| VIDEO | 1 (4%) | 3 (6%) | ~18 (6%) | Underused, high upside |
| LINK | 1 (4%) | 2 (4%) | ~12 (4%) | Blog posts / GitHub links |
Key pattern: IMAGE dominates the very top (top 10: 7 image, 2 gallery/video, 1 text), but TEXT dominates the substantive middle tier (ranks 10-50 are majority TEXT). The bimodal distribution matches the two-population community: memes vs. writeups.
What Format to Use For What
- Tool/app launches → TEXT with detailed writeup. Image-only tool launches score <100. If you must use an image, pair it with a detailed text explanation in the body.
- Production war stories → TEXT, 2,000+ words, first-person, specific. This is the archetype 3 playbook.
- Framework comparisons → TEXT with markdown tables. "I have written the same AI agent in 9 different python frameworks" (211, 0.99) is the template.
- Benchmarks / metrics → TEXT with tables, or IMAGE of a clean chart ("Which Format is Best for Passing Tables of Data to LLMs?" – 168, 0.97, IMAGE with a single table).
- Memes / cultural observation → IMAGE. A tweet screenshot or chart meme. Keep titles short.
- Demos → VIDEO (archetype 6 curiosity posts benefit from video). "Qwen is insane (testing a real-time personal trainer)" (189, VIDEO) and "DeepSeek R1 671B on Apple M2" (2,311, VIDEO) show video can clear the normal tool-launch ceiling when it shows something visibly impressive.
- Dev tool/landscape maps → IMAGE with legend ("AI Coding Agent Dev Tools Landscape 2026" – 403). These are evergreen and get reposted by the same author over time.
What Makes a Good Demo Video on r/LLMDevs
VIDEO is underused (6% of dataset) but has high upside when it shows something technically impossible-looking. From analyzing top video posts:
- Show the absurd thing working – "671B model on an M2 Ultra", "real-time personal trainer", "drawio live editor". The video's job is to answer "wait, this actually works?"
- No narration required – Top videos are screen recordings with on-screen text, not talking heads. Viewers are muted scrolling.
- Under 60 seconds – Assume 3-5 seconds of attention. Front-load the hook.
- Show the input → output gap – Cut from "this is what I typed" to "this is what it produced" quickly.
- End with a title card or URL if open source. Low_Acanthisitta7686's style of putting the GitHub link in a pinned comment also works.
7. Flair/Category Strategy
Raw Performance Ranking
- Discussion – Highest average score, widest usage, and highest ceiling (4,844). Default choice for 80% of posts.
- News – Highest average-per-post at the top tier (~155). Use for genuinely newsworthy external events (model releases, corporate drama, leaks). Do not use News for your own product.
- Resource – Safe, ratio-positive, but lower ceiling (~879 peak). Best for guides, curated lists, and technical writeups that link to external content.
- Tools – The official "I made a tool" flair. Technically correct but performance is depressed (~70 avg, peak 633). Community skepticism of tool launches is priced in here.
- Great Resource 🚀 – Similar to Resource but feels more aspirational; often applied to lists-of-lists and landscape maps. Peaks at 403.
- Help Wanted – For questions. 322 peak on a "help me with my open-source project" post that was actually a soft-launch.
- Great Discussion 💭 – Counterintuitively the lowest-ratio flair (0.86 avg). Avoid this flair unless you want to mark your post as provocative.
Distribution Utility Ranking (different from raw)
For someone trying to promote a product, distribution utility differs from raw score:
- Discussion – Best utility. Gives you credibility-by-association with the top tier. Use for archetypes 3, 4, 5, and 7.
- Resource – Second-best utility for launches. Frame your tool as a learning resource, not a product. "Here's how I built X and what I learned" beats "Check out X."
- Tools – Paradoxically the worst utility flair despite being the "correct" one. Using Tools flair signals "this is an ad" to the most skeptical readers. Only use if your post is also overtly archetype 6 (weird curiosity) or if you're a known recurring builder.
- Help Wanted – Under-rated utility. A genuine "I built this, what am I doing wrong?" post earns more comment engagement than a launch post. You get the tool exposure without the promotional cost.
- News – Do NOT use for your own product. Reserve strictly for third-party events.
- Great Discussion 💭 – Avoid.
Pricing Model Hierarchy (most to least community-friendly)
Based on Rule 5 (must be FOSS OR source-available + mod approval + disclaimer) and observed reactions:
- MIT/Apache-licensed open source – Highest tier. "EpsteinFiles-RAG" explicitly states "Python, MIT Licensed, open source" as a selling point. Unsloth's yoracale posts are in this tier.
- Open-core with no paywalled features – Rule 5 explicitly permits this: "The free version must be functionally identical to any other version — no locked features behind a paywall / commercial / 'pro' license." This is strict – if your paid tier has any extra features, you're out.
- Free tier with clear paid tier (third-party service) – Tolerated if disclosed upfront. "Free Model List (API Keys)" (182) is a good example of a service comparison that was welcomed.
- Source-available + clear disclaimer + mod approval – Rule 5 path for commercial products. Rarely seen in the top 100; requires proactive mod contact.
- Closed-source commercial SaaS – Effectively banned by Rule 5. Posts masquerading as discussion will be removed and the user may be permanently banned under Rule 10.
There is no title-prefix tag culture on r/LLMDevs. You will not see [OS], [FREE], [Giveaway], or [Updated] in top posts. Don't use them. The lack of such tagging is itself part of the sub's norms.
8. Title Engineering
Deconstructing the top 10 titles:
- "Soo Truee!" (4,844) – Pure meme title. Image does all the work. Punctuation used as vibe signal.
- "deepseek is a side project" (2,621) – All-lowercase, deadpan. Plays the "casual throwaway" register that lands on the meme crowd.
- "It's a free real estate from so called 'vibe coders'" (2,493) – In-joke reference ("free real estate" meme) + scare quotes as editorial commentary.
- "DeepSeek R1 671B parameter model (404GB total) running on Apple M2 (2 M2 Ultras) flawlessly" (2,311) – Technical spec porn. Specific numbers, specific hardware, "flawlessly" as the payoff word.
- "On to the next one 🤣" (1,761) – Context-free in-joke title with emoji. Depends entirely on the gallery content.
- "State of OpenAI & Microsoft: Yesterday vs Today" (1,653) – Before/after framing + corporate-hypocrisy tease.
- "Doctor vibe coding app under £75 alone in 5 days" (1,439) – Specific profession + currency + time. Implied concern about healthcare data.
- "Vibe coding from a computer scientist's lens:" (1,207) – Expert credential + topic. Trailing colon implies "here's the take."
- "Everything is a wrapper" (1,192) – Absolute statement that nihilists and builders both nod at.
- "I built RAG for a rocket research company: 125K docs (1970s-present), vision models for rocket diagrams. Lessons from the technical challenges" (955) – Domain specificity + scale + time span + technical hook + "lessons" payoff.
Title Formulas
Formula 1: "I built [system] for [specific domain] at [specific scale]: [lessons / war stories]"
- "I built RAG for a rocket research company: 125K docs" (955)
- "I Built RAG Systems for Enterprises (20K+ Docs). Here's the learning path I wish I had" (866)
- "Building RAG systems at enterprise scale (20K+ docs): lessons from 10+ enterprise implementations" (805)
- "I built RAG for 10K+ NASA docs (1950s–present) in 2 weeks" (284)
Formula 2: "[Weird/absurd] + verb + [normal technical goal]"
- "I accidentally built a vector database using video compression" (633)
- "I ran Claude Code in a self-learning loop until it successfully translated our entire Python repo to TypeScript" (213)
- "I trapped an LLM into an art installation and made it question its own existence endlessly" (87)
Formula 3: "Why the heck is [category] so [problem]?"
- "Why the heck is LLM observation and management tools so expensive?" (726)
- Implicit in "Not using Langchain ever !!!" (184)
- "Why we ditched embeddings for knowledge graphs (and why chunking is fundamentally broken)" (195)
Formula 4: "[N] months of [tech] in [setting]: what actually works (and what doesn't)"
- "7 months of Qwen in production enterprise: what actually works (and what doesn't)" (243)
- "What I learned running an Always-on AI Agent in production for months (10 lessons)" (24)
Formula 5: The low-case throwaway meme title
- "deepseek is a side project"
- "vibe coding:"
- "every ai app today"
- "its funny cuz its true"
- "Like fr 😅"
- "you do what you gotta do"
This formula is for memes only, never for tool launches or text posts. The lowercase is a deliberate "I didn't try hard" signal.
Formula 6: "I have written / tested / compared [N] [things] – here are my impressions"
- "I have written the same AI agent in 9 different python frameworks, here are my impressions" (211)
- "I Built 3 Apps with DeepSeek, OpenAI o1, and Gemini - Here's What Performed Best" (241)
- "Which Format is Best for Passing Tables of Data to LLMs?" (168)
Title Anti-Patterns (community-specific)
- "Check out [tool]" / "[Tool] just launched" – Almost no top-100 post has a generic launch title. Rewrite as an archetype.
- Star count / download count in the title – "My open-source project on different RAG techniques just hit 20K stars on GitHub" only scored 86. Growth metrics read as vanity and are ignored.
- Rocket emoji / fire emoji abuse – Minimal fire emoji usage in top posts. When emojis appear, it's usually 🤣, 🤨, 💭 – skeptical/comedic, not promotional.
- All-caps / sales-y titles – "Free Model List (API Keys)" is fine. "AMAZING NEW FRAMEWORK!!!" does not exist in the top 100.
- "Unpopular opinion:" – "Unpopular opinion: prompt engineering is just 'knowing how to talk to your coworker' rebranded" only scored 115 with 0.93 ratio. This framing reads as karma-farming.
- Vague questions from beginners – "Where to start from step 0" (3), "Am I not using LLM efficient enough?" (6). These get triaged to single digits. Rule 6 (No low-effort posts) is enforced socially.
- Day N of doing X – "Day 10 of showing reality of SaaS AI product" (2), "Day 15..." (1). Accountability-content style is explicitly rejected.
9. Engagement Patterns
Comments-to-upvote ratios by archetype:
| Archetype | Avg C/U Ratio | Interpretation |
|---|---|---|
| Vibe coding mockery | 0.15-0.25 | Discussion-heavy, people pile on |
| Honest failure post | 0.20-0.30 | Deep comment threads about shared pain |
| Enterprise RAG war story | 0.10-0.20 | Technical Q&A in comments |
| Framework slaughter | 0.20-0.30 | "Have you tried X?" additions |
| Cost/observability rant | 0.08-0.15 | Alternatives get suggested |
| Dev-insider meme | 0.01-0.08 | Passive upvotes, little discussion |
| Curiosity stunt post | 0.10-0.15 | "How did you do this?" questions |
| News (third-party) | 0.02-0.05 | Pure passive upvotes |
| Tool launch | 0.05-0.15 | Mostly skeptical comments |
| Help Wanted | 0.25-0.50 | People answer, even bad questions |
Conditional recommendation:
- If your goal is VISIBILITY, post a dev-insider meme or a News/leak item. Ceiling 2,500-4,800, passive scrolling, but maximum eyeball count.
- If your goal is RELATIONSHIPS and discussion, post an enterprise RAG war story, a framework comparison, or an honest failure post. Ceiling is lower (200-950) but comment counts and follower-gain are much higher.
- If your goal is DIRECT PRODUCT CLICKS, the archetype 6 "weird curiosity" post is the only format that reliably converts. "I accidentally built X" beats "I built X" by a large margin.
Highest-discussion topics (most comments regardless of score)
- Vibe coding and whether it's valid – "Doctor vibe coding" (366), "Vibe coding from a computer scientist's lens" (147), "CEO of Klarna claiming they are replacing Jira with a vibe coded app" (93), "power of coding LLM in 20+y dev" (172)
- Deepseek / Chinese lab drama – "DeepSee again" (261), "RAG is dead" (113), "deepseek is a side project" (87), "OpenAI calls for bans on DeepSeek" (44)
- RAG architecture debates – "Why we ditched embeddings for knowledge graphs" (83), "If RAG is dead, what will replace it?" (113), "Is RAG really necessary for LLM → SQL systems" (100), "50-100 PDFs, what RAG approach?" (95)
- LLMs killing critical thinking / mental health – "The more I learn about LLMs, I get genuinely upset" (247), "It feels like most AI projects at work are failing" (148)
- Model selection / framework fatigue – "What is currently the best production ready LLM framework?" (57), "9 frameworks impressions" (52), "Not using Langchain ever" (60)
10. What Gets Downvoted
Posts with ratios below 0.85 in the dataset:
| Title | Score | Ratio | Why it struggled |
|---|---|---|---|
| CEO of Klarna claiming they are replacing Jira with a vibe coded app | 88 | 0.78 | Perceived as CEO-worship; community distrusts the claim |
| Mythos is Opus 4.7… | 83 | 0.78 | Cryptic product-tease title with no substance |
| We beat Google Deepmind but got killed by a chinese lab | 79 | 0.80 | Self-aggrandizing framing, suspected promo |
| AI won't replace devs — but devs who master AI will replace the rest | 216 | 0.82 | Clichéd "hustle dev" framing |
| Found an open-source goldmine! | 186 | 0.83 | Vague clickbait title, felt like an ad |
| The more I learn about LLMs, I get genuinely upset at how most use AI | 255 | 0.83 | Moralizing tone, comes across as preachy |
| I stand by this | 188 | 0.83 | Low-context image post, unclear point |
| Carnegie Mellon just dropped one of the most important AI agent papers | 178 | 0.83 | Hyperbolic title ("most important") |
| Only this LLM books you need | 272 | 0.89 | Affiliate-adjacent reading list |
| How AI is transforming senior engineers into code monkeys | 187 | 0.90 | Controversial-by-design framing |
| Are LLMs Models Collapsing? | 408 | 0.85 | AI-slop-feeling writeup |
| Meta can now predict what your brain is thinking | 134 | 0.86 | Clickbait rewrite of a research story |
| NVIDIA dropped one of The most important AI paper of 2025 | 313 | 0.88 | Hyperbolic title ("most important") |
| It's DeepSee again | 645 | 0.87 | Pro-Deepseek-fatigue sentiment friction |
Ratio Tiers
- Above 0.94 – Universally well-received. Safe for technical writeups, memes, and honest failures.
- 0.85-0.94 – Net positive but with friction. Often provocative-by-design (controversial opinions, contested framings, viral memes). Expect pile-on comments.
- Below 0.85 – Controversial or community-hostile. Usually triggers one or more of: perceived promotion, hyperbolic claims, CEO/celebrity worship, moralizing, or suspicion of being karma-farmed/AI-generated.
Anti-Patterns (community-specific)
- "Most important [X] of [year]" hyperbole – The community distrusts superlatives. "NVIDIA dropped one of The most important AI paper of 2025" (313, 0.88), "Carnegie Mellon...most important AI agent papers" (178, 0.83).
- Suspected promotion flagged as "discussion" – Rule 10 says this earns a permanent ban. Posts that look like soft launches get rated down even if they survive mod removal.
- Moralizing about "how people use AI" – The "upset at how people use AI" tone reads as condescending. Technical criticism is welcome; value judgments about users are not.
- CEO / founder worship – The Klarna CEO post (0.78) and variants that take corporate PR at face value get punished. The community treats CEO claims as marketing until proven otherwise.
- Day N accountability content – "Day 10 of showing reality of SaaS AI product" (2 score). This is treated as spam regardless of content.
- Clickbait headlines on research papers – "Meta can now predict what your brain is thinking. read that again." (134, 0.86). The community prefers papers linked directly with neutral framing.
- AI-generated post tells – The rocket RAG post (955) defused this by disclosing "I used Claude for grammar/formatting polish." Non-disclosed AI-feeling posts (long, overly structured, perfect grammar, generic insights) cluster in the 0.82-0.89 friction band.
There is no public "hall of shame" or blacklist that I can detect. Enforcement is via: (1) mod removal for Rule 5/10 violations, (2) community downvote friction for the above patterns, (3) commenter callouts ("is this an ad?"). The permanent-ban threat in Rule 10 appears to be taken seriously based on the low volume of obvious astroturf in the top 100.
11. The Distribution Playbook
Phase 1: Pre-launch (Week -4 to 0)
Goal: Establish a posting identity the community recognizes before you ever mention your product.
- Read Rule 5 and Rule 10 carefully. If your product is not FOSS-licensed with free version identical to paid, contact mods first or do not post at all. The permanent ban is real.
- Post 2-3 low-stakes contribution posts first, spaced across 2-4 weeks. Good options:
- An honest-failure observation (archetype 7)
- A framework comparison you actually did (archetype 4)
- A technical rant about pricing in a category adjacent to yours (archetype 5)
- A curiosity/stunt post using LLMs (archetype 6)
- Comment substantively on other RAG/framework threads. The top comment-earners in the community (Low_Acanthisitta7686, Nir777, yoracale, Arindam_200) are recognized because they show up in comments, not just posts.
- Avoid the "Tools" flair for these warmup posts. Use Discussion or Resource.
Phase 2: Launch Day
Optimal launch formula (for FOSS tools):
- Format: TEXT, 800-2,500 words
- Flair: Discussion (NOT Tools) if your post is a writeup. Resource if it's a curated list or guide.
- Title: Use Formula 1 ("I built X for Y at Z scale: lessons") OR Formula 2 ("I accidentally built..."). Avoid generic launch phrasing.
- Opening paragraph: Who you are, what problem you hit, why existing tools didn't work. Be specific about the problem.
- Middle: The actual technical content. What failed first. What you tried. What worked. Include at least one "this is the weird hack that actually worked" moment.
- GitHub link: In the post body, mid-paragraph, not in the title.
- Closing: Offer to answer questions. Mention your license explicitly if MIT/Apache.
- Disclosure footer: If you used an LLM to polish grammar, say so ("Note: I used Claude for grammar/formatting polish" is the community-accepted template).
Timing: r/LLMDevs has global audience; US weekday mornings (13:00-17:00 UTC) work well based on the top posts' timestamps (rocket RAG 16:02 UTC, enterprise RAG writeups mostly 13:00-18:00 UTC). Avoid weekends – engagement drops ~40%.
Phase 3: First 24-48 Hours
Comment strategy – pre-written reply templates for the 4-5 most common objections:
-
"Is this vibe coded?" / "How much of this was written by AI?"
"The architecture and all technical decisions are mine. I did use [Claude/ChatGPT] for [grammar polish / boilerplate / docs]. The [core algorithm / retrieval logic / X] is hand-written. Happy to walk through any specific file."
-
"Why not just use LangChain/LlamaIndex/LangGraph?"
"I did try [framework] first. The specific thing that broke for me was [concrete issue]. For [simple case] LangChain is probably fine, but for [my use case] I ended up needing [specific control]. Here's the commit where I ripped it out: [link]."
-
"What's your pricing model?" / "Is this going to get paywalled?"
"MIT licensed, will stay MIT licensed. There is no paid tier planned. If I ever add hosting, the self-hosted version will remain functionally identical — that's a commitment I'll put in the repo."
-
"Does this actually work at scale?"
"Honest answer: I've tested it on [X documents / Y concurrent users / Z tokens/sec]. Above that, I haven't. If you hit the wall at larger scale, open an issue and I'll look at it with you."
-
"How does this compare to [competitor]?"
"[Competitor] does [X] better. Mine does [Y] better. The main reason to use mine is [specific thing]. If you don't need [that thing], use [competitor] – no hard feelings."
Monitor your ratio in the first 4 hours:
- Ratio > 0.94 + rising score: You're on track. Keep answering comments promptly.
- Ratio 0.85-0.94: Friction. Someone accused you of promotion or a technical claim is being challenged. Address the specific challenge in a top-level edit.
- Ratio < 0.85: Community is hostile. Do NOT edit the post to argue. Engage the top critic respectfully in comments. Consider that the post may need to be deleted and relaunched with different framing.
- Ratio > 0.94 but score flat after 4 hours: Not controversial but not viral. Content is probably fine; positioning failed. This is the most common outcome for tool launches. Learn and try a different archetype next time.
Phase 4: Ongoing Presence
Repeat cadence: Low_Acanthisitta7686 posts roughly every 2-6 weeks. yoracale posts every 3-8 weeks. Arindam_200 posts every 1-3 weeks. These are the cadences that produce a recognized identity without triggering "spam" callouts.
Building on a hit: If a launch post breaks 400, your next post should be a follow-up with NEW technical content (a deep-dive on one sub-problem from the original), not a v1.1 announcement. Version bump posts underperform.
Stealth distribution tactics:
- Answering questions in "Help Wanted" threads – The community has a constant stream of "how do I RAG 50 PDFs" / "which framework should I use" posts. Answer them substantively with a section mentioning your tool only if it's actually relevant. This earns more clicks than a launch post.
- Contributing to framework comparison posts – When someone posts "I tried 9 frameworks", comment with an 10th that happens to be yours.
- The observability rant inclusion – Posts like "Why is LLM observation so expensive?" list specific competitor pricing. A comment that lists your open-source alternative is welcomed.
- Crosslinking your own GitHub in a technical writeup – If your archetype 3 writeup naturally mentions a tool you built, include the link in-line, not as a CTA. Readers click through more when it feels incidental.
- Show up in comments of the RAG-guy posts – Low_Acanthisitta7686's comment threads are full of builders. Commenting substantively there (not pitching) gets your username in front of the core technical audience.
Score-Tier Calibration
Tell yourself the truth before posting:
- Tool launches on r/LLMDevs almost never exceed 500. If you need more visibility, you need a different archetype (meme, RAG writeup, or curiosity post).
- Enterprise RAG writeups realistically hit 200-950. If you don't have production experience, don't try this archetype – commenters will see through it.
- Memes can hit 4,000+, but require developer-insider cultural fluency and deliver zero direct product clicks.
- Framework comparisons hit 150-250 almost regardless of author, if the comparison is honest.
- Pure help posts cluster at 1-30. The "Help Wanted" flair is for getting comments, not upvotes.
Post-Publication Measurement
- 4 hours: If score < 20 and ratio < 0.90, the post has failed. Do not delete – a deleted post can look suspicious. Learn and try again in 2-4 weeks.
- 24 hours: Final score is roughly 70% set. A post at 150 after 24 hours will end at 200-250. A post at 500 after 24 hours will end at 700-900.
- 7 days: If the post is in a recurring-author pattern (e.g., your third writeup), watch whether the score exceeds the prior post. If flat or declining, consider whether your archetype is saturating.
- Clickthroughs vs upvotes: Expect roughly 2-5% clickthrough on GitHub links in text posts (much lower than r/LocalLLaMA's ~10%). This community reads and argues more than it clicks.
12. Applying This to Any Project
Quick-reference checklist (read before posting)
- My project is MIT/Apache or FOSS-licensed with free version functionally identical to any paid version (Rule 5)
- My post is not disguised advertising; I am not violating Rule 10
- I am using Discussion or Resource flair, not Tools (unless Tools is the clearest signal)
- My title is specific and uses Formula 1, 2, 3, 4, or 6 – not a generic launch phrase
- I have 800+ words of actual technical substance in the body
- My GitHub/product link is in-body, not in the title
- I have a pre-written reply template for "is this vibe coded?" and "why not use LangChain?"
- I disclosed any LLM-assisted writing ("used Claude for grammar polish")
- I have posted at least one non-promotional contribution in the past 2 weeks under the same username
- I am launching on a US weekday morning (13:00-17:00 UTC)
- My title contains zero "most important", zero all-caps, zero star/download counts
- I have my pricing model explicitly stated (or can answer it in comments within 5 minutes of posting)
Scenario-Based Launch Guides
Scenario 1: Your product is free and open-source (MIT/Apache)
- Optimal launch formula: Archetype 3 (production war story) OR Archetype 6 (weird curiosity) + Discussion flair
- Title template: "I built [specific tool] for [specific domain]: [lessons / scale / weird thing that worked]"
- Key risk: Over-claiming scale or production use when you haven't actually deployed it. The senior engineers in comments will ask follow-up questions that expose the gap.
Scenario 2: Your product uses one-time/lifetime pricing
- Optimal launch formula: Do NOT launch on r/LLMDevs. Rule 5 explicitly requires FOSS or (source-available + mod approval + disclaimer). Post on r/LocalLLaMA (friendlier to lifetime pricing) or r/macapps (where lifetime pricing is celebrated) instead.
- Key risk: Permanent ban under Rule 10 if you frame a commercial product as "discussion."
Scenario 3: Your product uses subscription pricing
- Optimal launch formula: Not on r/LLMDevs. See Scenario 2.
- If you must engage here: The only safe format is a "cost rant" post (archetype 5) where you enumerate competitor pricing including your own, framed neutrally as a category map. Do not editorialize that your tool is the answer – let readers find it.
- Key risk: Subscription-priced product launches are the single most common cause of permanent bans on this sub.
Scenario 4: Your product was built with AI (vibe coded)
- Optimal launch formula: Lead with the AI-assisted-ness as a confession, not a feature. "I used Claude Code to build X. Here's what it got wrong and how I had to intervene." Archetype 3 crossed with archetype 7 (honest failure).
- Title template: "I ran Claude Code in a loop until it [specific unreasonable task]. Here's what broke." (The "in a loop until it..." post scored 213, 0.92 ratio.)
- Key risk: Hiding AI assistance and being exposed in comments. Always disclose upfront.
Scenario 5: Your product is a RAG/agent framework competing with LangChain/LlamaIndex
- Optimal launch formula: Archetype 4 (framework slaughter) with your framework as one option among many. Do a genuine side-by-side with code, not a pitch.
- Title template: "I wrote the same [task] in [N] frameworks: here's what actually works"
- Key risk: Readers smell self-promotion if yours comes out suspiciously best. The "9 frameworks" post (211, 0.99) worked because it included "No framework" as the honest best-for-beginners answer.
Cross-Posting Guidance
Use the table below when reusing content across subs:
- On r/LocalLLaMA: Frame as "runs on hardware you own." Emphasize open weights, quantization, VRAM. GPU rig photos help.
- On r/ClaudeAI: Frame as "I built this with Claude Code." Emphasize the agentic workflow you used.
- On r/AI_Agents: Frame as "multi-agent system with [specific orchestration pattern]." Emphasize the diagram.
- On r/MachineLearning: Frame as methodology + paper citations. Emphasize reproducibility and benchmarks.
- On r/LLMDevs: Frame as a production war story OR a framework comparison OR a dev-insider observation. Emphasize what broke, what scale, what took 40% of your time. The "this is what nobody warns you about" frame is the highest-trust positioning this community offers.
The same content can generally be reframed 3-4 ways across these subs without issue, provided the title and opening paragraph are rewritten each time and the posts are spaced 3+ days apart.