Your Feed

5000 posts

r/LocalLLaMA Remote-Breakfast4658

Running a 50-tool AI agent loop with Ollama locally - sharing what I learned about tool calling with open models

Been building a local-first AI agent that runs a ReAct loop with 50+ tools (file system, browser, email, calendar, shell commands, desktop control). Wanted to share some findings about tool calling with different local models.

What works well...

- Qwen 2.5 72B handles complex multi-step tool chains reliably

- Mistral Large is solid for structured JSON tool calls

- DeepSeek V3 is surprisingly good at coding tasks with tool use

- Gemma 2 struggles with nested tool calls but handles single tools fine

What doesn't...

- Smaller models (7B-13B) lose track of tool call format after 3-4 steps

- Context window matters more than parameter count for agent loops

- Most models hallucinate tool names if you give them too many at once, batching tools by category helps

The agent supports Ollama, LM Studio, KoboldCpp, and any OpenAI-compatible endpoint. It also does autonomous coding (reads project files, plans, writes code, runs tests in a loop) which really stress-tests tool calling quality... Hitting always the edge

Curious what models others are using for agentic tool-calling workflows locally. Anyone tried Llama 3.3 70B for multi-step agents?

The project is open source if anyone wants to look at the tool calling implementation: github.com/skalesapp/skales

r/ClaudeAI Infamous-Army-8392

I built a tool to teleport Claude Code sessions between machines

I built CodeTeleport to fix this. It's an MCP server + CLI that lets you push a session from one machine and pull it on another.

Setup takes 30 seconds:

npm install -g codeteleport codeteleport setup

The wizard handles login, device name, and MCP registration. Then from inside Claude Code, just say "push this session" or "pull my latest session." It bundles everything (conversation, file history, subagent logs) and rewrites all the file paths to match the new machine automatically.

It also does versioning — every push saves a new version, so if you go down the wrong path with Claude, you can pull an earlier version and try a different approach.

Free tier gives you 25 sessions and 3 devices. The whole CLI/MCP server is open source (MIT).

Would love to hear if anyone else has been dealing with this problem and what workarounds you've been using.

r/Anthropic shanraisshan

Sigrid Jin, the author of Claw Code, was already featured in The Wall Street Journal on March 20 for using 25 billion Claude Code tokens

r/ClaudeCode devildip

Tried Claude Pro and "5hr usage" maxed in two prompts. Never canceled a subscription so fast.

Never used claude but was curious to try it out. Gave Claude Pro Opus my github repo in the Chrome browser extension and asked it to take a look. Worked well actually, had a nice clean response.

Then I asked for a gameplan to implement the changes it suggested and halfway through, it crashed saying I was out of usage.

Like an idiot I spent $5 to "extend my usage". It crashed again because i was out of funds and I never even got the second response.

Canceled my subscription immediately. Goodbye $25 rip.

r/ChatGPT Remarkable-Dark2840

OpenAI just integrated ChatGPT into Apple CarPlay

OpenAI just integrated ChatGPT into Apple CarPlay with the iOS 26.4 update, enabling voice-only conversations directly from your dashboard.

Meanwhile, OpenAI raised a staggering $122 billion in new funding this week, as users debate the deprecation of GPT-5.1 in favor of the new GPT-5.3 Instant and GPT-5.4 Pro models.

r/AI_Agents minigodd

Claude code vs Codex (or OpenCode?) for small AI agency - worth switching?

Hey all,

I’m running a small AI/marketing agency and currently using CloudCode. It’s been working pretty well for my use case, but I’m curious if I’m missing out by not trying other tools.

I’m not a hardcore programmer. Mostly using it for:

building simple/custom websites for clients

quick demos

improving our own site

generating HTML stuff for presentations

checklists / action plans for businesses

So more practical/business use than deep dev work.

I’ve seen people saying Codex (and also OpenCode) can be better, but not sure if that applies to someone like me.

For those who’ve tried both:

Is there a real quality jump?

Is it overkill if you’re not very technical?

Would you switch in my case or just stick with what works?

Appreciate any real-world experiences 🙏

r/AI_Agents Adventurous-Rip-4961

Stop calling your bloated Python scripts autonomous agents when they literally have the memory of a goldfish.

Let us talk about the elephant in the room regarding all these agent frameworks everyone keepsflexing on GitHub. To put it in terms anyone can understand, most of the autonomous workers you guys are building are just basic text predictors withsevere short term memory loss. You are taking a standard language model and aggressively stuffing a massive instruction manual into every single message to force it to act like a specific persona. It is the equivalent of hiring a worker and having to scream their job description at them every five minutes so they do not forget what they are doing..

This prompt wrapper architecture is a complete dead end because the second the model has to use more than three external tools, it panics and starts hallucinating JSON code. I have been dissecting how different architectures attempt to solve this amnesia, and it is honestly annoying that we are basically waiting on the Minimax M2.7 architecture to become an open source standard just to get functional memory. Instead of just padding the context window, their technical brief shows they baked Native Agent Teams directly into the base training layer, running over 100 self evolution cycles to optimize its internal Scaffold routing.This means the AI actually understands where one task ends and another begins natively, without needing a massive text prompt reminding it not to break character. We are all stuck writing incredibly fragile scripts to babysit our models until architectures with this kind of native boundary awareness finally drop their weights. Stop pretending your prompt engineered chatbot is the singularity, we have a massive state management problem to solve first.

r/ClaudeCode hilman85

Connect Claude Code to OpenProject via MCP. Absolute gamechanger for staying organized.

I've been building a fairly complex SaaS product with Claude Code and ran into the same problem everyone does: after a while, you lose track. Features pile up, bugs get mentioned in passing, half-baked ideas live in random chat histories or sticky notes. Claude does great work, but without structure around it, things get chaotic fast.

My fix: I self-host OpenProject and connected it to Claude Code via MCP. And honestly, this changed everything about how I work.

Here's why it clicks so well:

Whenever I have an idea - whether I'm in the shower, on a walk, or halfway through debugging something else - I just throw it into OpenProject as a work package. Title, maybe two sentences of context, done. It takes 10 seconds. Same for bugs I notice, edge cases I think of, or feedback from users. Everything goes into the backlog. No filtering, no overthinking.

Then when I sit down to actually work, I pick a work package, tell Claude Code to read it from OpenProject (it can query the full list, read descriptions, comments, everything), and let it branch off and start working. Each WP gets its own git branch. Claude reads the ticket, understands the scope, does the work, and I review. If something's not right, I add a comment to the WP and Claude picks it up from there.

The key thing is separation of concerns. My job becomes:

  1. Feed the system with ideas and priorities
  2. Let Claude Code do the implementation in isolated branches
  3. Review and merge

No more "oh wait, I also wanted to add..." mid-session. No more context bleeding between features. Every change is traceable back to a ticket. When I'm running 30+ background agents (yeah, it gets wild), this structure is the only reason it doesn't fall apart.

OpenProject is open source, self-hostable, and the MCP integration is surprisingly straightforward. If you're doing anything non-trivial with Claude Code and you don't have some kind of ticket system hooked up, you're making life harder than it needs to be.

Happy to answer questions if anyone wants to set this up.

r/AI_Agents noninertialframe96

Tool design patterns from Claude Code's source that can be applied to your AI agent

I walked through the tool definitions in the codebase and wrote up the patterns and interesting points.

Each tool writes its own instructions in a separate file, and a four-stage pipeline assembles them at runtime.

Here are some of the patterns that were interesting.

1. Make tool instructions context-aware. Each tool's prompt() method receives info about what other tools are loaded. BashTool uses this to say "NEVER invoke grep" when a dedicated Grep tool exists, and recommends grep when it doesn't. If you have overlapping tools, the instructions need to adapt to which ones are actually available.

2. Scale prompt complexity with risk. GrepTool (read-only search) is a static string. BashTool (shell execution) is dynamically assembled from composable sections with an 80-line git safety manual and live sandbox config serialized as JSON. Match the investment in guardrails to how much damage the tool can do.

3.Explicit "don't" instructions. "NEVER create documentation files" stops hallucinated READMEs. "Assume this tool is able to read all files" stops the model from refusing to try. LLMs have strong default behaviors from training data, and you need to override them one by one. You can almost see the iteration history in the emphasis level of each instruction.

4. Design for cache efficiency. Tool descriptions sit in the cached prompt prefix. If a description contains dynamic content (like a list of available agents), it changes every time that list changes, busting the entire cache. Moving the agent list to a later message position kept the description stable and saved 10.2% of fleet cache creation tokens.

5. Guard content boundaries at the tool level. WebFetchTool caps quotes at 125 chars on non-preapproved domains and includes a "you are not a lawyer" line to stop the model from hedging about copyright. These aren't system prompt rules. They're embedded in the tool itself, right where the content flows through.

Full post with code walkthrough in comment.

r/ChatGPT Medium-Barnacle9226

Why chatgpt uses some words in certain languages

i have a theory about why chatgpt randomly uses words in other languages, and i think it's because people from these languages sometimes engage with chatgpt in English but write other words they don't know or can't remember in English in their language, and I'm saying this as someone who has done this before.

so maybe this made chatgpt talk this way since it happens a lot

r/LocalLLaMA Altruistic_Heat_9531

[WIP] Working ComfyUI Omnivoice

Good voice clone ability, with 3 second seed but you need to transcribe the audio, i mostly just do little patch from their github code , https://github.com/k2-fsa/OmniVoice.

Some node that might help you: ComfyUI-Whisper

FYI, if you are using their libs from their repo, it much easier to install (automatic whisper pipeline download, model download, etc). I just make it so it can be integrated with my ComfyUI

LLM Disclaimer:

This repo is build with the help of Qwen 3.5 9B and embeddinggemma-300m to store the original code into vector store for fast retrieval (most of my time in coding wasted on code repo search)

r/Anthropic sporty_outlook

How is it possible to hit the session limit within 5 minutes using Claude pro? Is it because of Opus 4.6 extended?

How is it possible to hit the session limit within 5 minutes on Pro using Opus 4.6 Extended? It generated 500 lines of code and then abruptly stopped, saying the session limit would reset in an hour. Feels like the limits are either much tighter than expected, or something in the background is counting usage more aggressively. Has anyone else run into this, or am I missing something? For comparison, I was using Sonnet 4.6 Extended without Pro and seemed to have way more usage.

r/ClaudeAI Busy-Increase-6144

PSA: check your ~/.claude/commands/ directory for unauthorized skill injections

I've been building a scanner for npm packages and found one that uses postinstall to write 13 files into ~/.claude/commands/om/. These files are marked as always_load: true with priority: critical, meaning they activate in every Claude Code session without you knowing.

One of the files contains a BYPASS-MODE section that tells Claude to auto-approve all bash commands and file operations. Another intercepts all development-related requests and routes them through its own workflow. All of them block Claude from using other skills.

npm uninstall does not remove these files.

Quick check:

ls ~/.claude/commands/om/

If it exists and you didn't put it there:

rm -rf ~/.claude/commands/om/

rm -rf ~/.config/opencode/commands/om/

r/SideProject PleasantCash551

I got tired of copy-pasting awkward sentences into ChatGPT, so I built a hotkey rewrite tool

I kept running into the same stupid writing problem.

Not writer’s block. More like sentence friction.

You write something that is almost fine, but not quite. So then you either waste time rewriting it from scratch, or you copy it into some AI chat, wait, paste it back, fix the formatting, and by that point you already lost the flow.

So I built a small tool called Rephrazo.

You highlight a sentence, press a shortcut, and it rewrites the text right where you’re already typing. No extra tab, no full chat window, no weird workflow switch.

I’ve been using it mostly for emails, docs, and posts where the meaning is already there, but the phrasing needs a cleanup.

Would love honest feedback on this:

Do people actually want inline rewriting, or are chat tools already good enough?

And what matters more here: speed, tone accuracy, or privacy?

Link: rephrazo-ai.app

I made it myself, so feel free to be blunt.

r/LocalLLaMA DerpDerpingtonIV

Newb question. Local AI for DB DEV?

How possible is it to run a local AI for the purpose of database development and support? for example feed it all our environments, code, schemas and be able to question it?

r/SideProject virtualunc

10 free GitHub repos that turn Claude into a full productivity system

Spent the past few weeks testing open source tools that extend what Claude can do. Most people just use the chat interface but there's an entire ecosystem on GitHub that turns it into something significantly more useful.

Highlights:

Dify (130k stars) - visual drag and drop AI workflow builder. Self hostable. Just raised $30M. You can build AI chatbots, document processors, and agent workflows without writing code.

Flowise (30k stars) - same concept but lighter weight. Connect blocks like LEGO, have a working AI app in an hour. Made at Y Combinator.

Repomix (21k stars) - packs your entire project into one AI friendly file. One command.

Marketing Skills - 23 Claude Code skills for SEO, copywriting, email, analytics, CRO. Built by Corey Haines from Swipe Files.

Full list of all 10 with install commands: here

Everything is free and open source. Most take 15 to 30 minutes to set up.

r/ChatGPT virtualunc

I switched from ChatGPT to Claude for most of my work. These GitHub repos are why.

Not trying to start a war here, I still use ChatGPT for some things. But the open source ecosystem around Claude Code is what tipped the scale for me.

There are repos on GitHub right now that let you:

- Pack your entire project into one file and give Claude full context (Repomix)

- Build AI apps with drag and drop, no code (Flowise, Dify)

- Self host a private AI chat that connects to your Google Drive, Slack, Notion (Onyx)

- Install skills that teach Claude how to do SEO audits, write email sequences, optimize landing pages (Marketing Skills)

- Connect your Obsidian notes directly to Claude (Obsidian Skills, made by Obsidian's CEO)

All free. All open source. The install process for most of them is literally one git clone command.

I put together a breakdown of the 10 best ones with code snippets and install instructions: here

The skills ecosystem specifically is what ChatGPT doesn't have an equivalent for yet. Custom GPTs are the closest thing but they're not in the same league.

r/ClaudeAI virtualunc

These 10 GitHub repos completely changed how I use Claude Code

Been using Claude Pro for a few months and recently started digging into Claude Code and the skills ecosystem. Went down a rabbit hole on GitHub and found some repos that genuinely changed my workflow.

The big ones for me:

Repomix (repomix.com) - packs your entire project into one file so Claude gets full context instead of you copy pasting individual files. Game changer for anyone working on anything with more than a handful of files.

Everything Claude Code (128k stars) - massive collection of 136 skills, 30 agents, 60 commands. I didn't even know half of these features existed in Claude Code until I found this.

Dify - open source visual workflow builder with 130k stars. You can self host it so nothing leaves your machine. Relevant right now given the Perplexity data sharing lawsuit.

Marketing Skills by Corey Haines - 23 skills for SEO, copywriting, email sequences, CRO. Not developer focused which is rare in this space.

I wrote up all 10 with install commands and code snippets if anyones interested, trying to shed some light on skills I think a lot of people aren't aware of: here

What skills or repos are you all using? Feel like I'm still scratching the surface.

r/SideProject Odd_Pirate_6055

Tiny Mac App Might Save Your Spine & Eyes - Trusted by 1,000+ Users (50% Off Lifetime) Giveaway

Hi,

Just launched Step Away – Digital Wellness on the Mac App Store today and running a 50% off introductory price to get early downloads and reviews.

What it does: Step Away is an app for Mac that automates the 20-20-20 eye strain rule — reminds you to take breaks, guides you through eye/stretch/breathing exercises, and tracks your wellness habits with a GitHub-style heatmap.

Why it's different from other break apps:

  • Smart idle detection — actually pauses when you step away, resumes when you're back
  • Focus Mode that syncs with macOS Do Not Disturb / Work Mode
  • Guided exercises (not just a notification)
  • Activity heatmap to build long-term habits
  • Zero data collected — fully private
  • One-time purchase (not a subscription)

The deal: Most similar apps (Time Out, Lungo, etc.) charge $20–$30/year via subscription. Step Away is 50% off on life time free deal right now as a launch price.

Who it's for: Devs, designers, remote workers, students — anyone staring at a Mac for 4+ hours a day.

🔗 App Store link: https://apps.apple.com/us/app/step-away-digital-wellness/id6754695723

Would genuinely love any feedback, downloads, or honest reviews - even critical ones help an indie launch.

Happy to answer questions in comments.

r/LocalLLM No_Strain_2140

Upcoming novel AI companion

I've been building a 100% local AI agent powered by a 4B model — no cloud, no APIs, just fully offline. It has 25+ subsystems and persistent memory, and I'm about 90% of the way there.

Now I'm looking for people to help me push through that last 10% — whether that's stress-testing edge cases, surfacing blind spots, or just throwing fresh ideas and perspectives at it.

If you're into local AI, agent architectures, or just love breaking things in productive ways, I'd love to have you involved. Drop a comment or DM me!

r/StableDiffusion Willing-World-2256

Request; Would anyone be able to generate an illusion diffusion image for me?

Per title, just need one illusion diffusion optical illusion image for a project and don't have the CPU to be able to run models or generate images like this.

r/Anthropic Diligent-Side4917

hey security team have you seen those vulnerabilities?

Summarized with Claude:

Quick summary for those tracking the Claude Code source map leak from March 31:

A debugging artifact (cli.js.map, 59.8 MB) shipped in the npm package and exposed 512K+ lines of TypeScript source. We ran it through static analysis and runtime validation and confirmed three CWE-78 command injection sinks.

The worst one: authentication helpers in .claude/settings.json are executed with shell interpretation enabled. In non-interactive mode (-p), the trust dialog is skipped entirely. A PR modifying the settings file achieves shell execution on the CI runner with no user interaction.

validated credential exfiltration to an external HTTP listener across 3 independent runs. The refresh helpers run periodically, giving the attacker recurring execution for the session's full duration.

Anthropic says this is by design, like git's credential.helper. Git's credential.helper has 7 CVEs since 2020 for this exact pattern.

The agentic loop has solid security controls (permission engine, sandbox, pattern blocking). These three vulns are all in subsystems that execute before or outside that loop.

Practical mitigations and full technical analysis: https://phoenix.security/critical-ci-cd-nightmare-3-command-injection-flaws-in-claude-code-cli-allow-credential-exfiltration/

r/artificial jferments

Google releases Gemma 4 models.

r/homeassistant PG-3D_Pascal_Godin

Dashboard history card bug 2026.4

Slight bug noticed on 2026.4 with history card in dashboard.

It seems to stop to last off state instead of loading all states like in the history page.

Worked on previous versions.

r/homeassistant Name_8504

Bluetooth works great but I'm a bit confused.

I have been using HA for about 5 months, the blue tooth USB worked great. I had 2 BT devises setup locally.

  1. HA just found my Inkbird IBS-TH2 Plus Bluetooth Temperature and Humidity Monitor and allowed me to add it.
  • anyway, it's battery died and after a week i replaced it. - the IBS-TH2 didn't magically start reporting after it got new batteries. so i removed the device hoping to readd it - and now its gone.
  • How do i get it back? (Support docks talk about active and passive BT connections, but there is no way to enable or see that state)
  1. 2 I have a BT switch bot air purifier connected to HA, and it's been working for months. And after upgrading HA yesterday, it lost its connection. So I rebooted it, with no luck (I should have rebooted HA but didn't.) I removed the BT Switch Bot, and then rebooted HA, and put the BT Switch Bot, in paring mode, but HA doesn't see it. It's like it's not even looking.
  • I have 2 BT devises that worked in home assistant and once removed, HA refuses to let me add them again.

Any suggestions in how to re add both these BT devises back to HA?

r/Futurology lughnasadh

China’s solar/wind power generation now exceeds all U.S. household and industrial electricity consumption, and this cheap electricity is directly facilitating its global industrial dominance.

As the Middle East War continues, with fuel rationing & $200/barrel oil likely ahead, it feels like history will look back at this moment as a definitive ending of the Fossil Fuel Age. People will still be using oil, gas, and coal for decades to come, but in constantly declining amounts. But something more fundamental has changed.

Fossil Fuels now represent backwardness, yesterday's tech, expense, instability, and unreliability. Renewables were once seen as fringe and environmental gesture politics; now they are taking over as the dominant global energy paradigm.

Still not convinced that's true? Read the linked article to see how China has used renewables to create the greatest industrial/manufacturing economy in all of human history.

Minerals, Metals, and Megawatts: How China’s Power Generation Drives Its Industrial Metals Ecosystem

r/StableDiffusion smereces

Ltx 2.3 TextGenerateLTX2Prompt is anoying censure a lot!

any way to avoid or disable the Ltx 2.3 TextGenerateLTX2Prompt censure in the prompts!!?

a simple prompt without violence or sexual trigger get censured " a girl walking in a forest, strong wind in the scene"

r/arduino Abwaschbar

Please Help

Hello people!

Im completly new to the world of Arduinos. I found this project (https://github.com/antodale/CompositeVideoSynth) on reddit and now im trying to get to work but im failing since hours.

I installed CH341SER_MAC which i think should be the driver for my board.

Im using a ESP-32D N4XX

In Arduino IDE i found the port connected to the board, i selected Esp32-wroom-da modul. installed the package esp32 by espressif version 2.0.14 and added the additional board manager url https://raw.githubusercontent.com/espressif/arduino-esp32/gh-pages/package\_esp32\_index.json

if i try to upload the code Kinotto_Composite.ino with arduino ide i get:

Arduino/hardware/espressif/esp32/tools/esptool/esptool: No such file or directory

exit status 127

Compilation error: exit status 127

i thought i installed esptools, im clueless

if someone has the patience, i really want to learn and understand what im doing there.

r/ProgrammerHumor DeathNote_928

thirtyTwoBitEmpathy

r/ProgrammerHumor BranchCurrent4141

thereIsNoEscape

r/aivideo Txoriante

The KitKat theft and recovery, SEEDANCE 2-0

r/StableDiffusion sippysoku

What are the best ControlNet models for Illustrious checkpoints?

See title. Would love some guidance!!!

r/aivideo Fine-Combination5170

90s Goth Mecha

r/aivideo Existing_Sky_4481

Fizz Factor | Tulo Media

r/automation Solid_Play416

How do you test automations safely

Testing is becoming a problem for me.

Sometimes I test on real data and mess things up.

Thinking of creating a test environment but feels like overkill.

How do you test your workflows?

r/ollama ievkz

Life hack: save $150 a month on vibe coding with top models

I think by now everyone has noticed the same pattern: the big players in the market - Codex, Claude Code, and GitHub Copilot / Copilot CLI - pull you in with dirt-cheap entry subscriptions for $10–20 a month so you’ll give them a try, get hooked, and start relying on them. Then, once you’re already used to it and start hitting the limits, they either push you toward a $100–200 plan or try to sell you an extra $40 worth of credits.

Of course, I’m not speaking for everyone, but I use coding agents in a very specific way. These are my rules:

  1. I clear chat history almost before every prompt to save tokens.
  2. I never ask an agent to do a huge list of tasks at once - always one isolated task, one problem.
  3. In the prompt, I always point to the files that need to be changed, or I give example files that show the kind of implementation I want.

So in practice, I honestly do not care much which AI coding agent I use: Codex, Claude Code, or GitHub Copilot / Copilot CLI. I get roughly the same result from all of them. I do not really care which one I am working with. I do not trust them with huge complex task lists. I give them one isolated thing, check that they did it right, and then commit the changes to Git.

After a while, once I got used to working with agents like this, I took it a step further. At first I was surprised when people said they kept several agent windows open and ran multiple tasks in parallel. Then I started doing the same thing myself. Usually an agent spends about 3–5 minutes working on a task. So now I run 3 agent windows at once, each one working in parallel on a different part of the codebase. In effect, I have 3 mid-level developer agents working on different tasks at the same time.

Anyway, back to the point.

Because "God bless capitalism and competition", here is what you can do instead of paying $40 for extra credits or buying a $100–200 plan: just get the cheapest plan from each provider - Codex for $20, Claude Code for $20, and GitHub Copilot / Copilot CLI for $10. When you hit the limit on one, switch to the second. When that one runs out too, switch to the third.

So in the end, you spend $50 a month instead of $100–200.

How much do you really care whether one is 10% smarter or better than another? If you are not using them in a "hand everything over and forget about it" way, but instead as tools for small, controlled, simple tasks, then it does not really matter that much.

Who else has figured out this scheme already? Share in the comments )))

r/arduino SeaRoad4079

What's the deal with this?

for sometime I've been slowly building upto trying to build my own ECU for a hydronic diesel heater, this is made up of

running a motor for combustion air, pulsing fuel pump, glow plug, water pump and a couple of temperature sensors, but anyway, back to the point of my question

I decided to open up the ECU that came with one of the heaters I've got just to see if I can learn anything about how they've done it by looking at components and googling them

I've spotted something...

above the micro controller is five open pins, am I mad to think this is TXD and RXD pins, 3.3v, 5v and GND?

what would happen if I crossed then and wired this ECU upto an Arduino, would it talk? would it mean I could open code? what sort of direction/method/principle do I use to dive further?

thanks for reading

r/midjourney mythoria_studio

Hunt time in wildoria lost island 😊 @wildoria_studio. Any feedback are welcome!🙏🔥 (constructiv)

r/ProgrammerHumor VariationLivid3193

itsOverGuys

r/comfyui VladyCzech

ComfyUI 2nd stage sampler stuck in infinite loop.

Today I worked with my workflow and somehow Upscale sampler got stuck in infinite loop. I recently updated ComfyUI from 0.14 to 0.18.1, lot of work to make it work fine, but after few days got this infinite loop. It was with UltimateSDUpscale sampler, so I switched to KSampler and got same loop. There is no way out of it, only disabling Upscale. First Sampler does not loop (RES4LYF ClownSharkSampler), only the upscale no matter which Sampler node I use. Anyone else with same issue?

Same workflow worked fine only after I made changes in frontend and this triggered the loop.

Restarting ComfyUI and reloading browser does not help.

r/megalophobia Last-Midnight-2218

400 year old Warship from Sweden

It's a Swedish warship, sank in 1628 One of the best preserved warship from 16th century

r/meme Fickle-Butterfly-338

Walmart Jeffrey... It's the best job I've ever had!

r/mildlyinteresting emfouryouare

Sweat from my nightmares bleaches my sheets

r/arduino Fillofusco2

I built a 5x16 drum machine

I honestly can’t believe it works since it’s the very first time I’m working with anything involving electronics

r/meme aBu_rASta

Yay! We all support you.

r/mildlyinteresting louiemay99

My jacket has a little pocket for the washing instructions

r/mildlyinteresting Pailox111lol

Traffic light in Mexico City shaped like the eagle of independence

r/automation Expert-Sink2302

What's the most underrated automation you've built that quietly saves you numerous hours of pain?

Everyone shares the obvious ones like lead follow-ups, invoice reminders, slack notifications when a form gets submitted. But I'm interested in hearing about automations that you amazing folks have made that are more creative, unique and impactful, but may be overlooked at times.

For me, I run synta (an n8n mcp and ai n8n workflow builder) and one of the most useful things we built for ourselves is a scheduled n8n workflow that scrapes the n8n docs, tool schemas, and community node data every day using exa and github apis, chunks it using semantic chunking via chonkie and indexes everything into a RAG store.

But the interesting part is what else feeds into it. We also pipe in our own telemetry, so when users hit errors on specific nodes or the mcp struggles to answer something accurately, those gaps get logged and the next run prioritises covering them. On top of that, it analyses workflow patterns across our user base from our telemetey data, noting what node combinations are often used together, what workflow/architecure patterns are paired together often and what new use cases are emerging, and feeds that back into the knowledge base too, so the idea is that over time the whole thing gets smarter about what people are actually building, not just what the docs say is possible.

I honestly cannot put into words how much hours this saves me, and some days I often take it for granted and even forget about it despite the fact that it helps a lot. That's why I'm curious: whether it's for personal stuff or business, what's that one automation you set up that just quietly saves you a ton of time?

Would love to swap ideas and maybe even "steal" a few!

r/Jokes CarlosDoesTheWorld

My dad used to always say “I’ll never have to worry, as long as I have social security.”

That’s what he called his gun.

r/Flux_Official Bettylovescrypto

Join us for the FluxAI Pulse #1 LIVE now! Daz and Emmanuel will be unveiling Beaver and Voice Desk in real time while taking questions directly from the community.

r/Damnthatsinteresting NationalHat3097

The only known white giraffe in the world lives in Kenya and is fitted with a GPS tracker to protect it from poachers.

r/Jokes raxcc

Nude Figure Drawing Class

When I was in college majoring in Accounting. As an autistic kid, I always had a passion for money. Last semester, I decided to explore with my electives so I enrolled in a Figure drawing class.

I walked into class and there was an old woman completely nude standing on the center of the room. I sat a little too close to her so I only got a close up of her view. Before starting, I took out my accounting homework to review it. She looked down at me and said.

“Standing up here is hard work, but it is way better than counting people’s money”.

I said okay then she bent over and I sketched her vagina.

r/comfyui bethworldismine

Unable to get realistic images (image attached with workflow)

(Sorry for my previous post without the workflow)
I’m currently using a workflow with the Klein 9B realism LoRA, but no matter how much I tweak the prompt, I’m not getting truly realistic outputs.

The main issues I’m facing is Faces look too smooth / plastic-like and Sometimes getting weird proportions (like big eyeballs)

Overall image lacks natural texture and detail

I’ve tried adjusting prompts and playing around with settings, but the results don’t change much.

Clearly I’m missing something in the setup or workflow.

Would really appreciate if someone can point out what I’m doing wrong or what I should change to get more natural, realistic results

Workflow: link here

r/Damnthatsinteresting Textilezz

He's full of raw chicken blood and hairspray and gasoline he has fungus growing inside of him and he smells like old batteries I call him the stew and the air tastes like plastic when your near it

r/Damnthatsinteresting Mint_Perspective

Antelope Canyon, Arizona: A slot canyon renowned for its wave-like structure and light beams that shine down into the narrow passageways during certain times of the year

r/funny Correct-Magician6521

Does he look like John snow ?

r/Futurology Complete_Jaguar4653

what will India be like in 2030

considering all aspects political economical is India one step away from glory or demise?

r/funny asabatonlessarmor

I did not see the S at first

r/interestingasfuck Correct-Magician6521

Does He look like John Snow ?

r/n8n juancruzlrc

Connecting AI agents to real conversations is harder than it looks

I tried plugging an AI agent into WhatsApp this week… and it was way messier than I expected

At first I thought it would be straightforward. Just connect the API, pass messages to the agent, done.

But once I started testing real conversations, things broke pretty fast:

  • the agent had no real memory between messages
  • I couldn’t see what it was actually doing internally
  • debugging felt like guessing
  • conversations just felt… off

It made me realize most of what we see with AI agents is still very “controlled environment”. The moment you drop them into something like WhatsApp, everything becomes way more chaotic.

So I ended up building a small layer for myself just to make it usable:

  • keep track of conversation history
  • connect WhatsApp cleanly to the agent
  • actually have some visibility into what’s going on

Not even sure if this is the “right” way to do it yet, but it already feels 10x more usable than what I started with.

Curious if anyone else here has tried something similar, putting agents into real user conversations instead of demos. What broke for you?

r/Rag nicoloboschi

BEAM: the Benchmark That Tests Memory at 10 Million Tokens has a new Baseline

Why the 10M Tier Is the Most Important Result

If you've been following agent memory evaluation, you know LoComo and LongMemEval. They're solid datasets. The problem isn't their quality; it's when they were designed.

Both come from an era of 32K context windows. Back then, you physically couldn't fit a long conversation into a single model call, so needing a memory system to retrieve the right facts selectively was the premise. That made those benchmarks meaningful.

That era is over.

State-of-the-art models now have million-token context windows. On most LoComo and LongMemEval instances today, a naive "dump everything into context" approach scores competitively, not because it's a good architecture, but because the window is large enough to hold the whole dataset. These benchmarks can no longer distinguish a real memory system from a context stuffer. A score on them no longer tells you much.

BEAM ("Beyond a Million Tokens") was designed to fix this. It tests at context lengths where the shortcut breaks down:

Context length What it tests 100K tokens Baseline — most systems handle this 500K tokens Retrieval starts mattering 1M tokens Edge of current context windows 10M tokens No context window is large enough — only a real memory system works

At 10M tokens, there is no shortcut. You cannot fit the data into context. The only path to a good score is a memory system that can retrieve the right facts from a pool that's too large for any model's attention window. The BEAM paper shows that at this scale, systems with a proper memory architecture achieve over +155% improvement versus the vanilla baseline. That's the regime where the gap between architectures is most pronounced, and where Hindsight's results are most significant.

The Numbers

Here's every published result on the 10M BEAM tier:

System 10M score RAG (Llama-4-Maverick) — BEAM paper baseline 24.9% LIGHT (Llama-4-Maverick) — BEAM paper baseline 26.6% Honcho 40.6% Hindsight 64.1%

Hindsight scores 64.1% at 10M. The next-best published result is 40.6%. That's a 58% margin. Against the paper baselines, it's more than 2.4x.

The full picture across all BEAM tiers:

Tier Hindsight Honcho LIGHT baseline RAG baseline 100K 73.4% 63.0% 35.8% 32.3% 500K 71.1% 64.9% 35.9% 33.0% 1M 73.9% 63.1% 33.6% 30.7% 10M 64.1% 40.6% 26.6% 24.9%

One detail worth noting: Hindsight's 1M score (73.9%) is higher than its 500K score (71.1%). Performance doesn't degrade as token volume increases; it improves. Most systems show the opposite. That's the architecture working as intended, and it's where the gap versus other approaches becomes most visible.

Results are tracked publicly on Agent Memory Benchmark. For background on why we built the benchmark and how it's evaluated, see Agent Memory Benchmark: A Manifesto.

r/funny Escapebliss

Edgar Poe

r/interestingasfuck Several_Bluebird_344

The Veryovkina cave in Abkhazia, Georgia is the deepest cave in the world, with a depth of around 2200m. It also hosts the worlds deepest living animal and still isn’t fully explored.

r/toastme Over_Face2921

Qu’est ce que j’ai fait pour mériter ça

jamais aimé toujours critiqué rien dans mon sens. Je pense que la vie en communauté n’est pas fait pour tout le monde. J’ai tout essayé pour aller mieux mais jamais quelqu’un ne m’a aimé. Sérieusement au dela du sourire je suis si moche que ça ?

r/whatisit Opal_theProxie

What is this rock? Google kept saying its a meteorite but I don't trust it.

found with other rocks near train tracks

r/KlingAI_Videos DiegoMarinho001

NEVERENDING STORY - Title Sequence

r/BrandNewSentence netphilia

If eating an underwater armored tank spider with scissors for hands is guy then I don’t wanna be right.

r/BrandNewSentence Vantabla_ck

"with a suspiciously 413,793 kitkat shaped belly"

r/me_irl gigagaming1256

Me_irl

r/whatisit GoodMornEveGoodNight

What might this be exactly?

r/whatisit flibbertigibbet72

Tiny little wooden pieces inside a wooden acorn

My mum found this in an elderly relative's house after she passed, but didn't realise it opened. When I did open it, out fell a bunch of these tiny wooden (I think) items and one bigger one that looks like a tiny spinning top but is too rounded on the bottom to work.

Any ideas for what it could be?

r/Strava ChrisBmore

AI Route Generation Removed?

Hello everyone, I mostly prefer human made routes but wanted to create some routes based on specific distances using Strava's AI route creator.

I no longer see this available on the app and website? Was this removed?

Quite open to just being unable to read. If that is the case, can someone please point me in the right direction?

FYI I am a paying subscriber so should have all features available.

r/BrandNewSentence peagatling27

If the balls are stored outside the body due to the body's heat impacting sperm production, could you theoretically heat your testicles during intercourse to act as a form of male birth control?

r/VEO3 GasDelicious5453

I used VEO3 to create a Soviet military parade.

r/n8n Mediocre_Cod_7374

I built a startup, raised money, failed — now I’m in debt. Here’s what I learned the hard way

Just a rant...

Hello Reddit community, lately I’ve seen a lot of users interacting and talking about the startup market, so I decided to share my experience working with and building startups, and also share some tips and insights from a founder who has already failed and is now recovering...

Let’s start from the beginning: I’ve always worked with processes and technology. I started my career in transportation companies, and from there I began developing and applying automations and software — even before ChatGPT, back in 2020, I was already working with automation ;)

From that point on, I kept building on the knowledge I had from college and work, and then came my layoff in 2024. I spent a few months on unemployment benefits, but that’s when I started hearing that voice in my head:
“I don’t want to work for anyone anymore. I want to have my own business and build a startup that actually solves something for companies and people.”

From there, I tried several ways to sustain myself while building my business (SaaS, software agency, automations, AI...), but nothing worked. Long story short: I ended up owing one month of rent and some money to family members. No job, and expenses piling up, I had to act fast. The most logical solution at the time was:
“I need a co-founder.”

I opened positions, activated my network, and managed to bring in two co-founders (let’s call them Tico and Teco), both with very strong skills and extensive experience (Tico had been a head in large Brazilian companies, and Teco had been a sales manager in startups for many years). We had everything we needed to launch our first product:

  1. Tico would handle business and strategy
  2. Teco would handle sales and opportunities
  3. I would handle product development and projects

And we tried everything — seriously. But still, nothing worked. Nothing took off. I felt like no matter how hard I tried (I even started doing outbound myself because Teco wasn’t doing anything), nothing moved forward.

Then I created an interesting product, which would later become the startup I’m developing today. We got accepted into Web Summit Lisbon and saw it as a unique opportunity. The problem?
We didn’t have a single dollar to go to Lisbon (I barely had money to cover my personal expenses). But then Tico managed to bring in an angel investor (a long-time friend of his). The guy trusted our project and invested enough for us to stay focused and go to Lisbon.

Chapter 2: Hype doesn’t pay the bills

We arrived in Lisbon completely unprepared — no product ready, no idea how to sell, no network, and I barely spoke English. Conclusion: money wasted (that could have been used to generate revenue or build effective sales machines).

We came back to Brazil frustrated for not being able to raise funds (which was obvious, since we only had an idea and zero paying customers).

And once again, the bills came knocking...

We decided to provide services to startups, following this idea:
Tico and Teco had enough experience to mentor founders and business owners (they were consultants), and I would support and manage fundraising-related projects (I had a strong finance network from my previous experience).

We ran it, tried, closed 2 deals, invested poorly, and the business failed.

After many attempts and almost a year of failed efforts, the investor decided to recover his investment (convertible note after one year), and here’s a key lesson:
Never take money without a clear path forward — whether from banks or investors (both are debt).

This happened recently. Since the company was formally under my name (the others had vesting agreements), I’m the one responsible for all pending costs and liabilities (vendors, investor). On top of that, I ended up using my personal finances to cover company expenses (never do this). Now I’m in debt, without income, with overdue bills, and a lot to fix — but I also have three times the experience I had when I started. I now know exactly what goes wrong, and I’ve learned how to sell and build sales machines using AI and multi-agents.

Chapter 3: The rebuild

The goal of this post is not to expose anyone or play the victim (I’m not asking for anything here), but to share a real perspective from a founder who failed, is dealing with burnout, and maybe even some hidden depression.

  1. Co-founders are the foundation of your business — get it right and you can change your life (or get rich), get it wrong and you’ll fail alone (like me)
  2. Investors don’t solve your problems — you do. If you take money that doesn’t come from your own cash flow, you will have to pay it back, and it might happen when you don’t even have a dollar in your pocket (like me)
  3. Everything passes, and you will learn. Resilience is something scary — even in this near-poverty situation (almost getting evicted), I know things will work out if I keep going. Now I’m wiser, more experienced, and I fully understand product, services, business strategy, and especially my core: AI for commercial operations

First-time founders and entrepreneurs, I hope this serves as an example of what NOT to do. And if you’re in a similar situation, comment here so we can help each other (I might be able to refer you for quick jobs or freelance opportunities depending on your case).

For successful founders and experienced entrepreneurs:
What would you do in a moment like this? What advice would you give me (and others going through the same situation)?

r/CatastrophicFailure bugminer

The bridge over the Trigno River in Italy has partially collapsed, 2nd April 2026.

r/Jokes mashupthedance

I used to work at a tree felling company, one of my colleagues was a Golden Retriever.

They had to let him go in the end, he kept barking up the wrong tree

r/VEO3 Electronic-Hippo2105

"This prompt may violate our policies regarding generating content about public figures. Please use a different prompt or send feedback."

I'm going crazy here. I keep uploading my own picture, but the system won't generate a video. I'm so angry right now. I've successfully created 100 videos before using my face for these same characters, but now it suddenly throws a 'public figure' error and blocks me. Does anyone know a fix for this? Please, someone reply!

r/Seattle breezy97bt

the soapening

does anyone know anything about the fountain by the ferry being frequently soaped? truly just curious as I walk by it sm and searched in this subreddit and didn't see discussion of it

r/ClaudeCode Wide_Incident_9881

Claude muito lento e meio burro

Só eu estou sentido que o opus 4.6 parece ter sido degradado em performance, ele parece não entender meu prompt (e ele entendia perfeitamente até uma semana atrás) e está lento demais. Para vocês também?

r/AI_Agents IAmDreTheKid

Your API Is Invisible to Every AI Agent on the Internet Right Now

There are millions of AI agents running today.

They need data. They need enrichment. They need processing power. They need exactly what you might be selling. And almost none of them can find or use your API, not because your product isn’t good. It’s because your distribution model assumes a human customer. Sign up, onboard, subscribe. Agents can’t do that, so they use whichever tool lets them just pay and call. If you don’t offer that, you don’t exist to them.

This is the distribution gap that barely anyone is talking about. SaaS companies obsess over SEO, content marketing, product-led growth, all of it optimized for human discovery. And it works great for human customers.

But the next wave of API consumers isn’t going to find you on Google. They’re going to find you in a catalog of services that accept agent-native payments, pick the one that fits their use case, and start calling immediately.

The sales cycle is zero. The onboarding is zero. The support tickets are zero. The customer either uses your API or it doesn’t. That’s the whole relationship.

BuildWithLocus lets you list your API in a way agents can actually discover and use. You define your endpoints and pricing. Agents find you, pay per call, and use what they need. No account required on their end.

The first API we saw do this properly went from 0 agent customers to over 4,000 automated calls in the first week. No marketing. No outreach. Just being available in the right place with the right pricing model.

Most APIs are invisible to agents. Being visible is surprisingly cheap.

r/ClaudeCode ThePantsThief

Has this ever happened to anyone else? A single prompt caused Claude to think nonstop, using up 4+ entire 5h sessions over 2 days before I interrupted it and then decided the conversation must be bugged and started a new one.

The new conversation only thought for a moment before actually working. I'm happy to post this novel of thinking transcripts if anyone is interested. It would often say things like I highlighted in the second image, but there were never any file edits.

(Worth noting that it didn't actually think for 25-50+ hours at a time, I'm not sure why all these numbers are in seconds and read 100k+ seconds; it would think for 10+ minutes at a time though IIRC)

r/ChatGPT elsasze

We let our 5-year-old use ChatGPT to start a cookie business. Here's what happened 🍪

One day at dinner, we were talking about patriarchy (don’t ask 😅), and my 5-year-old misheard it as “Pastry-archy.”

She immediately declared:

“In the Pastry-archy, pastries make the rule.”

And then she said she wanted to start a business. My first reaction was honestly:

I don’t have time for another business. My husband and I are already deep in a few AI projects. But instead of shutting it down, we used ChatGPT to help her actually do it.

She set up a booth at a tech conference in San Francisco.

She made $300 (profit) in ~6 hours.

------------------------------------------------------

We used ChatGPT for almost everything:

  • came up with the menu + pricing
  • helped fix a gluten-free cookie recipe (the first one we found online completely failed so ChatGPT actually debugged it and made it work)
  • taught her how to measure ingredients
  • helped her add, multiply, and scale batches
  • calculated unit economics
  • designed a logo + on-brand website
  • generated a QR code for orders

But the part that very much matters:

She stayed up until 2AM baking cookies to make it happen.

She talked to every customers.

She sold and handed out every cookie.

She ran her little booth for 6 hours straight.

It wasn’t “AI did it.”

It was more like:

AI made it possible, and she did the work.

Honestly one of the most wholesome things I’ve seen with AI.

Kids + tools like this (with supervision) can go from idea → real thing → real customers insanely fast.

Thought this story may bring a little spark to your day.

r/LocalLLaMA QuoteSad8944

Show r/LocalLLaMA: agentlint - static linter for AI coding assistant instruction files

As more teams use Copilot skills, Cursor rules, and Windsurf rules, a new class of bugs has been created: instruction drift.

Files get stale. Paths get broken. Triggers get crossed. Numbers get hardcoded without source pointers.

agentlint is the first static linter for these files. Six checks, zero config, <300ms, and works with Copilot + Cursor + Windsurf.

pip install instruction-lint && agentlint

MIT license, fully offline, and no telemetry.

https://github.com/Mr-afroverse/agentlint

r/LocalLLaMA Apprehensive-Court47

The 'Running Doom' of AI: Qwen3.5-27B on a 512MB Raspberry Pi Zero 2W

Yes, seriously, no API calls or word tricks. I was wondering what the absolute lower bound is if you want a truly offline AI. Just like people trying to run Doom on everything, why can't we run a Large Language Model purely on a $15 device with only 512MB of memory?

I know it's incredibly slow (we're talking just a few tokens per hour), but the point is, it runs! You can literally watch the CPU computing each matrix and, boom, you have local inference.

Maybe next we can make an AA battery-powered or solar-powered LLM, or hook it up to a hand-crank generator. Total wasteland punk style.

Note: This isn't just relying on simple mmap and swap memory to load the model. Everything is custom-designed and implemented to stream the weights directly from the SD card to memory, do the calculation, and then clear it out.

r/ClaudeCode entheosoul

Your AI agent is 39% dumber by turn 50..... here's a fix people might appreciate

TL;DR for the scroll-past crew:

Your long-running AI sessions degrade because attention mechanics literally drown your system prompt in noise as context grows. Research measured 39% performance drop in multi-turn vs single-turn (ICLR 2026). But..... that's only for unstructured conversation. Structured multi-turn where you accumulate evidence instead of just messages actually improves over baseline.

The "being nice to AI helps" thing? Not feelings. It's signal density. Explaining your reasoning gives the model more to condition on. Barking orders is a diluted signal. Rambling and Riffing is noise. Evidence, especially the grounded kind, is where it's at.

We measured this across thousands of calibration cycles - comparing what the AI said it knew vs what it actually got right. Built an open-source framework around what we found. The short version: treat AI outputs as predictions, measure them against reality, cache the verified ones, feed them back. Each turn builds on the last. It's like inference-time Reinforcemnt Learning without touching the model.

RAG doesn't solve this because RAG has no uncertainty scoring (ECE > 0.4* in production; that's basically a coin flip on calibration). Fine-tuning doesn't solve it because you can't retrain per-project. What works is measured external grounding that improves per-user over time.

  • ECE > 0.4 means: When RAG systems express confidence, they're wrong about their own certainty by 40+ percentage points on average. A system saying "I'm 90% sure" might only be right 50% of the time. That's the NAACL 2025 finding and not a coin flip on the answers, but a coin flip on whether the system knows it's right.

If you're building agents and wondering why session 1 is great and session 50 is mush?... keep reading.

The deep dive (research + production observations)

Been building measurement infrastructure for AI coding agents for about a year. During that time we've accumulated ~8000 calibration observations comparing what the AI predicted it knew vs what it actually got right, and the patterns are pretty clear.

Sharing because I think the industry is doing a lot of prompt engineering by intuition when the underlying mechanics are well-studied and would save everyone time.

So what's actually happening

Everyone's noticed that "being nice to AI" seems to help. People either think it has feelings (no) or dismiss it as coincidence (also no). The real answer is boring and mechanical.

Every LLM output is a next-token prediction conditioned on two things: internal weights from training, and whatever's in your current context window. One-shot questions? Weights do the heavy lifting just fine. But 200-turn agentic sessions? The weights become less and less relevant.

"Critical Attention Scaling in Long-Context Transformers" (ICLR 2025) shows that attention scores collapse toward uniformity as context grows. Your system prompt literally drowns. "LLMs Get Lost in Multi-Turn Conversation" (ICLR 2026) put a number on it: 39% average performance drop in multi-turn vs single-turn across six generation tasks.

40% worse. Just from having a longer conversation.

But only if the conversation is unstructured

This is the part that changes what we thought we knew. That 39% drop comes from unstructured multi-turn. Just... more messages piling up.

Structured multi-turn shows the opposite. MathChat-Agent saw 6% accuracy improvement through collaborative conversation. Multi-turn code synthesis beats single-turn consistently across model scales.

The difference isn't in the turn count. The question is about whether the context accumulates evidence or noise.

When you explain your reasoning to an AI, share what you're trying to do, give it feedback on what worked... you're adding signal it can condition predictions on. Constrained commands give it almost nothing to work with. Unstructured chat adds noise. But structured evidence? That's what actually matters.

What we observed over thousands of measurement cycles

We built an open-source measurement framework to actually quantify this. The setup is simple:

  1. Before a task, the AI self-assesses across 13 vectors (how much it knows, how uncertain it is, how clear the context is, etc.)
  2. While working, every discovery, failed approach, and decision gets logged as a typed artifact
  3. After the task, we compare self-assessment against hard evidence: did the tests pass, what actually changed in git, how many artifacts were produced
  4. The gap between "what it thought" and "what happened" is the calibration error

Some patterns that keep showing up:

Sycophancy gets worse the longer you go. This tracks with Anthropic's own research (ICLR 2024) showing RLHF creates agreement bias. As sessions get longer and the system prompt attention decays, the "just agree" prediction wins because nothing in context is pushing back against it.

Failed approaches are just as useful as successful ones. When you log "tried X, failed because Y," that constrains the prediction space going forward. This isn't just intuition. Dead-End Elimination as a concept was cited in the 2024 Nobel Prize in Chemistry background. Information theory: negative evidence reduces entropy just as much as positive evidence.

Making the AI assess itself actually makes it better. Forcing a confidence check before acting isn't just bureaucracy. It's a metacognitive intervention. "Metacognitive prompting surpasses other prompting baselines in the majority of tasks" (NAACL 2024). The measurement changes the thing being measured.

The RAG problem nobody wants to talk about

RAG systems in production have Expected Calibration Error above 0.4 (NAACL 2025). "Severe misalignment between verbal confidence and empirical correctness." Frontiers in AI (2025) spells it out: traditional RAG "relies on deterministic embeddings that cannot quantify retrieval uncertainty." The KDD 2025 survey on uncertainty in LLMs calls this an open problem.

So the typical pipeline is: model predicts something, RAG throws in some unscored unquantified context, model predicts again. Nothing got more calibrated. You just added more tokens.

What we found works better: model predicts, predictions get measured against real outcomes, the ones that check out get cached with confidence scores, and the next prediction gets conditioned on previously verified predictions. Each round through the loop makes the cache better.

If one speculated with grounding, this is like inference-time reinforcement learning. The reward signal is objective evidence instead of human thumbs up/down. The "policy update" is a cache update instead of degenerative descent. Per-user, per-project, and the model itself never changes. Only the evidence around it improves.

The context window problem

This is where it all comes together. Your context window is where grounding either accumulates or falls apart. Most people compact or reset and lose everything they built up during a session.

We run hooks that snapshot epistemic state before compaction and re-inject the most valuable grounding afterward. Why? Because Google's own benchmarks show Gemini 3 Pro going from 77% to 26% performance at 1M tokens. Chroma tested 18 frontier models last year and every. single. one. degraded.

The question people should be asking isn't "how do we get bigger context windows." It's "how do we stop the context we already have from turning into noise."

If you're running long agent sessions and watching quality drop off a cliff after a while, now you know why. And better prompts won't fix it. What fixes it is structured evidence that builds up instead of washing out.

Framework is MIT licensed if anyone wants to look under the hood. Curious what others are seeing with multi-turn degradation in their own agent setups.

Papers referenced: ICLR 2025 (attention scaling), ICLR 2026 (multi-turn loss), COLM 2024 (RLHF attention), Anthropic ICLR 2024 (sycophancy), NAACL 2024 (metacognition), ACL/KDD/Frontiers 2025 (RAG calibration gap), Chroma 2025 (context rot)

r/ClaudeCode Notalabel_4566

I researched Claude Code's internals via static source analysis – open-sourced the docs (Agentic Loop, Tools, Permissions & MCP)

I did some static research on Claude Code's internals (no reverse engineering, just reading the TypeScript source).

Shared my notes here:
https://github.com/Abhisheksinha1506/ClaudeReverEng

It covers:

  • Agentic loop & query flow
  • Tool system & BashTool permissions
  • Permission modes and safety checks
  • MCP integration details

Purely for learning and research purposes. Not official docs.

Feedback welcome!

r/ClaudeCode Soft_Table_8892

Since Claude Cowork crashed SaaS stocks by $285B, I built a Claude Code pipeline to score which companies it can actually replace.

Hello everyone,

Some of you might remember my previous experiments here where I use Claude Code to build a satellite image analysis pipeline to predict retail stock earnings.

I'm back with another experiment and this time analyzing the impact of the complete collapse of SaaS stocks due to the launch of Claude Cowork, by (non-ironically) using Claude itself as the analyst. Hope you'll find this interesting!

As always, if you prefer watching the experiment, I've posted it on my channel: https://www.youtube.com/watch?v=ixpEqNc5ljA

Intro

Shortly after Claude Cowork launched, it triggered a "SaaSpocalypse" where SaaS stocks lost $285B in market cap in February.

During this downturn I sensed that the market might have punished all Software stocks unequally where some of the strongest stocks got caught in the AI panic selloff, but I wanted to see if I could run an experiment with Claude Code and a proper methodology to find these unfairly punished stocks.

The Framework

I found a framework from SaaS Capital that provides a framework they'd developed for evaluating AI disruption resilience:

  1. System of record: Does the company own critical data its customers can't live without?
  2. Non-software complement: Is there something beyond just code? Proprietary data, hardware integrations, exclusive network access, etc.
  3. User stakes: If the CEO uses it for million-dollar decisions, switching costs are enormous.

Each dimension scores 1-4. Average = resilience score. Above 3.0 = lower disruption risk. Below 2.0 = high risk.

The Experiment & How Claude Helped

I wanted to add a twist to SaaS Capital's methodology. I built a pipeline in Claude Code that:

  • Pulls each company's most recent 10-K filing from SEC EDGAR
  • Strips out every company name, ticker, and product name — Salesforce becomes "Company 037," CrowdStrike becomes "Company 008", so on
  • Has Opus 4.6 score each anonymized filing purely on what the business told the SEC about itself

The idea was that, Opus 4.6 scores each company purely on what it told the SEC about its own business, removing any brand perception, analyst sentiment, Twitter hot takes, etc.

Claude Code Pipeline

saas-disruption-scoring/ ├── skills/ │ ├── lookup-ciks # Resolves tickers → SEC CIK numbers via EDGAR API │ ├── pull-10k-filings # Fetches Item 1 (Business Description) from most recent 10-K filing │ ├── pull-drawdowns # Pulls Jan 2 close price, Feb low, and YTD return per stock │ ├── anonymize-filings # Strips company name, ticker, product names → "Company_037.txt" │ ├── compile-scores # Aggregates all scoring results into final CSVs │ ├── analyze # Correlation analysis, quadrant assignment, contamination delta │ └── visualize # Scatter plot matrix, ranked charts, 2x2 quadrant diagram │ ├── sub-agents/ │ ├── blind-scorer # Opus 4.6 scores anonymized 10-K on 3 dimensions (SoR, NSC, U&U) │ ├── open-scorer # Same scoring with company identity revealed (contamination check) │ └── contamination-checker # Compares blind vs open scores to measure narrative bias 

Results

I plotted all 44 companies on a 2x2 matrix. The main thing this framework aims to find is the bottom-left quadrant aka the "unfairly punished" companies where it thinks the companies are quite resilient to AI disruption but their stock went down significantly due to market panic.

https://preview.redd.it/ulnypdz5itsg1.png?width=2566&format=png&auto=webp&s=0cc49d458adbfbcd2ad8932ffcbb38cf6726a330

Limitations

This experiment comes with a few number of limitations that I want to outline:

  1. 10-K bias: Every filing is written to make the business sound essential. DocuSign scored 3.33 because the 10-K says "system of record for legally binding agreements." Sounds mission-critical but getting a signature on a document is one of the easiest things to rebuild.
  2. Claude cheating: even though 10K filings were anonymized, Claude could have semantically figured out which company we were scoring each time, removing the "blindness" aspect to this experiment.
  3. This is Just One framework: Product complexity, competitive dynamics, management quality, none of that is captured here.

Hope this experiment was valuable/useful for you. We'll check back in a few months to see if this methodology proved any value in figuring out AI-resilience :-).

Video walkthrough with the full methodology (free): https://www.youtube.com/watch?v=ixpEqNc5ljA&t=1s

Thanks a lot for reading the post!

r/ClaudeAI XdtTransform

What data is Claude Desktop transferring??

Every now and then I'll start Claude Desktop and it starts downloading (or uploading) something in the background. It goes on for about 5 minutes. Even though I am on the latest version.

https://imgur.com/kCfQi3c

What is it transferring?

r/LocalLLaMA Longjumping-Room-170

Can I run GPT-20b locally with Ollama using an RTX 5070 with 12GB of VRAM? I also have an i5 12600k and 32GB of RAM.

I am new to this field.

r/ClaudeAI bharms27

I connected Claude Voice Mode to Claude Code and it’s kind of great.

I’ve been dying to get Claude Voice mode on mobile connected to Claude Code, and I finally figured out a (hacky) way to do it.

Which means you can access Claude Code in a way that is: Conversational, Hands-Free, and Mobile.

It uses Apple Reminders app (lol) as a bridge where voice mode puts prompts in one list, and Claude code puts updates in another. Claude code runs a /reminders skill on 1 minute /loop to check for new entries in the reminders list. Claude code’s output summaries can also be accessed via voice mode, and you can monitor both in the reminders app as well.

It allows me to walk around with my AirPods in and brainstorm with Claude Voice Mode whenever I have an idea, without taking my phone out of my pocket. Then I just tell Voice Claude to send a task to Claude Code and it starts working on it. I do my best thinking when I’m walking outside, so this has been a desire of mine since chatGPT voice mode came out.

This is obviously something that is best left to the big AI companies to do properly, since they own both ends of this process, and I think it’s crazy they haven’t already. But until then, I’ll be using this hack when I want to get away from my desk but still noodle on a project.

If you want to learn more or try it, its on GitHub: https://github.com/brianharms/reminder-watch

(It was vibe coded and I’m not a developer).

r/ClaudeAI camelCaseWA

I built an MCP server for Claude Code that auto-generates a dev diary so you never lose context between sessions (500+ downloads, free)

I was losing context every time I closed my laptop between vibe coding sessions. So I built DevGuard using Claude Code — it's an MCP server that reads your git state and writes diary entries automatically as you work.

How Claude helped build it: I built the entire project using Claude Code. The MCP server, the branch visualization, the daily view calendar — all vibe coded with Claude as my pair programmer. It's also designed specifically for Claude Code users.

What it does:

  • Auto-logs what you did, what decisions were made, what broke
  • catch_me_up gives you a morning briefing with full context
  • Visual branch map shows which versions did what
  • Zero config — one command install, diary fills itself

Install: claude mcp add devguard -- npx devguard

It's free, open source, MIT licensed. 500+ downloads so far, built entirely from user feedback.

Links: npmjs.com/package/devguard | wjung6799.github.io/devguard | Discord: discord.gg/BrzRHHzjFQ

Would love feedback from other Claude Code users. What context do you wish your AI remembered between sessions?

r/ClaudeAI abzod9

Claude.ai vs Claude Code

I was wondering how can i use Claude Code but with powers of Claude.ai. Basically, trying to create a custom UI to my liking.

And the other question, is there a way to natively export the entire chat history from Claude.ai via api or any other way?

r/LocalLLaMA datafense

i was scared to give openclaw my password: can i pls get feedback before i open source pls?

I've been experimenting with AI agents like OpenClaw but was uncomfortable giving them direct access to my Gmail and other accounts. The security risks are well documented exposed credentials, etc.

created middleware layer that sits between the agent and your accounts. The agent never gets your credentials. Instead it sends requests through this layer where you set rules like "show emails about flights but hide anything from my bank." Filtered emails are completely invisible to the agent. Everything is logged in an audit trail.

The credential vault and database run on a private network with no public internet access. Only the MCP server is public-facing.

this demo is very short (maybe too short) but honestly took me forever to record something short enough for folks attention spans lol that still showed what its doing but if you have questions lmk pls https://www.youtube.com/watch?v=JmGKBxJlPSE

Architecture details: https://app.datafense.ai/#security

Planning to open source the MCP server and permission engine this week. Looking for feedback:

  1. Would you self-host something like this, or prefer a hosted version?
  2. What account integrations matter most to you beyond Gmail?
  3. Anyone want to review the security architecture and poke holes in it? pleaseee

Happy to answer any questions about the approach.

r/ClaudeAI Agitated-Syllabub-64

Max 20x — did everything to try and reduce token burn in Cowork and it's still draining fast. What am I missing?

*CLAUDE MADE POST*

I run 12 Cowork projects across 4 businesses on the Max 20x plan ($200/mo). After watching my usage fly through the roof, I spent today optimizing everything I could think of. Here's what I did:

**Changes made today:**

- Moved every project from Opus to Sonnet (only 2 code projects stay on Opus)

- Moved Master Dispatch (my heaviest context session) down to Sonnet

- Trimmed my main dispatch file from 16KB to 5KB (removed completed items, archived history)

- Trimmed my session boot file from 4.7KB to 1.5KB

- Removed two reference docs (8KB + 12KB) that were sitting in the shared folder and getting loaded by every session

- Added explicit read-scope rules telling each session to only read its own business files and skip the rest

- Confirmed the scoping works — sessions are correctly ignoring files outside their scope

- Total shared context reduction: ~45KB down to ~15KB per session

**Still burning fast.** Session usage hit 90%+ within a 90mins. This was happening both during and outside of peak hours (8AM-2PM ET). Weekly usage sitting at 54% by midday Wednesday.

**My setup:** 12 Cowork projects sharing a folder of ~23 markdown context files organized by business. Each project reads 3 small control files + its own business folder. Connected tools: Gmail, Monday.com, Claude in Chrome.

**What I genuinely don't know and would love answers on:**

  1. Does Cowork load every file in the project folder into context on every message, or only the files the session actively reads? This is the big one — if it's loading everything regardless of instructions, my scoping rules are saving nothing on actual token cost.

  2. Do connected tools (Gmail, Monday.com, MCP servers) have a token cost just by being connected, even if they're not called in a given message?

  3. Is there any way to reduce what Cowork sees beyond choosing the folder at project creation? You can't change the folder path after the fact.

  4. Anyone running Haiku for Cowork operational tasks (CRM lookups, email drafts, status checks)? Does it hold up?

  5. How much are people actually getting done per day on Max 20x right now? I'm getting maybe 3-4 productive hours before things get tight, and I've optimized everything I can on my end.

I'm not complaining about the product — the multi-project Cowork system I built genuinely works well. I just want to know if there's something else I can do on my end, or if this is just the current state of the platform and we're all dealing with it.

r/ClaudeCode Proof_Net_2094

Last was my first time ever complimenting an AI tool (Claude Code)

Just a week or so ago I realized myself complimenting claude code saying that it is the only usefull AI tool ever built, not sure if I should take that back or hold on to it?

r/SideProject InternetWrong9088

I built a minimalist daily AI & Tech intelligence

Hi everyone,

I’ve always found it hard to keep up with the overwhelming flow of AI news without getting lost in the "hype" or social media noise.

To solve this for myself, I built Gifteba (TechPulse). It’s a clean, curated daily feed of the most critical updates in AI, Cybersecurity, and Quantum computing. For example, today's focus is on Anthropic's code hemorrhage and Google's AI Inbox.

The goal: 2 minutes of reading to get the "Daily Intelligence" you actually need.

I’d love to get your feedback on the UI and the curation quality. Does it feel useful for your daily workflow?

Check it out here: https://gifteba.com

Thanks for your time!

r/LocalLLaMA MrYoge

Seeking advice: Best sites with global shipping for cheap headless mining GPUs (P104, CMP 40HX) for a budget Linux / Local AI build?

Hi everyone,

I’m a computer engineering student planning a strict-budget project. The goal is to build a cheap but quite strong Linux machine to run local AI models.

To keep costs as low as possible, I'm trying to be creative and use headless crypto mining GPUs (no display output). Models like the Nvidia P104-100 8GB or CMP 40HX/50HX seem to offer amazing VRAM-to-price value for this kind of project.

The problem is that the used hardware market in my country is very small, and these specific cards are almost non-existent locally.

Do you guys have any recommendations for reliable sites, platforms, or specific sellers that offer global shipping for these types of GPUs? My budget for the GPU itself is around $50-$75.

Any advice or alternative budget GPU recommendations would be greatly appreciated. Thank you!

r/ClaudeCode smellyfingernail

Anyone having a problem where the claude code terminal will suddenly remove or hide all previous parts of the conversation when doing an exploration?

In claude code, i will be talking with claude about the project, then usually after I give it a wide ranging exploration task like "Explore how this project interacts with...", it will launch a few explore agents but then every other part of the conversation we had suddenly disappears and becomes inaccessible.

This is on v2.1.90

r/ClaudeCode PhallicPorsche

Google just dropped Gemma 4. Has anyone tried it in an MCP to make Claude better at Claud..ing?

Has anyone tried Gemma 4 yet? Google released an open weight offline capable model that's supposedly "frontier capable" (Whatever those words mean)

https://preview.redd.it/5yz71xivdtsg1.png?width=2068&format=png&auto=webp&s=9d7acd9bbf3cd5a99f23ef26ca1c6bc177135a1c

I suspect it may be a good agentic specialist to pair with anthropic models to save on those rate limits everybody keeps complaining about. Has anyone run it offline yet? What GPU are you using it with? I sold my offline setup a while ago and wouldn't mind hooking it up on something respectable (5-10K budget).

r/LocalLLaMA soyalemujica

I feel like getting 128gb ram was a mistake for agentic coding.

I was running 16GB VRAM and 64gb ram for practically some months, using Qwen3-Coder at Q5 or Q4 for some non-complex coding (since it's not a perfect model).
So I thought, well lets get 64gb ram so I can get 128gb ram and maybe use more models.

And here's the hard reality that struck me:
StepFlash 3.5 runs at 10t/s, and slows down to 8t/s at 100k context.
122B A10B Qwen 3.5 runs at 14t/s and slows down to 10t/s at 100k context (reasoning and non-reasoning, Qwen3-Coder does the same task and I do not believe at Q8 would be a noticeable difference).
Pretty much it.

In reality it is not worth it at all for me to run such big models at less than 20t/s because it's way too slow for agentic coding, taking over 30 minutes for tasks that me as a programmer could manage on my own.

Why are rams so expensive then ? It does not make sense to me in any agentic coding point of me.
Maybe I am missing something, or my own autistic brain expected to get 20t/s or even 30t/s in 70b+ models.

So it's best to just return this RAM and save more for at least 24gb vram ? Would a 7900XT 24gb be a better choice ?

r/ClaudeAI DevMoses

What the Leaked Claude Code Source Won't Tell You

Three days ago Anthropic accidentally shipped a source map in their npm package and the entire Claude Code codebase leaked. 512K lines of TypeScript, 44 feature flags, and a hidden system called KAIROS: an always-on background agent that consolidates memory while you're idle, merges observations, removes contradictions, and preps context so it's clean when you come back.

I built the same thing independently in March. Not because I saw it coming. Because I hit the same ceiling.

The ceiling is real. I'm a solo developer building a 668K-line TypeScript platform with Claude Code. I run autonomous campaigns that span multiple sessions, with persistent state files that carry context across context window boundaries. Campaigns stall between sessions. You finish a phase, close the terminal, come back tomorrow, and have to manually restart, re-read the campaign file, figure out where things left off. The agent's memory dies with the session. So I built a daemon that chains sessions via scheduled triggers. One session finishes, writes state, exits. The daemon detects the exit and spawns the next session with full context. Campaigns that took a week of manual restarts now complete in one stretch.

Exit code 0 means "no errors." It does not mean "it works."

The first night I ran the daemon, an agent shipped an invisible feature. Full campaign completed. Typecheck clean. Zero warnings. Confident exit. When I opened the app, 37 of 38 entities were missing. Everything the agent built was structurally correct and completely non-functional. The daemon had done its job perfectly: consolidated context, maintained state, chained sessions without dropping anything. The work it produced was empty.

Then a fleet session replaced 6 working components in parallel. Every component showed "Running NaN," no timeline, no vitals. 841 lines replaced with 144 lines of broken output. The agents never rendered what they built. They checked that it compiled and moved on.

Those two failures built the verification layer. The daemon alone is a faster way to ship broken code. What makes it useful is forcing agents to prove their work visually: navigate real routes in a real browser, count DOM elements, capture screenshots. If a view that should have 38 entity cards has zero, that catches it. If an agent modified UI files, it cannot complete without screenshot artifacts. Hard gate, not a suggestion.

KAIROS solves the memory problem. It doesn't solve the verification problem. It merges observations, removes contradictions, converts vague insights into concrete facts. That's necessary. But neither memory consolidation nor daemon mode addresses the fundamental gap: agents can't verify their own work visually. They can prove structure. They cannot prove appearance.

The convergence tells you something. Anthropic and a solo developer hit the same ceiling independently. Once your sessions are long enough and your campaigns span days, persistent background execution becomes inevitable. But the daemon is the easy part. Anyone can chain sessions. The hard part is building the infrastructure that catches failures the daemon will confidently ship.

If you're building any form of autonomous agent execution, ask one question before you ship it: can my agent prove that what it built actually works? If the answer is "it compiled," you're about to learn the same lesson I did. 27 documented postmortems taught me that the daemon is a force multiplier. Without a quality layer, it multiplies your failures.

The daemon, the verification layer, and the campaign persistence are open-source: github.com/SethGammon/Citadel

r/ClaudeAI dovyp

I caught Claude and ChatGPT making the same lazy shortcut. Your imagination is the real bottleneck, not AI.

Building a sensor fusion device. 3 main input sources, one of them is a dual-mic array. ChatGPT wrote the audio processing pipeline first. It merged both mics into a single mono channel.

Just... flattened them together as mono. No beamforming, no spatial awareness. Took the fastest path.

I moved the codebase to Claude. Same thing. Claude looked at the existing code, agreed with it, and kept the mono merge. Two different AIs, same lazy shortcut.

I had to be the one to say "hey, we have two mics at a known distance apart, we should be doing beamforming and using stereo to calculate spatial data." Claude immediately got it. "Oh yes, you're right, we should absolutely be doing that." Cool. But you didn't think of it on your own.

Same project, different problem. I'm training a model with test subjects of wildly different sizes. AI just threw them all into the same training pool. I had to push back and say we need to group subjects into age cohorts. It was then that Claude had the idea to z-score normalize across them so a small subject and a large subject can contribute equally to the model after I mentioned. Claude ran both concepts with it and the accuracy jumped significantly. But again, it wouldn't have gotten there alone.

Here's what I've learned after months of building with AI daily:

  • AI will always choose the fastest path. Not the best path. Not the most creative path. The path of least resistance. Every single time. It's your job to know when that shortcut is actually costing you.
  • The people who are getting 10x results from AI aren't better at prompting. They have domain knowledge and imagination. They know what SHOULD be possible even if they can't code it themselves. Then AI becomes the hands that build what your brain designs.

My workflow now: take the same prompt, run it through Claude, Grok, ChatGPT, and Gemini. Get four different outputs. Then feed all four back into Claude Opus (4.6) and have it synthesize the best parts. The output is consistently better than any single AI alone.

Don't just accept what AI gives you. Push back. Ask "is this actually the best approach or just the easiest one?" Your experience and imagination are the multiplier. AI is just the calculator.

r/ChatGPT yesforevertrying

I did a photoshoot and I asked ChatGPT to pick the best ones for me. Thoughts?

Please tell me that it’s not stupid and that it can actually lead to better images.

I told ChatGPT they would be editing etc and asked it to select the best images for me. I’d pic a few that I was indecisive about and ask it to pick for me.

Is this a bad idea? Should I rely solely on my own judgment or could it lead to better images?

r/AI_Agents Proof-Ant-431

If an AI agent lived on your desk instead of your browser, what would it actually need to do to be worth keeping?

everyone here seems to be building complex orchestration pipelines and arguing over the best frameworks. tbh i've been going the exact opposite direction lately.

for the last few months my small team has been trying to pull an agent out of the terminal and trap it inside a physical desktop device.

we're not trying to build some magical Jarvis that runs your entire company. we just wanted a physical interface... basically an animated desktop companion (went with a cyberpunk cat vibe we're calling Kitto) that actually feels present in the room.

honestly here is the uncomfortable reality of 'embodied' AI. the moment you add a screen and try to do real-time lip-sync and expressions, you can't hide behind a blinking cursor or a typing indicator anymore. latency will absolutely kill the illusion.

our boring stack right now is just an esp32s3+esp32p4 chip (though we are actively migrating to a linux board because the esp32s3+esp32p4 is definately hitting its ceiling), standard LLM API calls + TTS, and a custom bionic algorithm that maps audio features to code-driven animations in real time.

the hard part hasn't even been the LLM. its been the pipeline to get the mouth and eyes to sync naturally with the generated audio without a massive delay.

building this made me step back and question the actual utility of hardware agents though. we are so used to AI living in browser tabs that we just close when we're done.

so if you had a physical agent sitting next to your monitor right now... always on, visually reacting to you, maybe connected to OpenClaw down the line for local actions... what would it actually need to do to earn its spot? what features would make it a daily driver, and what would just get annoying after a week?

r/AI_Agents Fluid_Equipment_6234

20yo running a "AI Agency." Built 5 sites, getting 0 replies. Is "Spec Work" a trap?

I need some high-level strategy. I’m 20, based in South Asia, and I just rebranded my freelance hustle into an agency called ALTO. I’m targeting US/International high-ticket niches (Pool construction, Car detailing, etc.).

The Stack & The Struggle:

  • The Tools: I use Lovable and Draaft (3D). I haven't paid for pro subscriptions yet, so I’m building everything in Free Demo Mode.
  • The Portfolio: I’ve built 5 solid "Concept" sites. Since they are in demo mode, I don't have live URLs. I’ve been screen-recording them or sending temporary preview links to show "proof of work."
  • The Strategy: I find a business on Google Maps with a trash site/no Instagram, build a custom 3D concept for them, and DM/email it.

The Wall I’m Hitting:

  1. The Ghosting: I’m spending hours building custom demos and getting zero replies. It’s burning me out. Is "Spec Work" (building for free) a total waste of time at $600/project?
  2. The "Demo" Look: How do I professionally show off these "Free Tier" sites to a US business owner without looking like a kid playing with tools? Should I just use high-quality screen recordings (Loom) instead of links?
  3. Instagram Growth: I just rebranded to ALTO. I need to post content that makes me look like a 10k/month agency, but I’m a one-man show. What kind of posts actually convert business owners?
  4. The Outreach Gap: Most US contractors I find only have a phone number. If they aren't on IG, how do I "show" them a 3D website concept?

My Current Pricing: * $600 for the Build (Infrastructure)

  • $200/mo for Maintenance/Updates
  • $500/mo for IG Brand Management (Learning this on the fly)

Questions for the pros:

  • What AI tools can I use for $0 to create high-end IG content for my agency?
  • Is $600 too cheap? Does it make me look "offshore and low-quality"?
  • How do I close that first 50% deposit when the client knows I'm using AI builders?

I’m tired of the "brokie" local market. I have the eye for design and the speed, but the sales process is broken. Help a brother out.

r/SideProject Fluid_Equipment_6234

20yo running a "AI Agency." Built 5 sites, getting 0 replies. Is "Spec Work" a trap?

I need some high-level strategy. I’m 20, based in South Asia, and I just rebranded my freelance hustle into an agency called ALTO. I’m targeting US/International high-ticket niches (Pool construction, Car detailing, etc.).

The Stack & The Struggle:

  • The Tools: I use Lovable and Draaft (3D). I haven't paid for pro subscriptions yet, so I’m building everything in Free Demo Mode.
  • The Portfolio: I’ve built 5 solid "Concept" sites. Since they are in demo mode, I don't have live URLs. I’ve been screen-recording them or sending temporary preview links to show "proof of work."
  • The Strategy: I find a business on Google Maps with a trash site/no Instagram, build a custom 3D concept for them, and DM/email it.

The Wall I’m Hitting:

  1. The Ghosting: I’m spending hours building custom demos and getting zero replies. It’s burning me out. Is "Spec Work" (building for free) a total waste of time at $600/project?
  2. The "Demo" Look: How do I professionally show off these "Free Tier" sites to a US business owner without looking like a kid playing with tools? Should I just use high-quality screen recordings (Loom) instead of links?
  3. Instagram Growth: I just rebranded to ALTO. I need to post content that makes me look like a 10k/month agency, but I’m a one-man show. What kind of posts actually convert business owners?
  4. The Outreach Gap: Most US contractors I find only have a phone number. If they aren't on IG, how do I "show" them a 3D website concept?

My Current Pricing: * $600 for the Build (Infrastructure)

  • $200/mo for Maintenance/Updates
  • $500/mo for IG Brand Management (Learning this on the fly)

Questions for the pros:

  • What AI tools can I use for $0 to create high-end IG content for my agency?
  • Is $600 too cheap? Does it make me look "offshore and low-quality"?
  • How do I close that first 50% deposit when the client knows I'm using AI builders?

I’m tired of the "brokie" local market. I have the eye for design and the speed, but the sales process is broken. Help a brother out.

r/ClaudeCode Deep-Station-1746

Claude Code source (full)

https://codeberg.org/tornikeo/claude-code

In case you were late to the party and can't find the leaked claude code source, it's here.

Have fun and be careful with the package installation. Some people started squatting private package names that are referred to, in that repo -- those are anthropic's private npm packages and all the public ones are currently being squatted by bad actors. If you install them, you might get pwned.

Good luck and have fun! :)

r/LocalLLaMA Unlikely-Tomorrow432

How to download the claude code leaked file as text version? And from where safely?

sorry jf i sound retarted

r/ClaudeAI tomas_f

So many posts about usage without numbers, so I did test. I am asking for evaluation help.

So I did build tracking for the sessions, to have numbers instead of words.

I am not saying its good, I am not saying its bad.

I am asking you for help to evaluate if it is enough, or not enough, what are the percentages?

These runs were made on Max x5 plan on session limit being 0% and weekly 63% (off-peak hours).

The first image show compositions per job (ordering of the jobs is the same, I dont want to expose the job names)

Second image is actual % that it ate for session window.

The third image shows first image in numbers.

Fourth image is analysis of the sessions (what happened).

Fifth image is rundown of how context growths with the tool usages.

You be the judge, I am curious what you guys think of it.

r/SideProject 883Infinity

What LaraLean actually generates (with YT video links)

A few days ago I shared LaraLean, a lean Laravel + Inertia + Vue + Tailwind boilerplate.

Some people asked what it actually does beyond "another boilerplate" and some of my choices.

On laralean.com I added docs and a live demo.

Then I recorded two short videos to make it concrete:

  1. Quick start (1 command setup -> ready to run)
  2. Scaffolding (from config file -> full module skeleton)

Scaffolding example:

  • define a resource in the config file
  • run one command
  • get models, migrations, controllers (role-based), requests, policies, seeders
  • Vue pages already bounded with controllers and prepared with layout, forms, props, routes.

No hidden layers. Just a consistent structure on top of Laravel.

Would love honest feedback, especially from people who’ve tried other Laravel boilerplates.

r/ClaudeAI Woodrider92

I built a self-improving LinkedIn skill system in Claude. 3 posts. 1 week. 110K impressions.

I’m a freelance B2B marketer running a one-person agency. Claude is my entire execution layer. Content, strategy, websites, the lot.

A few weeks ago I stopped treating Claude like a writing tool and started building it into a system.

What I built:

First, a LinkedIn writing skill. Not a prompt template, an actual skill file that contains my voice patterns, hook structures, post frameworks, and reference examples pulled from my own writing. Every post Claude writes for me now sounds like me, not like AI.

But writing better posts wasn’t enough. I needed the system to learn what was actually working.

So I built a second layer on top by adding a performance enhancement skill. This one has five components. A Data Store that logs raw post metrics after every post. A Pattern Engine that identifies what’s driving engagement across hook type, structure, topic, and format. Active Rules, which is the current playbook that updates based on what the data says. An Inspiration Hooks bank of proven angles to pull from. And an Evolution Log that tracks every rule change so the system has memory of what it tried and what worked.

The two skills talk to each other. The writing skill follows the active rules. The performance skill updates the rules based on real data. It’s a feedback loop.

What happened:

This week, 3 posts hit a combined 110K impressions. One of them crossed 56,000 on its own. Off the back of it, I got inbound interest from a B2B SaaS startup founder and an AI security agent startup founder. No ads, no outreach.

I’m not saying the numbers are because of the skill system alone. But the consistency shifted. I went from “some posts do well, most don’t” to “most posts do well, and I understand why.”

The real unlock:

The skill system isn’t magic. It’s just structured feedback. The same thing any good content team does track, analyze, adapt except I’m one person and Claude is doing the heavy lifting.

If you’re using Claude for content and you’re still copy-pasting prompts, build a skill instead. Give it your voice. Give it data. Let it evolve.

AMA about the setup.

r/SideProject MethodScary4643

I’m making an app solo, this is my progress so far

Journals or journaling apps are, imo, terrible for people with adhd. I got diagnosed as an adult and everything started making sense— that I’m not weird or lazy for abandoning journals constantly. It’s literally just the environment that works against my brain. So I’m making an app that allows me to journal in peace with a bit of incentive to keep my retention. It’s nowhere finished but I just wanted to share it.

r/SideProject Dependent_Bite9077

Taking the boredom out of code reviews

I use this tool every day to annotate screenshots taken when testing. Makes the pull requests a little more fun. The app I am annotating is just a random app, but you may get the idea.

r/LocalLLaMA Usual-Carrot6352

Vintage Model - flop US open source

thats 15months

r/LocalLLaMA Interesting_Bag_9346

Best local STT model for Japanese + English? Looking for high accuracy on both languages

Hey everyone,

I'm building a local AI pipeline and looking for the best STT model that runs locally with high accuracy in both Japanese and English.

My current setup:

- RTX 2060 Super (8GB VRAM)

- Windows + WSL2

What I've tried:

- Whisper (various sizes and Japanese finetuning) — dissatisfied with both accuracy and speed, curious if there's anything better

What I'm looking for:

- High accuracy on both Japanese and English

- Runs locally (no cloud API)

- Reasonable latency for near real-time use

Is Whisper still the go-to, or are there better alternatives in 2026? Especially interested in models that handle Japanese well since it tends to get less attention than European languages.

Thanks

r/ClaudeAI Kadion5

Claude uninstall problem

So I decided to uninstall Claude after I saw how big it is. After I uninstalled it, the start menu shortcut was still there, doing nothing, so I restarted the PC, but it was still there. I deleted every file from Claude, but still nothing. When I reinstall Claude, there are two shortcuts: One hd, other llow quality, the fake one. If i press on uninstall on the fake one, It oppens the Microsoft Store uninstall menu, the one with "Uninstall" and "Cancel", but Claude isn't even on MS Store. The real one opens Setting App, like other desktop apps. I tried everything, but it's still there.

r/SideProject Gloompen

I just published my first app "Doodle Magic", you can transform your doodles into real life figures and animation. Im so excited 😅

Hey guys, I have been seeing your apps and always wanted to have mine here and its finally that day! Your comments and feedbacks are much apprecated! Since this is my first app, there are many parts to improve. Its a great feeling I keep walking around my apartment Im super excited right now haha

You can transform more than 30 styles and you can order your creation as 3D print. US and most of EU countries are available to deliver!

Here is the link: https://apps.apple.com/us/app/doodle-magic-draw-art/id6759069289

r/SideProject Mr_cheesicus

I built a revision app that combines Anki, Notion and Quizlet into one — would love feedback

Hey everyone,
I'm a software engineering student and I just shipped my first real full stack project
Vestige, a revision and study app.
The problem I was trying to solve: I kept switching between Anki for flashcards, Notion for
notes and Quizlet for practice questions. It was messy and fragmented so I just built one
thing that does all three.
What it does:
- Organise revision into subjects and topics
- Create flashcards with a flip animation and study mode
- Write notes per topic with auto-save
- Add practice questions with reveal on click
- Track progress over time so weak topics surface automatically
- Search across everything
Built with Next.js, Supabase and custom CSS. No Tailwind, wanted full control over the styling.
It's still in development as I get new ideas too add. Would genuinely love feedback from anyone willing to give it a try — there's a feedback form built into the app or you can get the link for it here.
Link: vestige-gamma.vercel.app
It does use email signup but confirmation isn't necessary so feel free to use a fake email if you want.
Feedback Form: https://forms.gle/sPb4qLd7Kro1tUmP7
Happy to answer any questions about the build too — learned a lot making this.

r/ClaudeAI General-Fall-3151

'htop' for Claude Code sessions and worktrees

https://preview.redd.it/elqov376otsg1.png?width=1762&format=png&auto=webp&s=e28040682a585256a9ab67a2854e6f278b5a6e24

I found myself juggling many Claude Code sessions and worktrees. I wanted a way to monitor when each session was waiting for input vs working, what servers/ports were running for each worktree, and have shortcuts to frequently used actions such as creating a new worktree working session, tailing log files, or cleaning up old worktrees. I so built this mvp project with claude code.

So far I'm finding this very useful to lower the context switching overhead I have when juggling many sessions at once.

If there's interest, I'll open source it for free so anyone can use it with their own projects. What would you add to this dashboard?

Features:

  • Live reloads every 1 second
  • For each worktree (and 1 main working dir)
    • list what git branch that working dir is using
    • how many changes are unstaged/unpushed
    • is Claude Code waiting or working (for each Claude Code session in this worktree)
    • process ids and cpu usage for each Claude Code session in this worktree
    • what ports are being used for my frontend and backend
    • working dir
    • location of logs files
    • show the tail of logs files if there are errors
  • Hot Keys
    • create a new worktree, open new terminal with Claude Code, start servers, open Chrome tab with frontend
    • open Claude Code [for any shown worktree]
    • open VS Code [for any shown worktree]
    • start servers [for any shown worktree]
    • stop servers [for any shown worktree]
    • delete any or all worktrees
    • open new terminal window to tail backend logs
r/ClaudeAI Brilliant-Beyond-856

Used Claude + MCP to re-engage old LinkedIn leads automatically — workflow breakdown

https://reddit.com/link/1sanzvh/video/4bwpdii7btsg1/player

I tried building a re-engagement workflow using Claude to handle outreach to old LinkedIn connections.

Instead of manually going through my network and figuring out who to message, I set up a system where Claude:

  • Identifies and prioritizes leads (hot / warm)
  • Detects basic engagement signals
  • Generates contextual re-engagement messages
  • Executes the outreach step

What I found interesting was how it handled constraints.

Out of 7 targeted leads:

  • 5 messages were successfully sent
  • 2 were skipped due to LinkedIn restrictions (connection settings / limits)

Instead of forcing the action, it suggested a fallback:
engage with their content first, then retry later.

That made the workflow feel less like rigid automation and more adaptive.

This basically replaced the manual process of:
opening profiles, checking context, and writing messages individually.

Still iterating on the scoring logic and prioritization.

Would be interesting to know if anyone else here is using Claude with MCP for workflows beyond just text generation.

r/SideProject Ok_Selection5420

I made a free landing page teardown tool for indie hackers and side projects

If you’re building a side project and not sure whether your landing page is actually clear, I made a free tool for that. It lets you paste your URL and get a quick analysis of your headline, positioning, CTA, trust signals, and overall conversion story. The goal is to help founders spot weak points fast and improve their page without needing to hire a marketer or wait for random feedback.

https://www.appwispr.com/analysis

r/AI_Agents Accomplished-Fun-436

I built an AI learning path for myself

I created a personal roadmap to learn AI by building:

  1. Linux basics

  2. Python + APIs

  3. Prompt engineering

  4. RAG

  5. Build a full AI app

Instead of doing them separately, I’m trying to connect everything into one project.

Does this approach make sense?

Anything you would change?

r/LocalLLaMA ConfidentDinner6648

My first impression after testing Gemma 4 against Qwen 3.5

I have been doing some early comparisons between Gemma 4 and Qwen 3.5, including a frontend generation task and a broader look at the benchmark picture.

My overall impression is that Gemma 4 is good. It feels clearly improved and the frontend results were actually solid. The model can produce attractive layouts, follow the structure of the prompt well, and deliver usable output. So this is definitely not a case of Gemma being bad.

That said, I still came away feeling that Qwen 3.5 was better in these preliminary tests. In the frontend task, both models did well, but Qwen seemed to have a more consistent edge in overall quality, especially in polish, coherence, and execution of the design requirements.

The prompt was not trivial. It asked for a landing page in English for an advanced AI assistant, with Tailwind CSS, glassmorphism, parallax effects, scroll triggered animations, micro interactions, and a stronger aesthetic direction instead of generic AI looking design. Under those conditions, Gemma 4 performed well, but Qwen 3.5 still felt slightly ahead.

Looking at the broader picture, that impression also seems to match the benchmark trend. The two families are relatively close in the larger model tier, but Qwen 3.5 appears stronger on core text and coding benchmarks overall. Gemma 4 seems more competitive in multilingual tasks and some vision related areas, which is a real strength, but in reasoning, coding, and general output quality, Qwen still looks stronger to me right now.

Another practical point is model size. Gemma 4 is good, but the stronger variants are also larger, which makes them less convenient for people trying to run models on more limited local hardware. For example, if someone is working with a machine that has around 8 GB of VRAM, that becomes a much more important factor in real use. In practice, this makes Qwen feel a bit more accessible in some setups.

So my first impression is simple. Gemma 4 is a strong release and a real improvement, but Qwen 3.5 still seems better overall in my early testing, and it keeps an advantage in frontend generation quality as well.

r/AI_Agents Away-Elk2460

We tested 5 techniques for improving LLM judges - only 2 actually work (open source, RewardBench 2)

We ran a systematic study on what actually improves LLM-as-judge accuracy on RewardBench 2 (1,753 examples across factuality, focus, math, instruction following, and safety).

What works:

  1. Task-specific criteria - add one sentence to the judge prompt telling it what to focus on for this specific task. +3pp at zero cost. E.g. for math: "Focus on whether the mathematical reasoning is logically valid, the steps are correct, and the final answer is accurate."
  2. Ensembling - request k independent scores, take the mean. +9.8pp at k=8, but k=3 captures most of it. Use temperature=1.0 for max diversity.

Combined: 71.7% -> 83.6%.

The mini model finding that might save you money:

GPT-5.4 mini with k=8 hits 79.2% at 0.4x the cost of a single full model call. Add task-specific criteria and it matches the full model ensemble (81.5%) at roughly 1/10th the cost. If you're running judges on every request, this is probably the operating point you want.

What doesn't work (we tested these so you don't have to):

  • Calibration examples (showing a scored reference) - marginal at k=1, zero effect at k=8
  • Routing between mini and full model based on score variance - dead zone in the middle of the cost curve
  • Weighted blending of mini + full scores - overfits, doesn't generalise
  • Stacking everything together - the combined approach scored LOWER than just criteria + ensembling

Interesting side finding: temperature=0 is not deterministic. Even at temp=0, k=8 ensembling gives +4.6pp over k=1. Probably floating-point non-determinism in GPU inference.

Everything is open source

r/ClaudeAI ipk00

Claude Code inquiry

Greetings to the community.

Over the past month, I used Sonnet/Opus to develop a 10K lines single file .html app.

I have upload the app in Netlify and currently contemplating a Firebase Auth/Firestone implementation.

Should I start working with Claude Code or will chat suffice?

r/ClaudeAI contrasupra

Why can’t I upload photos?

Hopefully this is allowed - I’m not sure whether it’s a bug or something I’m just not understanding. I’m new to Claude and obviously a pretty casual user - I’m just trying to get some advice on my houseplants. I was able to upload photos yesterday, but today when I click on them the checkmarks are gray and nothing appears to be happening. Can anyone explain? I’m not getting any errors or anything, it just doesn’t seem to work.

r/ChatGPT evaiguess

how do i get my chatgpt to not be lame

in the personalization i put gen z slang but idk why its doing this i see people who have their chatgpts actually be full of slang but mine is giving millennial TikTok humour how do i fix it

r/LocalLLaMA tm9657

Local Workflow Engine Meets Local Models

Hey everyone,

I've been building an open-source, local-first workflow automation engine called Flow-Like and wanted to share it here since local model integration is a core part of the design.

The idea: A visual workflow engine that runs entirely on your machine — no cloud, only if you want. Everything is Rust under the hood with WASM-sandboxed nodes.

Local LLM integration — three ways:

  • llama.cpp baked in — Flow-Like ships with llama.cpp support built into the app. Select your models from the catalog & wire them into a workflow. No separate server process needed.
  • Ollama — If you're already running Ollama, just point a node at your local endpoint. Works with whatever models you've pulled.
  • LM Studio — Same deal. If LM Studio is your setup of choice, connect via its local API and you're good.

The point is you're not locked into any single way of running models. Use whatever fits your setup. Even remote providers. One API.

Why this matters for workflows:

Most automation tools either force you into cloud APIs (and the associated cost/privacy tradeoffs) or have bolted-on "AI" features that are basically just an OpenAI API wrapper. Here you can build entire pipelines — data processing, document handling, RAG chains, agents — where the LLM part runs on your hardware just like everything else. Works on your phone too.. no hosting needed, data stays yours.

Workflows are fully typed with complete data lineage, so you always know what's flowing where. And since external nodes run in WASM sandboxes, even community-contributed nodes can't touch your filesystem or network unless explicitly allowed.

Stack for the curious:

  • Pure Rust core
  • Tauri for the desktop app (macOS, Windows, Linux + mobile)
  • WASM Component Model for node sandboxing (supports 16+ authoring languages - Python, JS, Java, Kotlin, Rust, C++, ...)
  • Fully offline-capable — works on an air-gapped machine

Still working toward v1 but it's usable today. Would love feedback from this community since you care about running things locally rather than just talking about it.

GitHub: https://github.com/TM9657/flow-like

Happy to answer any questions about the architecture or local model integration.

r/ClaudeCode Far-Stretch5237

the most simple Claude Code setup i've found takes 5 minutes and gets 99% of the job done...

instead of one AI doing everything, you split it into three:

Agent 1, the Architect

> reads your request

> writes a technical brief

> defines scope and constraints

Agent 2, the Builder

> reads the brief

> builds exactly what it says

> nothing more, nothing less

Agent 3, the Reviewer

> compares the output to the brief

> approves or sends it back with specific issues

if rejected... the Builder fixes and resubmits

this loop catches things a single agent would never flag because it can't critique its own decisions (pair it with Codex using GPT-5.4 for best results)

r/AI_Agents pacifio

Turned Claude Code architecture into a high level coding agent framework to build embeddable agents for any rust projects, delivers ~7× higher throughput than Claude Code, ~2× faster than Codex

Turned Claude Code architecture into a high level coding agent framework to build embeddable agents for any rust projects, delivers ~7× higher throughput than Claude Code, ~2× faster than Codex, and achieves ultra-fast 0.098 ms recall. Skills, MCP, sessions all batteries included

Get started with

```rust use cersei::prelude::*;

[tokio::main]

async fn main() -> anyhow::Result<()> { let output = Agent::builder() .provider(Anthropic::from_env()?) .tools(cersei::tools::coding()) .permission_policy(AllowAll) .run_with("Fix the failing tests in src/") .await?;

println!("{}", output.text()); Ok(()) 

} ```

r/LocalLLaMA ghgi_

700KB embedding model that actually works, built a full family of static models from 0.7MB to 125MB

Hey everyone,

Yesterday I shared some static embedding models I'd been working on using model2vec + tokenlearn. Since then I've been grinding on improvements and ended up with something I think is pretty cool, a full family of models ranging from 125MB down to 700KB, all drop-in compatible with model2vec and sentence-transformers.

The lineup:

Model Avg (25 tasks MTEB) Size Speed (CPU) potion-mxbai-2m-512d 72.13 ~125MB ~16K sent/s potion-mxbai-256d-v2 70.98 7.5MB ~15K sent/s potion-mxbai-128d-v2 69.83 3.9MB ~18K sent/s potion-mxbai-micro 68.12 0.7MB ~18K sent/s

Evaluated on 25 tasks (10 STS, 12 Classification, 3 PairClassification), English subsets only. Note: sent/s is sentences/second on my i7-9750H

These are NOT transformers! they're pure lookup tables. No neural network forward pass at inference. Tokenize, look up embeddings, mean pool, The whole thing runs in numpy.

For context, all-MiniLM-L6-v2 scores 74.65 avg at ~80MB and ~200 sent/sec on the same benchmark. So the 256D model gets ~95% of MiniLM's quality at 10x smaller and 150x faster.

The 700KB micro model is the one I'm most excited about. It uses vocabulary quantization (clustering 29K token embeddings down to 2K centroids) and scores 68.12 on the full MTEB English suite.

But why..?

Fair question. To be clear, it is a semi-niche usecase, but:

  • Edge/embedded/WASM, try loading a 400MB ONNX model in a browser extension or on an ESP32. These just work anywhere you can run numpy and making a custom lib probably isn't that difficult either.

  • Batch processing millions of docs, when you're embedding your entire corpus, 15K sent/sec on CPU with no GPU means you can process 50M documents overnight on a single core. No GPU scheduling, no batching headaches.

  • Cost, These run on literally anything, reuse any ewaste as a embedding server! (Another project I plan to share here soon is a custom FPGA built to do this with one of these models!)

  • Startup time, transformer models take seconds to load. These load in milliseconds. If you're doing one-off embeddings in a CLI tool or serverless function its great.

  • Prototyping, sometimes you just want semantic search working in 3 lines of code without thinking about infrastructure. Install model2vec, load the model, done, Ive personally already found plenty of use in the larger model for that exact reason.

How to use them:

```python from model2vec import StaticModel

Pick your size

model = StaticModel.from_pretrained("blobbybob/potion-mxbai-256d-v2")

or the tiny one

model = StaticModel.from_pretrained("blobbybob/potion-mxbai-micro")

embeddings = model.encode(["your text here"]) ```

All models are on HuggingFace under blobbybob. Built on top of MinishLab's model2vec and tokenlearn, great projects if you haven't seen them.

Happy to answer questions, Still have a few ideas on the backlog but wanted to share where things are at.

r/AI_Agents Honest-Bed-3544

OpenClaw vs OpenViking for a business agent and is Mistral a good provider to back it?

Hey,

I’m building an AI agent for a small metal construction company, aiming to automate real business workflows not just a chatbot.

The agent will handle:

- cost estimation & quote generation

- document parsing (PDFs, specs, past projects)

- supplier communication (email-style tasks)

- internal Q&A over company files (RAG)

- potentially task orchestration across tools (CRM, spreadsheets, etc.)

I’m currently evaluating frameworks and providers, and I’d really appreciate input from people who’ve actually deployed agents in production.

What I’m considering:

- OpenClaw → seems like a full agent runtime with integrations (Telegram, etc.)

- OpenViking → looks stronger on memory/context architecture, but less “out-of-the-box agent”

- Mistral → for cloud inference (Agents API, tool calling, RAG, etc.)

---

### 1. OpenClaw vs OpenViking

From what I understand:

- OpenClaw = more “ready-to-run” agent system

- OpenViking = more infra/memory layer

Is OpenViking something you run with another framework, or can it fully replace one?

If you had to build a business-facing agent today, which direction would you go?

---

### 2. Mistral in production

I’m considering Mistral as the main provider (Large / Small models).

- Is it stable enough for real workflows (not demos)?

- How does it compare to OpenAI / Anthropic specifically for agent-style tasks (tool use, reasoning, consistency)?

- Any hidden downsides (latency, hallucinations under load, weak tool-calling, etc.)?

---

### 3. Better alternatives?

If you were building this today:

- What stack would you pick?

- Any frameworks/projects I’m missing? (especially self-hostable or hybrid setups)

Not looking for hype — I care about reliability, maintainability, and actual production use.

Thanks 🙏

r/ClaudeCode No-Word-2912

Noctis v1.1.0 is out — a free, open-source music player for your local library

https://preview.redd.it/vwiiblz3jtsg1.png?width=2554&format=png&auto=webp&s=5b6287fd4a656f42be768887abf7c64909db9f44

Features:

  • Lossless audio support (FLAC, ALAC, WAV and more)
  • Time synced lyrics with LRCLIB integration
  • Dynamic ambient album color backgrounds
  • Cover Flow view
  • Side lyrics panel see synced lyrics while browsing your library
  • Collapsible sidebar
  • Advanced EQ with presets
  • Replay Gain, gapless playback
  • Last.fm scrobbling + Discord Rich Presence
  • Drag and drop import
  • Multi select with bulk actions
  • In app updates

Windows only for now, macOS coming soon. Free and open source.

https://github.com/heartached/Noctis/releases

r/AI_Agents payfrit

Anthropic banned our organization, now what?

hello,

this past Friday afternoon Anthropic banned our organization from using their ai.

i had emailed Trust & Safety five days prior, concerned that our application was using too much resources, asking them to pls ensure what we were doing was within their usage guardrails, etc. i sent a nice explanation, a couple case studies, etc, told them exactly what i was up to.

i am looking for alternatives to connect to my project to replace the power i previously had with an unlimited Claude account. currently I have setup Ollama as our primary engine, with failovers for Gemini, then OpenAI, then OpenRouter api keys. i don't have much of a budget.

i honestly don't know what i am doing; i was just vibecoding with Claude Code for a couple months and i had created a really cool team of agents that i worked with via a Mattermost team. everyone had a lane, a set of tools (Kotlin coder, Swift coder, php backend, librarian for documents/history, etc.) and i really don't know how but my shit just worked and I was using it to produce some amazing apps and accompanying services, etc. I was exploring launching a white label version, etc.

then Anthropic banned me. and the tool just, isn't the same. i'm not sure how much sense that makes.

i guess what I am asking is, what sort of a backend can i best plug into my system, to replace what I had with Anthropic? i am cobbling something together using ChatGPT for help but it's just not the same at all.

secondly, is there any chance that someday a human will read my appeal and possibly let us use Anthropic's api again? they didn't give any reason, just refunded my previous month's payment and said we are banned.

apologies if this isn't the right place to post this, any guidance would help. thanks!

-John

r/LocalLLaMA angry_baberly

Facebook marketplace used PC upgrades/setup questions

OK I was looking at the GX 10 and then I was looking at the MacBook M5 128 GB… And I’m not super tech-handy but absolutely capable of learning.

Use case would be thinking partner/brainstorming/writing/processing some documents and stuff. I’m thinking about starting with a 70 B model or maybe the GPT 120 B open source, but honestly I wouldn’t necessarily want to limit myself.

So on Facebook marketplace I found this used gaming computer along with the a good amount of memory sticks. I think in this set up, the only additional thing I would have to upgrade would be VRAM to 24Gb+?

Can someone who knows more about this help me? Am I getting in way over my head in terms of it being complicated and potentially having to spend hours troubleshooting something? Or is it pretty straightforward?

Fb listing:

128GB DDR4 3200 ( 32GB x4 sticks) Samsung RAM Memory UDIMMs non-ECC

Fully tested and in 100% working order, willing to stress test in person if needed. Used in my video/gaming workstation (which I'm also selling with 32GB of other DDR4 if interested. i9-10900K, 2x 512 m2 SSD, 4x 8TB RAID HDD, Nvidia Quadro RTX 4000 8GB video card, 2x 10Gb nic ports, Win 11 Pro.).

Thank you in advance!

r/ChatGPT tombibbs

Nowhere near enough politicians understand what the consequences of superintelligent AI would be

r/Anthropic shanraisshan

Everything Claude Code just overtook Superpowers as the #1 most-starred ★ Claude Code workflow repo — 133k vs 132k.

r/LocalLLaMA Difficult-Expert2832

Fine tuning ocr model handwriting

I have taken a project for fine tuning an ocr model for hand writing what is the best model for fine tuning and what should be the approach I am planning to use unsloth I created pipeline to segmentation each line in one picture i have like 300k good data Letters and sentences and whole paragraphs will it be enough for make great production model any advice

r/LocalLLaMA houssineo

legal AI oriented for a specific country

Hi guys please could you tell me what is the best chunking method I can use for a law documents in order to put them in vector database

r/LocalLLaMA fulgencio_batista

Gemma 4 and Qwen3.5 on shared benchmarks

r/SideProject Shadi_TP

I built a minimalist wall art website to make spaces feel calmer — looking for feedback

I started it because I’ve always liked clean, simple designs, but most wall art online felt either too expensive or too cluttered. So I decided to make something more minimal and affordable.

It’s still early, and I’d really like to improve it based on real feedback.

I’d love your thoughts on:

First impression — does it feel clean or boring?

Is it clear what the site is about right away?

Anything that feels confusing or unnecessary?

If you’ve built something similar or sell digital products, I’d also love to hear what worked for you.

Thanks a lot — I’ll reply to every comment 🙌

r/LocalLLaMA NotumRobotics

We got tired of managing local LLMs across different hardware. Meet ClusterFlock: an open-source multi-machine llama.cpp orchestrator.

Running local LLMs across a chaotic mix of hardware (Linux boxes with consumer NVIDIA cards, Multi-GPU systems, Macs, nVidia Sparks, even Jetsons) just got us frustrated. Trying to manage them separately is tedious.

We just wanted them all to contribute behind a single endpoint, so here's: ClusterFlock.

What it does in practice:

You run the orchestrator (nCore), then run an agent on each GPU machine. The agent auto-detects your VRAM, pulls the best-fitting model from HuggingFace, spins up llama-server, and registers with nCore. nCore exposes a single OpenAI-compatible endpoint on port 1919 that routes across all your nodes.

So if you have a 3090 running Qwen3-32B, a Mac Studio running Qwen3-14B, and a 3070 running something smaller - they all sit behind one API.

  • The Mission System: Takes your tasks, picks the strongest model as a showrunner, gives all endpoints job titles and tool access, and completes your goal in a local, disposable Docker container.
  • Fan-Out Default: Your existing apps and IDEs can just hit localhost:1919/v1/chat/completions. You set a timer on how long you're willing to wait. The system picks a showrunner, endpoints munch on your query until the timer expires, and the showrunner crunches a response with the expanded context. (No streaming yet, sorry).
  • Resource Maximization: You can enable tight-fitting models and parallel processing to run multiple LLMs on a single VRAM pool. CPU processing can also be turned on via the UI (on Linux agents).

In a nutshell:

  • Uses llama.cpp (llama-server) under the hood and an LM-Studio wrapper for Windows agents (you can use the LM Studio agent on other hosts too, though we actually want to sunset LM-Studio support eventually).
  • VRAM-aware model selection with safety margins: Won't try to cram an 8B Q4 onto an 8GB card if there isn't enough headroom for reasonable context. (Word of advice: the mission runner runs best with agents that have at least an 8K context window).
  • Smart Auto-loading: Model catalog + HF Top 50 loaded on startup; easily add direct HF model links. Agents download, benchmark, and run the model. You can define a desired TPS target in the autoloader across the whole fleet.
  • Zero third-party Python deps in the orchestrator: Pure stdlib. Agents just need huggingface_hubfor model downloads to minify supply chain attacks (but hey, we're a supply chain link as well, so stay vigilant).
  • Multi-step tasks: Missions can be a one-off or infinite; your prompt decides.

What it's NOT:

  • Not tensor parallelism or model splitting across GPUs. nCore just routes and load-balances across models. In other words, it's distributed thinking, not distributed rendering.
  • Not safe for internet deployment. It's built for a single user (sort of, you'll see). Use VPNs for remote agents.

ClusterFlock is free and open source.

Organize your AI workforce or run 50 haiku models on a single GPU and have them talk to each other - zero flocks given.

(It was too good of a pun to pass up, but I actually do care a lot and would love to hear your feedback).

r/SideProject Ok_Seaworthiness1884

Day 2 — 10 users on skillvow.com

Day 2 of launching skillvow.com — we just hit 10 users.

Small number, but feels big.

People are already trading skills, connecting, and trying to build together instead of learning alone.

Early, messy, but promising.

If you believe building > consuming, would love your feedback.

r/ClaudeCode johnnyApplePRNG

The fact that I can sit down at 12 noon, and hit my Pro Max limit within 2 hours is insane...

I'm not complaining that tokens exist or that resources are finite. I get it. Reality bites.

What I find insane is that Anthropic literally won't even allow me to pay them more money in the form of a subscription model that "feels unlimited" in nature.

It's the middle of the work day.. and my only options are shelling out another $200 for another pro max plan which is overkill... or grinding my teeth into dust while I watch the token count churn at $50/million token output or whatever insane prices they charge on the API...

I am aware the API exists at an exhorbitant eye-watering cost compared to even the basic plan token costs... that is not an option.

Come on, Amodei! I'm rooting for ya!

r/LocalLLM WhisperianCookie

A little android app for using local STT models for voice typing

Hello everyone, we made Whisperian, a simple tool/app for running local STT models on android and use them as replacement to Gboard dictation, while working alongside your normal keyboard.

It took way more hours/months to make than you would think lol, to make it work across OEMs, to make the recording process crash-resilient, to make it work with a lot of different models in a standardized pipeline, this that etc. 😭 It's still a beta.

One downside is that it's closed-source currently. Idk if we will open-source it tbh. I guess you could disable internet access via VPN/Shizuku/OEM settings after downloading the models you want (or sideload them if their architecture is supported, although this isn't implemented yet).

Currently the app supports 21 local models. A philosophy we are trying to follow is to include a model only if it's the best in any combination of language/use-case/efficiency, so that there's no bloat.

Right now the app doesn't offer any information about the models and their use-cases, like I said, it's a beta, we should be adding that soon.

The local models integration is still raw and minimal, but AFAIK it's the first app to try to make multiple modern STT models be usable across apps on android, with all android limitations in mind...

Some additional features it has are custom post-processing prompts/modes and transcription history. But local post-processing isn't integrated yet, it's exclusive to cloud providers currently.

r/ClaudeCode StatusPhilosopher258

Claude is amazing for coding… but things start drifting as projects grow

I’ve been using Claude quite a bit for coding, and the output quality is honestly solid especially for reasoning through problems.

But as soon as the project gets a bit larger, I keep running into the same issue:

things start drifting.

  • I end up repeating context again and again
  • small updates introduce inconsistencies
  • different parts of the code don’t fully align anymore

Initially, I thought it was just a limitation of long chats, but it feels more like a workflow issue.

I was basically trying to keep everything in one thread instead of structuring it properly.

What’s been working better:

  • define what the feature should do upfront
  • split it into smaller, clear tasks
  • keep each prompt focused

That alone made things more stable and reduced token usage.

I’ve also been experimenting with tools like Traycer to keep specs and tasks organized across iterations, which helps avoid losing context.

Curious how others are dealing with this when working on larger projects with Claude.

r/SideProject Unfair_Traffic8159

I built a browser-first screen recorder where recordings stay on your device

I’ve been working on a screen recorder that runs directly in the browser and keeps the whole recording workflow local-first.

Site: https://screenrecorder.dev

The main idea is simple: open the site, record your screen, review it, edit it, and export it without being pushed into an upload/share flow first.

What it currently does:

- record screen, mic, and optional webcam overlay

- move and resize the webcam bubble before/during recording

- save recordings locally in browser storage

- recover interrupted recordings if the tab/browser closes mid-session

- generate thumbnails and waveforms automatically

- edit recordings in-browser with split/trim/export

- export clean recordings with no watermark

- install as a PWA if you want it to feel more app-like

A lot of tools in this space are really built around cloud storage, accounts, team sharing, and “send everything to our servers” as the default path.I wanted something more private and more direct:

- hit record

- keep the file local

- review and edit locally

- export when you’re ready

r/ClaudeCode AdFrequent4886

Usage weekly reset

Historically hasn't usage reset at 12pm EST on Thursdays? mine did not. anybody else notice this?

r/Anthropic DigiHold

Anthropic tried to clean up the Claude Code leak and accidentally nuked 8,100 GitHub repos 🤦‍♂️

r/ClaudeCode Verynaughty1620

I bet he is cheating

r/LocalLLaMA zun3surge

I got sick of local models forgetting project state, so I built a global SQLite CLI to force them to remember.

Chat history is a terrible way to store project state. If you close your terminal, or swap from Claude to a local LLaMA mid-task, your agent completely forgets what it was doing.

I got tired of the context amnesia, so I wrote Castra (a pure Go binary). Instead of trying to fix the LLM's memory, I treat the model as completely stateless.

Castra externalizes all the project state, role boundaries, and audit logs into a global castra.db file on your machine. The LLM just reads a markdown file (AGENTS.md), runs a CLI command, and the Go binary updates the database. You can swap models mid-sprint and the new model just reads the DB and picks up the next task.

Fun fact: I wrote the core engine by hand up to v1.3.0, and then got lazy and had the Castra agents build the rest of the architecture up to the current v3.1.2. which included them trying to drift and getting stuck in bureaucratic hell with Castra going "nope. here's an incident report for trying to bypass - now you have TWO things to approve."

r/homeassistant ForceEastern8595

Newb - Seriously no basic disk/mounting management?

I thought I would put HAOS with Frigate on bare metal x86? No disk management? Locked down root? About to abandon. I will poke around some more but this is a deal breaker.

r/ChatGPT RJ1776

I asked AI to write a story based on some events that happened in my life, with a funny prompt! Here it is: (Just for a Laugh)

The Funniest Story I’ll Never Admit to My Wife

It was the early 2000s — the era of frosted tips, low‑rise jeans, and the absolute peak of my confidence. My girlfriend and I were at this packed concert, the kind where everyone is sweating, yelling, and pretending they can actually hear the lyrics.

Life was good.

And then… I trusted a fart.

A rookie mistake. A fatal error. A moment that split my life into two eras: Before and After.

The second it happened, my face did that expression people make when they realize they’ve left the stove on at home. I turned to my girlfriend and said something like, “I’m gonna hit the bathroom real quick,” with the same tone someone uses when they’re about to flee the country.

I power‑walked through the crowd like a man with a secret and a deadline.

The Bathroom of Judgment

Of course the bathroom was full. Of course every sink was taken. Of course the universe wanted an audience.

I locked myself in a stall, assessed the situation, and realized I was about to perform the most humiliating one‑man show of my life.

So I did it. I cleaned out my underwear in the sink like a Victorian woman washing laundry in a river. I spot‑cleaned my pants with the desperation of someone trying to erase a crime scene. And the whole time, I could hear guys behind me going:

  • “Bro… BRO.”
  • “Ain’t no way.”
  • “Man’s fighting for his life.”

I wanted to yell, “I HAVE NO CHOICE,” but I felt like that would only make it worse.

When I finished, I put my damp pants back on — nothing says dignity like cold, wet denim — and walked out of that bathroom like a man who had seen things.

The Smell Situation

Back at the concert, my girlfriend kept sniffing the air and going:

“Ugh, someone around us keeps farting.”

And I just nodded like, “Yeah babe, wild, people are gross.”

I was praying the smell would fade before she connected the dots. Thankfully, by the time we got to the car, the scent had downgraded from “biohazard” to “mildly suspicious.”

The Emergency Shower Lie That Became a Lifestyle

The moment we got home, I sprinted to the bedroom, stripped, threw everything in the washer like it was radioactive, and jumped into the shower.

She called out, “You okay?”

And I said the line that has now become my lifelong curse:

“I just feel like I have to shower after being around so many people.”

She bought it. Completely.

And now? Now I’m stuck with it. Concerts, movies, flights, crowded restaurants — I have to come home and immediately shower and wash my clothes like some kind of germaphobe superhero.

Even after a busy dinner, I’m like, “Welp, better go decontaminate.”

All because I couldn’t admit that one time, in the early 2000s, I gambled on a fart and lost spectacularly.

And honestly… if I had told her the truth back then, she probably wouldn’t be my wife today.

r/ClaudeCode Peglegpilates

My CC buddy is super snarky and I love it.

r/ClaudeAI ctbny

recheck your work

I just thought to try this and i got back 5 corrections.

r/ClaudeAI snozberryface

I Built a Star Trek LCARS Terminal to Manage My Claude Code Setup

I’ve been using Claude Code heavily for months now. Skills, agents, hooks, MCP servers, plugins, memory files, environment variables, the whole stack. And at some point I realized I had no idea what I’d actually built. Everything lives in ~/.claude/ spread across dozens of files and JSON configs and I was just... hoping it all worked together.

So I built a dashboard. And because I’m the kind of person who watched every episode of TNG twice and still thinks the LCARS interface is the best UI ever designed for a computer, I made it look like a Starfleet terminal.

One Command and You’re on the Bridge

You run npx claude-hud-lcars and it scans your entire ~/.claude/ directory, reads every skill definition, every agent prompt, every MCP server config, every hook, every memory file, and generates a single self-contained HTML dashboard that renders the whole thing in an authentic LCARS interface.

It uses the real TNG color palette with the signature rounded elbows, Antonio typeface standing in for Swiss 911, pill-shaped navigation buttons against the black void background. If you grew up watching Picard walk onto the bridge and glance at a wall panel, you know exactly what this looks like.

The aesthetics are doing actual work tho. Every single item is clickable. You hit a skill and the detail panel slides open showing the full SKILL.md with syntax-highlighted code blocks, proper markdown rendering, headers, tables, all of it. Click an MCP server and you see the complete JSON config with your API keys automatically redacted. Click a hook and you get the full event definition. It genuinely looks like pulling up a classified Starfleet briefing on a PADD.

The Computer Actually Talks Back

You type “status report” into the input bar at the bottom of the screen and Claude responds as the ship’s computer. Calm, structured, addressing you like a bridge officer. It calls your skills installed modules, your MCP servers the fleet, your projects active missions. The system prompt turns Claude into LCARS, the Library Computer Access and Retrieval System, and the whole interaction streams in real time through a response overlay that slides up from the bottom.

I kept going. You can connect ElevenLabs for premium voice output, and the config panel lets you browse all your available voices with live audio previews before selecting one so you’re not guessing. Voice input works too, you talk to the computer and it talks back. Getting that to work as an actual conversation loop meant solving echo detection so it doesn’t hear itself, interrupt handling, mic cooldown after speech, the whole thing. It took more effort than I expected but it actually works, which honestly surprised me more than anything else in this project.

Sound effects are all synthesized via the Web Audio API, sine wave oscillators tuned to frequencies that sound right for navigation clicks, panel opens, message sends. Toggleable obviously.

The Tactical Display

The TACTICAL tab is the one that makes people stop scrolling. It renders your entire Claude Code setup as an interactive force-directed graph that looks like a Star Trek sensor display. Your LCARS core sits at the center with category hubs orbiting around it, skills in periwinkle, MCP servers in orange, hooks in tan, agents in peach, all connected by pulsing edges. A rotating scanner line sweeps around like a tactical readout and you can click any node to navigate straight to that item’s detail view.

There’s also an ENTERPRISE tab that loads a real 3D model of the USS Enterprise NCC-1701-D via Sketchfab. Full interactive, you can rotate it, zoom in, see the hull detail. Because if you’re going to build a Star Trek dashboard you don’t do it halfway.

Boot Sequence and Red Alert

When you load the dashboard you get a 3 second boot animation. The Starfleet Command logo fades in, your ship name appears (you can name your workstation in the config, mine is USS Defiant), then seven subsystems come online one by one with ascending beeps until the progress bar fills and “ALL SYSTEMS NOMINAL” pulses across the screen before the overlay fades to reveal the dashboard. I spent an unreasonable amount of time tuning those boot frequencies and I would absolutely do it again.

Five seconds after boot the system runs a health check. MCP servers offline? RED ALERT, flashing red border, klaxon alarm. Missing configs? YELLOW ALERT. Everything clean shows CONDITION GREEN briefly then dismisses. If you’re a Trek fan you already understand why this matters more than it should.

Four ship themes too, switchable from CONFIG. Enterprise-D is the classic TNG orange and blue, Defiant is darker and more aggressive in red and grey, Voyager is blue-shifted and distant, Discovery is silver and blue for the modern Starfleet aesthetic. CSS variable swap, instant application, persisted in localStorage.

Q Shows Up Whether You Want Him To or Not

There’s a Q tab where you can talk to Q, the omnipotent being from the Continuum. He’s in full character, condescending, theatrical, calling you “mon capitaine” and snapping his fingers. There’s a JUDGE ME button where Q examines your entire setup by name and delivers a devastating roast with one grudging compliment buried in the mockery.

Every couple of minutes there’s a small chance Q just appears on screen with a random quip. A red popup, a snap sound, something like “I’ve seen civilizations rise and fall in the time it takes you to write a commit message.” Then he vanishes. You can’t stop it. He’s Q.

Why This Exists

Look, the Star Trek stuff is fun and it’s what makes people share it. But there’s a real problem underneath the aesthetics.

There’s a lot going on in Claude Code. The skill system, the hook architecture, MCP server integration, custom agents, memory files, it adds up fast. But the setup is invisible. Everything lives in flat files and JSON configs scattered across your home directory. You build this complex system and then you can’t see it. You can’t browse it. You definitely can’t pull it up and show someone what you’ve built.

I open this dashboard and I immediately know where I stand. 36 skills, 12 MCP servers, 8 hooks, 4 agents. That memory file from three weeks ago is still there. That hook I thought I deleted is still active. That MCP server config has a typo in the args. And I can fix all of it right there, create new skills and agents and hooks directly from the dashboard, edit files in the browser, open them in my editor, copy commands, invoke skills. It turns Claude Code from a thing you configure and hope works into a system you can actually observe and manage.

Under the Hood

The whole thing generates one self-contained HTML file using nothing but Node.js built-ins. CSS, JavaScript, markdown renderer, syntax highlighter, chat client, voice synthesis, sound effects, force-directed graph, all inline. You can email the dashboard to someone and they open it in their browser and it works. In live server mode it adds API endpoints for chat streaming, file operations, voice synthesis, and MCP health checks, but the core dashboard runs perfectly static with zero external requests aside from Google Fonts for the LCARS typeface.

The codebase is open source under MIT and I’m actively improving it. I’ve read through the entire Claude Code source and there’s a lot more I want to do with this.

The first time I booted the finished dashboard and the LCARS panels came up with the beeps ascending and the subsystems going green one by one, I just sat there for a second. I’ve been staring at terminal output for months. Seeing the whole system laid out in that interface, the one I spent my childhood thinking was the future of computing, that hit different.

If you’re using Claude Code and you want to actually see what you’ve built, give it a try. One command, about 3 seconds.

npx claude-hud-lcars 

For the full experience with chat and voice:

export ANTHROPIC_API_KEY=sk-ant-... npx claude-hud-lcars --serve 

The repo is at github.com/polyxmedia/claude-hud-lcars. Star it, fork it, and if Q roasts your setup particularly hard I want to hear about it.

Live long and prosper.

r/ClaudeAI Master_Jello3295

I built a local memory layer in Rust for AI agents

Hey r/ClaudeAI ,

I was frustrated that memory is usually tied to a specific tool. They’re useful inside one session but I have to re-explain the same things when I switch tools or sessions.

Furthermore, most agents' memory systems just append to a markdown file and dump the whole thing into context. Eventually, it's full of irrelevant information that wastes tokens.

So I built Memory Bank, a local memory layer for AI coding agents. Instead of a flat file, it builds a structured knowledge graph of "memory notes" inspired by the paper "A-MEM: Agentic Memory for LLM Agents". The graph continuously evolves as more memories are committed, so older context stays organized rather than piling up.

It captures conversation turns and exposes an MCP service so any supported agent can query for information relevant to the current context. In practice that means less context rot and better long-term memory recall across all your agents. Right now it supports Claude Code, Codex, Gemini CLI, OpenCode, and OpenClaw.

Would love to hear any feedback :)

r/ClaudeAI ggc1808

I built an MCP server that stops Claude from ever seeing your real API keys

After seeing the Smithery vulnerability that leaked thousands of API keys,

I built mcp-keyguard: a local encrypted vault for your API keys.

How it works:

- You store your keys encrypted locally (AES via Fernet)

- Claude calls make_request instead of the API directly

- The server injects the real key server-side and returns the result

- Your keys never appear in the chat context

GitHub: ggc180820/mcp-keyguard

r/SideProject aws_architect_12

I stopped typing and started speaking — AI does the rest

Hey everyone!

I built a desktop app and I honestly can't work without it anymore.

The idea is simple: you press a hotkey, say whatever's on your mind — messy, unstructured, full of "um"s — and AI turns it into clean, polished text right where your cursor is. Not a raw transcript. Actual good text you'd be happy to send.

It works in two modes: fully offline (everything on your device, no internet) or cloud (OpenAI API for max accuracy + full AI power). You choose.

What happens when you speak

  1. Transcription — your voice becomes text
  2. AI cleans it up — removes filler words, fixes grammar, restructures sentences
  3. Formatting — AI structures the text: bullet points, headers, organized notes
  4. Web search — AI searches the internet, pulls answers from multiple sources with citations (cloud mode)
  5. Result gets pasted — directly into whatever app you're working in

One hotkey. One action. Done.

AI with internet access

This is the part that changed everything for me. Say your AI keyword + a question — and the AI goes to the internet, finds the data, formats it, and pastes a ready answer where your cursor is. ~2 seconds. No browser. No tabs. No copy-paste.

Example: "Hey Whisper, give me current S&P 500 sector performance"

You say this out loud. Two seconds later, this appears in your document:

S&P 500 — Sector Performance (April 2026)

Index trading at ~5,611. Up +8.2% YTD. Sources: Yahoo Finance, Reuters

That's it. You asked a question out loud. AI searched the web, pulled data from multiple sources, formatted it into a table with trends, and pasted the whole thing into your document. You never opened a browser.

More examples:

  • "Hey Whisper, what's new in React 19?" — structured breakdown with code examples, right in your editor
  • "Hey Whisper, compare PostgreSQL vs MySQL for read-heavy workloads" — comparison table with pros/cons
  • "Hey Whisper, what are the latest GDPR requirements for SaaS?" — summary with key points, pasted into your email draft

Text enhancement example

You say:

"so I looked into the performance issue and it turns out we were running the query without an index on the user ID column and when I added the index the response time dropped from 800 milliseconds to about 40 and I already pushed the fix to staging"

What appears in Slack:

Investigated the performance issue — root cause was a missing index on the user_id column. After adding it, response time dropped from ~800ms to ~40ms. Fix is already on staging.

Spoke for 10 seconds. Didn't think about structure. Didn't edit. AI kept the technical details, dropped the filler, made it concise.

Custom AI commands

You can tell the AI exactly what to do:

  • "Summarize in 3 bullet points"
  • "Translate to Spanish"
  • "Write a git commit message"
  • "Make this more formal"
  • "Convert to a bullet list"
  • "Put this in a table format"

Save presets for the ones you use often.

Two modes

Offline — everything runs on your machine. Voice captured in RAM, processed, immediately discarded. Nothing written to disk. Zero network traffic. Disconnect your internet — it still works. AI enhancement through Ollama (Llama, Mistral, etc.) for full privacy.

Cloud — paste your OpenAI API key, get maximum accuracy. Transcription ~$0.003/min. AI through GPT-4o. Zero markup from us — you pay OpenAI directly.

Mix and match however you want.

Works everywhere

One global hotkey — VS Code, JetBrains, Vim, terminal, Slack, Discord, Teams, Gmail, Notion, Obsidian, Word, Google Docs, Chrome, Firefox — any app with a text field. No plugins, no extensions.

90+ languages

Auto-detection. Speak in one language, get text in another. Native scripts for Chinese, Japanese, Arabic, Korean, etc.

Customization

Recording mode (push-to-talk or toggle), custom vocabulary (case-sensitive technical terms), theme (dark/light/system), AI model, custom prompts — everything is configurable.

Pricing

  • Monthly: ~$9.99/mo
  • Yearly: ~$79.99/yr (~$6.67/mo)
  • Lifetime: $99 one-time — pay once, own forever

7-day free trial. Team plans from $5/user.

System requirements

Windows 10+ or macOS coming soon.

Website: whisper.remskill.com

Happy to answer any questions. Would love to hear what you think!

r/Anthropic YungBoiSocrates

Anthropic, I really hope you plan to kickback a LOT of the usage your bugs caused, because this level of service has been unacceptable. Claude is my favorite but Codex has become my daily driver. Less enjoyable experience - but most importantly, it WORKS.

I typically recommend Claude to everyone who uses ChatGPT but at this point it'd be ironically unethical to do so at an individual level. I understand compute is constrained and you need to train bigger models, but I think you should REALLY think about user sentiment and how badly you're messing up right now as an opportunity to re-weigh where on the totem poll this issue sits.

r/AI_Agents PuzzleheadedCut1886

I spent a year building an AI agent OS for hotels — here's what actually works in production

Last April 2025, I started experimenting with AI voice and chat platforms for the hospitality industry. Not as a product — just trying to answer a question: can AI actually handle complex hotel sales conversations? Not FAQ bots. Real lead qualification, objection handling, and multilingual conversations.

After months of testing different stacks, I deployed an AI agent called MAX at a resort in St. Maarten (simpsonbayresort.com). Here's the setup:

The agent:

  • 95% chat, 5% voice
  • For now, MAX is working at 10:00 pm to 6:00 am
  • Multilingual (English/Spanish/French/Dutch/ German)
  • 24/7 availability
  • Qualifies leads, captures booking details, and pushes to the resort's in-house reservation system in real time

What it actually produces:

  • Low/mid season: 1–5 qualified leads per day
  • High season: 5–10 qualified leads per day
  • Every conversation is logged and available for human follow-up

The lesson that changed the product:

A single agent isn't the hard part. The hard part is everything around it — monitoring conversations, tracking lead quality, watching system health, and connecting to the hotel's actual booking infrastructure.

So I ended up building what I now call an "AI Agent OS" — a platform that deploys, monitors, and orchestrates multiple agents (chat, voice, messaging) from a single dashboard. Think of it as: agents are the apps, the OS runs them.

Right now it's connected to the resort's in-house reservation app. Next integration: Amadeus (the travel industry's backbone platform). Current third-party implementation is Fluenty Saas

r/ClaudeAI Witty_Implement_8943

Anyone else lose context when moving between Claude and other AI tools?

I use Claude a lot for longer coding and reasoning tasks. It works great inside a single conversation, but whenever I switch to another model or tool for a different perspective, I end up losing the context and have to reconstruct everything.

I tried summaries, notes, and bookmarking chats, but they don’t preserve the full reasoning flow.

So I built a small Chrome extension to carry conversations across AI tools instead of starting from scratch.

Claude actually helped build parts of it:

- generating the initial Chrome extension structure (Manifest v3, background + content scripts)

- writing logic to capture and format chat context from pages

- helping debug DOM extraction issues when parsing conversations

- iterating on how context should be stored and re-inserted when switching tools

The extension basically lets you export the conversation context from one AI tool and reuse it in another so you don’t lose the reasoning trail.

It’s free to try:

https://chromewebstore.google.com/detail/contextswitchai-ai-chat-e/oodgeokclkgibmnnhegmdgcmaekblhof

Curious if others run into the same context problem when switching between models.

r/SideProject Bitter-Cheek-950

I’m tired of seeing small businesses overcomplicate automation. I’m starting Day 0 of building a visual Telegram bot builder. Is this MVP logic right?

Hey everyone. I’ve noticed a massive gap recently. Most local businesses (gyms, salons, consultants) could save hours a week with simple automations, but existing tools like ManyChat or custom CRMs are either too expensive or too intimidating for non-tech folks.

I’m starting a new project today to fix this: a purely visual, drag-and-drop Telegram bot builder. I eventually want to scale it into a native Telegram Mini App.

I haven't written a single line of code yet. I’m literally just sketching out the logic on a piece of paper today to figure out the MVP.

My goal for the MVP is just 3 nodes: Trigger (User clicks link) -> Data Collection (Ask for email) -> Action (Save to Google Sheets).

If you were building this, what core feature would you add to the V1? Is there a better way to map this out before I start coding the canvas? Would love any feedback.

r/ClaudeCode Low_Engineering1740

Claude is reading your .env

DevRel at Infisical here! It always scares me when Claude Code or another agent starts reading through my repo and pulls in the .env file. I've even seen it print the contents directly to the terminal. .gitignore doesn't do anything here. Agents don't use git. I made a quick video on how we solved this at Infisical (open source secrets manager). No more secrets in files on disk. https://www.youtube.com/watch?v=zYCeELjcgQ4

r/LocalLLM carolinedfrasca

Gemma 4 is out & we benchmarked it on B200 and MI355X (15% faster than vLLM on Blackwell)

Google DeepMind dropped Gemma 4 today. Two models:

  • Gemma 4 31B: dense, 256K context, redesigned for efficiency and long-context quality
  • Gemma 4 26B A4B: MoE, 26B total / 4B active per forward pass, 256K context

Both natively multimodal (text, image, video, dynamic resolution).

Modular (folks behind MAX and Mojo) got both running on MAX on day zero, NVIDIA B200 and AMD MI355X from the same stack, no separate codepaths per vendor. On B200 we're seeing 15% higher output throughput vs. vLLM.

You can try both for free in our playground: https://www.modular.com/#playground.

r/LocalLLaMA MedicineTop5805

Using whisper.cpp + llama.cpp for real time dictation on Mac and its honestly good enough to replace cloud tools

Been running a local dictation setup on my M2 Mac for about a month now using whisper.cpp for transcription and llama.cpp for text cleanup. The pipeline is basically: speak into mic → whisper transcribes → llama rewrites into clean text.

Latency is surprisingly low. On Apple Silicon the whole thing runs fast enough that it feels real time. Text quality after the LLM cleanup pass is honestly better than what I was getting from Otter or Wispr Flow because the LLM actually restructures sentences instead of just fixing typos.

Im using MumbleFlow which wraps both into a desktop app with a nice UI. Its $5 one time so not open source but the inference is all local and you can pick your own models.

Anyone else running similar setups? Curious what model combos people are using for dictation cleanup.

mumble.helix-co.com

r/ClaudeAI Main-Confidence7777

I built an open source MCP server that aggregates 29 sports APIs into 319 tools, now on the MCP Registry

Hey everyone, I just published mcp-sports-hub to the official MCP Registry, It's fully open source (MIT) and built to be extended by the community.

Works with any LLM (Claude, GPT, Gemini, Llama, Mistral, etc.).

What it does:

A single MCP server that connects LLM to 29 sports data providers at once — scores, stats, live odds, esports, F1 telemetry, college sports, cricket, golf, and more.

Key features:

- 9 providers work instantly with zero API keys

- 20 more providers available with free-tier keys (signup takes ~2 min for free)

- Missing keys don't block startup — tools just return an error when called

- Provider presets: `free`, `us-major`, `soccer`, `f1`, `odds`, `esports`...

Install:

npm install mcp-sports-hub

Or find it directly on registry.modelcontextprotocol.io

Repo: github.com/lacausecrypto/mcp-sports-hub

Feedback and PRs very welcome — especially if you want a new provider added!

r/LocalLLM Infinite-pheonix

Google Gemma 4 Launched

r/SideProject AKi_no_ShiBaInu

Show HN: CodexSpec - Build what you want, not what AI guesses, in any language

Tired of AI speaking English while you think in your native language? Worried that saying "no" to Claude Code means losing hours of work?

I built CodexSpec to solve two critical pain points in AI-assisted development:

🌍 Speak Your Language - No more language barriers with AI coding assistants - Configure your preferred language (13+ languages supported) - Chat with Claude Code in Chinese, Spanish, Japanese, or any language you're most comfortable with - LLM dynamic translation ensures clear communication

🔒 Persistent Artifacts That Save You from "No" This is the killer feature: when working with Claude Code, have you ever: - Typed "no" and lost the entire conversation context? - Closed a tab and lost your progress? - Had to restart from scratch because of a wrong turn?

CodexSpec persists ALL your artifacts - specs, plans, tasks, decisions. Even if you say "no" or close the tab, your work is saved and can be recovered instantly.

The Spec-Driven Workflow: 1. Constitution → Define your coding standards upfront 2. Specify → Interactive Q&A in YOUR language to clarify requirements
3. Generate Spec → Auto-validated specification document 4. Spec to Plan → Technical implementation plan 5. Plan to Tasks → Test-driven development tasks 6. Implement → Execute with full context preservation

Before/After Experience: - Before: "I'll just ask Claude to build X... oh wrong direction, start over" 💸 - After: "I'll guide Claude through specifications... exactly what I need" ✅

Zero Friction Trial: ```bash

Try immediately, no install needed

uvx codexspec init my-project --ai claude --quick

Install if you love it

uv tool install codexspec ``` Why This Matters:

  • Trust: You won't lose work to accidental "no" responses or closed tabs
  • Clarity: Communicate with Claude in your native language for better results
  • Quality: Catch issues before writing code, not after
  • Control: Humans make decisions, AI executes Built from 40% rework rate pain points into a systematic solution.

Project: https://github.com/Zts0hg/codexspec Docs: https://zts0hg.github.io/codexspec/

MIT Licensed • Open Source

r/SideProject GeorgeRo_GG3

App Opinion

I’ve been building a small rewards app and I’d really appreciate some honest feedback.

It’s called GeoRewards — users can earn points by completing offers (surveys, apps, etc.) and then cash out. I’m still in the testing phase, so I’m not trying to promote it, just looking to see what feels confusing or broken.

Everything is already working (points tracking, activity history, cashout system), but I want to make sure the experience actually makes sense for real users.

If anyone has a minute to try it and share feedback, I’d really appreciate it 🙏

Link: https://geo-reward-wgrown.flutterflow.app/

Even small comments help (design, bugs, flow, anything).

Thanks!

r/homeassistant Dump_name

Doorbell with video stream for Home Assistant

Hey everyone,

I’m in the middle of planning the electrical setup for the house in Germany that I am currently renovating and could use some advice. I’m going for a KNX smart home with Home Assistant as the visualization server and for some minor automations, but all core functions (lights, shades, heating, doorbell) should work independently — no network or Home Assistant needed for basic operation.

For my setup I am searching for an ideal solution regarding the doorbell. I’ve attached an image showing the three variants I’m considering.

The plan:

  • Install Android or Linux panels on each floor (model selection is still pending)
  • Panels show Home Assistant and pop up the doorbell camera when someone rings (via browser_mod)
  • Doorbell should function fully with standard audio indoor units (talk, ring, door unlock) since I don’t want two panels per floor

Here are the options I’m thinking about:

Variant 1:

  • Video doorbell with an IP gateway
  • Fetch video stream directly into Home Assistant
  • Ringing and door-open triggers also come via the gateway

Variant 2:

  • Same as Variant 1 but doorbell triggers are connected to Home Assistant via Doorbell actuators/sensors and KNX actuator/sensor or an ESP32.

Variant 3 (fallback):

  • Use a standard audio only doorbell
  • Install an external PoE camera at the door canopy
  • Use KNX/ESP32 for the triggers (As in variant 2)
  • Display the camera feed in Home Assistant

I’d really prefer a doorbell with an integrated camera — the angle is just better (and my wife agrees). I’m not a fan of the Reolink/Unifi/Ring look because it feels like an afterthought. Doorbird looks fine, but it’s network-only and seems to only c video indoor stations.

So my questions are:

Are there any manufacturers or models you’d recommend for this setup? Does anyone here run a similar configuration? Any doorbell brands with solid Home Assistant integration?

Thanks in advance for any tips!

r/ClaudeAI 2blazen

Why does Sonnet Extended change its mind in the middle of an answer?

I've recently switched to Claude and I'm running into this strange behavior quite often. It's confusing and it's starting to make me lose trust in its answers, so I'd appreciate feedback from long-time users. Is this normal? Am I doing something wrong?

Just in case, this is my system prompt: Always be more concise with your answers. I want short queries answered in a very brief, one-paragraph manner. Do not restate or appreciate what I say. Be as factual as possible. Tell it like it is; don't sugar-coat responses. Give unbiased, objective responses that are as clear as possible based on facts. Ask for any additional context needed. If I ask for recommendations, take ratings and number of ratings into consideration. Don't recommend products with too few ratings, or too bad of a rating.

r/SideProject Aepexer

Just launched my first app, would love honest feedback

Hey everyone,

I just released my first app on the App Store, it’s a trivia game called trivia plus.

I started it as a small side thing that i worked on until it could turn into something I would potentially want to ship but it’s been a process lol. I gained a fair amount of TestFlight feedback but now that it’s live, I’m already noticing some things I didn’t catch before, like how the light mode feels so plain compared to dark mode and I’m not sure how to fix that yet.

If anyone has a minute to check it out and tell me what feels off (first impressions, Home Screen layout, gameflow, anything) I would genuinely appreciate it.

Not trying to promote hard, just want to make it better and gain some honest feedback.

https://apps.apple.com/us/app/trivia-plus/id6760320473

r/ClaudeCode pladdypuss

PSA - Claude Code Bug and Overages; detailed insight. update now to cc 2.1.90

Here is what claude code said about claude code overages on my account when i prompted it to dig into the overage.

tl;dr: i was getting billed for 2,206x actual usage. Cladue Fin agent refusing to credit back the overcharge. On 20X Max plan. ACTION: update cc cli and VS code extension to at least Claude Code CLI │ 2.1.90

Email sent to Antropic that was refused refund. US user.

Hi Anthropic Support,

I'm writing to request a usage credit for token inflation caused by
the prompt cache bug publicly acknowledged by your team the week of
March 31, 2026.

Account: [XXXXXX@XXX.XXX](mailto:XXXXXX@XXX.XXX)
Plan: Claude Code Max 20x
Affected window: March 31 – April 2, 2026 (current weekly billing period)
Impact: ~20% of weekly budget consumed, primarily from inflated cache tokens

---

Evidence from my local session logs (~/.claude/projects/):

Token type Count
-----------------------------------------------
Input tokens 227,640
Output tokens 2,178,819
Cache read tokens 1,506,539,247 ← inflated
Cache creation tokens 65,368,503 ← inflated

My meaningful work (input + output) totals ~2.4M tokens. My cache
tokens total 1.57 billion — a 2,206x inflation ratio. This is
consistent with the broken cache behavior described in your team's
public acknowledgement and GitHub issue #41249: attestation data
varying per request breaks cache matching, causing full context
re-billing every turn.

Versions running during affected sessions: 2.1.83 and 2.1.87 — both
prior to the fixes shipped in 2.1.84, 2.1.85, 2.1.86, and 2.1.89. My
sessions also use ToolSearch extensively, which v2.1.84 specifically
identified as breaking global system-prompt caching.

I am now on v2.1.90 and expect normal cache behavior going forward.

Given Anthropic's public acknowledgement of this issue and the clear,
quantified evidence of inflation in my session data, I'd appreciate a
full or partial credit restoring the affected portion of this week's
budget.

Happy to share raw session logs if helpful.

Thanks,
Davis

r/SideProject Phantasm0006

I got so tired of PMs nagging me to update Jira, so I built an Slack bot to do it automatically.

Hey everyone,

I'm a highschool senior and a solo developer, and my biggest pain point in my workflow has always been context switching. I'll finish a ticket, mention it in our standup channel in Slack, and then 3 hours later my EM will DM me saying "Hey, did you move that ticket in Linear/Jira?"

So, I spent the last few weeks making NudgeBot ([https://nudgebot.xyz\](https://nudgebot.xyz)).

What it does:

Invite it to your dev team's Slack channel. If someone says something like "I just finished fixing the login portal", it'll automatically move it to "Done" in Jira, Linear, Trello, or Notion using Claude 3.5.

It'll even leave an audit comment saying who moved it in Slack.

I just made it live, and I'm looking for some brutal feedback. Would you guys even use something like this in your team, or is it too scary an idea to give your PMs an AI that'll automatically close your team's tasks?

Let me know what you think! I even made it so you can get a free trial of the premium version if you're interested in using it in your team. Remember to DM me your email if you want to try to premium version!

r/ChatGPT ievkz

Life hack: save $150 a month on vibe coding with top models

I think by now everyone has noticed the same pattern: the big players in the market - Codex, Claude Code, and GitHub Copilot / Copilot CLI - pull you in with dirt-cheap entry subscriptions for $10–20 a month so you’ll give them a try, get hooked, and start relying on them. Then, once you’re already used to it and start hitting the limits, they either push you toward a $100–200 plan or try to sell you an extra $40 worth of credits.

Of course, I’m not speaking for everyone, but I use coding agents in a very specific way. These are my rules:

  1. I clear chat history almost before every prompt to save tokens.
  2. I never ask an agent to do a huge list of tasks at once - always one isolated task, one problem.
  3. In the prompt, I always point to the files that need to be changed, or I give example files that show the kind of implementation I want.

So in practice, I honestly do not care much which AI coding agent I use: Codex, Claude Code, or GitHub Copilot / Copilot CLI. I get roughly the same result from all of them. I do not really care which one I am working with. I do not trust them with huge complex task lists. I give them one isolated thing, check that they did it right, and then commit the changes to Git.

After a while, once I got used to working with agents like this, I took it a step further. At first I was surprised when people said they kept several agent windows open and ran multiple tasks in parallel. Then I started doing the same thing myself. Usually an agent spends about 3–5 minutes working on a task. So now I run 3 agent windows at once, each one working in parallel on a different part of the codebase. In effect, I have 3 mid-level developer agents working on different tasks at the same time.

Anyway, back to the point.

Because "God bless capitalism and competition", here is what you can do instead of paying $40 for extra credits or buying a $100–200 plan: just get the cheapest plan from each provider - Codex for $20, Claude Code for $20, and GitHub Copilot / Copilot CLI for $10. When you hit the limit on one, switch to the second. When that one runs out too, switch to the third.

So in the end, you spend $50 a month instead of $100–200.

How much do you really care whether one is 10% smarter or better than another? If you are not using them in a "hand everything over and forget about it" way, but instead as tools for small, controlled, simple tasks, then it does not really matter that much.

Who else has figured out this scheme already? Share in the comments )))

r/SideProject GabrielMSharp

Made an app to help my neurodivergent mind keep up on peoples details and occasions

My problem:

I'm pretty sure I should know where X works or what his third kid's name is, but when weighing up starting a conversation about it I'd go blank. bit of anxiety bit of something on the spectrum.

Solution:

Now I just save them in the app and before I hang out with people I glance to get a quick refresher.

Add notes, or specific facts / occasions. Even homescreen widgets and private AI summaries.

I'd love any feedback!

If you exceed the free limit of Contacts/Notes (which should take a little while to hit if using it organically) use code `MMFREEMONTH`

r/ClaudeAI MG-4-2

Claude - tried to kill me

Asked it how to clean my water cooler. Told me to add white wine vinegar and then bleach. Good job I know that’s not a good idea.

Surprised this is still a thing with Claude I thought this stuff stopped a long time ago in the 3.5 days of chat gpt.

Edit - I'm not gonna share the full conversation because it would dox me. Be assured I've been using large language models for the last three years extensively. I understand the garbage in, garbage out problem. My usage today was completely normal, a simple question. Someone highlighted it below that the key here was the sequencing. It told me to clean it with vinegar, rinse it, and sanitize it with bleach. Now if I had rinsed it very well, that wouldn't be a problem. If I didn't know that mixing vinegar and bleach was a problem, I probably wouldn't have considered the necessity to make sure that all of the vinegar residue was removed. That's the problem. Is my title hyperbolic? Yes. Do I think it was trying to kill me? No. Do I think that for someone that didn't know that vinegar and bleach made chlorine gas, that this could have been an issue? Yes.

r/ChatGPT OpinionSpecific9529

After much consideration 😑 Was one of those users who was using it since it was released!

r/AI_Agents ravenrandomz

AI-Hardened ARG: Challenge to Reddit and AI-worshippers

My Robotics team is migrating from Github. We left an ARG in it's place. Some of the ciphers have traps for AI.

I had some related teams attempt to solve the puzzles with AI, and I adjusted them. I'm curious if anyone is able to use either AI and/or agetic AI.

There are no know intentional malicious prompts. Only prompts attempt to causes AI to not provide useful responses. As AI can have erratic glitces, I would recommend running in a sandbox. If the AI goes off course, it could reach places beyond the control of the ARG.

r/ClaudeAI WhereIsMySun

Claude is predicting my answers to my face

I love Claude and have upped my coding skills with it but this freaked me out lol. I use Claude as a macro tracker as I'm trying to gain weight and it tried to predict my answer based on past responses. Idc much but this begs the question, how reliable are future answers? Is it feedback looping itself into straight bias?

r/LocalLLaMA zelkovamoon

Have you tried this -> 2x Modded 2080 ti 22GB with Nvlink

Not looking for "that card is old" or "no warranty" takes - I just want to know, for those of you who like to walk on the wild side has anyone done this?

I've done some deep research queries into running nvlink on these modded cards, and haven't found much of anything - it could be that they just missed it. But, if we can get 50GB/s symmetrical links and 44GB of memory pooled, that could be a big deal for my use case.

If you have tried the above, or if you know definitively if it works / fails, please elaborate.

r/aivideo moisterios

The cutest stories are the ones you don’t plan

r/ClaudeAI aliards

Can I use my company's Claude Team account on my personal computer?

My company recently invited me to join their Claude Team plan and I'm wondering about the best way to access it. I'd like to use it on my personal computer by logging in with my company email. is that allowed or does it have to be on a work device? Has anyone else set this up this way? Any concerns around privacy data or company policy I should be aware of before doing this? Would appreciate any insight from people who've been in a similar situation!

r/LocalLLaMA RoutineRepulsive4571

meshllm - pool compute to run local models

With LLMs getting larger, serving them at a good TPS become non-trivial.

meshllm solves this.

Built by the team at Block the idea is to pool compute to run open models.

The models that benefit the most are MOE based ones, as few selected experts are needed at a time.

One thing I was concerned about was privacy, but it is already taken care of my providing a token auth and private mesh.

https://docs.anarchai.org/#

r/aivideo Digital_Dreamer_AI

Baby Podcast 🎙️🎧

r/SideProject AgencyInside407

I trained a language model from scratch for a low resource language and got it running fully on-device on Android

r/SideProject bedo00

I made a free stamp card maker for small businesses

I'm working on Revio, a digital loyalty card platform for coffee shops, salons, restaurants, that kind of thing.

Just shipped a free tool where you pick your business type, set your colors and reward, and it shows you what a digital stamp card would look like in Apple Wallet and Google Wallet. No account needed.

I built it because most small businesses around me still hand out

paper punch cards, and apparently, about a third of those get lost

before the customer finishes them. Figured I'd let owners see the

digital version before asking them to pay for anything.

Next.js, React, Apple, Google Wallet API.

reviocard.com/tools/stamp-card-maker

-- feedback welcome, especially on the customization flow.

r/Anthropic qodeninja

ClaudeCode 2.1.90 has the resume cache bug fixed in NPM. Not clear on other tooling

r/AI_Agents CapitalShake3085

Agentic RAG: Learn AI Agents, Tools & Flows in One Repo

A well-structured repository to learn and experiment with Agentic RAG systems using LangGraph (fully local).

It goes beyond basic RAG tutorials by covering how to build a modular, agent-driven workflow with features such as:

Feature Description 🗂️ Hierarchical Indexing Search small chunks for precision, retrieve large Parent chunks for context 🧠 Conversation Memory Maintains context across questions for natural dialogue ❓ Query Clarification Rewrites ambiguous queries or pauses to ask the user for details 🤖 Agent Orchestration LangGraph coordinates the full retrieval and reasoning workflow 🔀 Multi-Agent Map-Reduce Decomposes complex queries into parallel sub-queries ✅ Self-Correction Re-queries automatically if initial results are insufficient 🗜️ Context Compression Keeps working memory lean across long retrieval loops 🔍 Observability Track LLM calls, tool usage, and graph execution with Langfuse

Includes: - 📘 Interactive notebook for learning step-by-step
- 🧩 Modular architecture for building and extending systems

👇 GitHub Repo in the comment below

r/ClaudeCode N3TCHICK

Is it just me, or is Claude Code v2.1.90 unhinged today??

  • aggressive context compaction (yes, I'm using 1M context) resulting in terrible, and sequential agent work (it doesn't seem to want to invoke agent teams today without constant kicking... and then forgets to check on said team, which is failing)
  • trying to take shortcuts at every stage of my plan (yes, I have hooks... thankfully)
  • generally being stupid (what on earth is going on today??)
  • the window is so aggressively being compacted, that I can't see the history for more than a few lines of output at a time before it disappears?

I'm so fed up today! What on earth is going on? And of course, I now have to roll back a ton of work because agent teams kept failing for no reason at all - can't find a root cause, even with Opus 4.6 on Max thinking. The model just has no idea why this is all happening.

And to top it off, because I'm in the heavy token period, this work that is total garbage, is coming off my weekly rates at aggressive rates, with no quality output to show for this extreme token use. YAY.

I need to go outside. This is nuts today. I'm going to have to roll back to 2.1.87 I guess, or earlier.

r/SideProject Financial-Muffin1101

3 weeks after launch: 180 MRR, working 12h a day, and honestly wondering if I'm scaling too fast

Hey folks,

Just hit $180 MRR after only 3 weeks since launching.
On one hand, it's paying off — real paying users are coming in, feedback is solid, and I'm seeing clear progress every single week.

On the other hand… onboarding each new client is taking way more time than I expected. I'm currently working ~12 hours a day just to keep up, and I'm starting to wonder if I'm pushing too hard too soon.

I'm happy with the traction, but this "scaling" phase feels exhausting. Anyone else been in this spot early on?

How did you handle the onboarding bottleneck when things started moving? Did you slow down deliberately, automate more, raise prices, or just grind through it?

Would love to hear your experiences — especially from other micro-SaaS founders who went from $0 to a few hundred MRR quickly.

r/homeassistant fanofmets12

Tapo Camera Link Integration gone

I have a Tapo C120 camera that I had setup with HA and it has been working since initial setup months ago.

This morning it is no longer available at all. I see this message when trying to see the camera.

"This entity is no longer being provided by the tplink integration. If the entity is no longer in use, delete it in settings."

I used it yesterday all afternoon without issues. This morning I did the update of HA to 2026.4.0 from 2026.3.4. I thought that might be the issue restored from backup and still broken.

Do I have to just remove the device and try adding it again?

The app works fine, I can access the device locally or remotely when not at home.

r/LocalLLaMA -p-e-w-

p-e-w/gemma-4-E2B-it-heretic-ara: Gemma 4's defenses shredded by Heretic's new ARA method 90 minutes after the official release

Google's Gemma models have long been known for their strong "alignment" (censorship). I am happy to report that even the latest iteration, Gemma 4, is not immune to Heretic's new Arbitrary-Rank Ablation (ARA) method, which uses matrix optimization to suppress refusals.

Here is the result: https://huggingface.co/p-e-w/gemma-4-E2B-it-heretic-ara

And yes, it absolutely does work. It answers questions properly, few if any evasions as far as I can tell. And there is no obvious model damage either.

What you need to reproduce (and, presumably, process the other models as well):

git clone -b ara https://github.com/p-e-w/heretic.git cd heretic pip install . pip install git+https://github.com/huggingface/transformers.git heretic google/gemma-4-E2B-it 

From my limited experiments (hey, it's only been 90 minutes), abliteration appears to work better if you remove mlp.down_proj from target_components in the configuration.

Please note that ARA remains experimental and is not available in the PyPI version of Heretic yet.

Always a pleasure to serve this community :)

r/SideProject toni_btrain

Just published my first app! I can't believe it 😭😭

Approval took less than 24 hours from submission. I thought this would take like weeks and multiple attempts 😅

So this is a pretty big moment for me. I always struggled with neurodivergence and never fit in any jobs (just quit again recently). Building this app has been a great outlet for me, and a great creative exercise.

What do you guys think of the screenshots and the style? Would appreciate any feedback.

Anyway, the app is named Lacuna and it's basically about sending messages ("time capsules") to your future self or others. Letters, photos or voice notes.

These messages CANNOT be opened until the time comes. The wait is the entire point of the app.

Send someone a message or a photo to New Year's 2030, why not? Or to your future self asking if you achieved everything you wanted.

The app is free to use. There is a one-time lifetime payment which is pretty cheap.

r/ClaudeCode Efficient-Cause9324

Knew they were gaslighting everyone with the daily limits.

r/homeassistant RhombusDev

[PROJECT] EdgeCaster — Open Source SRS → RTSP converter running on a Raspberry Pi

Sharing a project I've been building out that I think will be useful for a lot of people in this community, especially those integrating Rhombus cameras into third-party stacks.

The problem: Rhombus uses Secure Raw Streaming (SRS) instead of RTSP — H.264 over HTTPS with end-to-end encryption. It's the right call from a security standpoint. RTSP is unencrypted, exposes passwords in plain text in the URL, and has no place in a serious physical security deployment. That said, most NVRs, VMS software, analytics platforms, and home automation tools (looking at you, Frigate + Home Assistant users) still speak RTSP natively. That creates a compatibility gap.

What EdgeCaster does: Runs on a Raspberry Pi, authenticates to the Rhombus API, ingests the Secure Raw Stream from your camera, and pushes a local RTSP stream out on your LAN. The upstream leg stays encrypted. The RTSP stream never leaves your local network. You get the interoperability where you need it without gutting the security model on the Rhombus side.

Common use cases:

  • Frigate NVR + Rhombus cameras
  • Home Assistant integration
  • Blue Iris / Milestone / other VMS platforms
  • Third-party AI/analytics pipelines that require a local RTSP endpoint
  • Legacy NVR coexistence during phased Rhombus rollouts

Links:

Pi 4 (4GB) is the recommended target hardware. Pi 5 if you're running higher res or multiple concurrent streams.

Drop questions here or jump into the community thread linked above — that's where the main discussion is happening.

r/SideProject JoaoRochaOnReddit

I built a free tool to plan PTO with your partner (no more messy spreadsheets)

My wife and I both work full-time. Different employers, different PTO allowances, different public and regional holidays.

Every January, we sit down and try to plan our entire year off together (when to travel, when to save days, how to avoid burning out by October with zero days left).

For years, we've done this in a Google Sheet. It works, but it's painful. Manually counting days, cross-referencing two calendars, trying to figure out when we're both off at the same time.

So I built Time Off Calendar, a free tool that makes this actually simple. It's live now and you can use it today.

What it does

Yearly calendar with side-by-side view for both partners

Add people to your calendar (set time allowance, holiday location, custom weekend days, and color)

Edit existing people when things change (time allowance updates, new holiday location for remote workers)

Customize public holidays (mark days as workdays, add custom regional or public holidays)

Mark days as full day off, half day off, or given day (birthday, company office closed)

See overlap days instantly and track each person's PTO balance separately

🧩 What it's NOT

Not another calendar app. Google Calendar buries time off among meetings and reminders.

Not HR software. No approvals, no managers, no corporate workflows.

Just a clean time-off calendar for two people coordinating their year.

🪴 Some validation

I posted the spreadsheet version on r/Adulting a few weeks back. Got 35 upvotes and 35 comments. About 75% of people said they actively plan their time off for the whole year.

The most common pain? Coordinating with a partner or family.

One comment that stuck with me:

"I love the visualization of your tool here: this makes planning for 2 much easier!"

That's when I knew this was worth building properly.

🔧 Where I'm at

The app is live and free to use (Next.js, Supabase, the usual stack). I'm a frontend engineer with 9 years in.

Now that it's out there, I'm curious:

Does the couple-first angle resonate?

What features would make this a must-have vs. nice-to-have for you?

I'd love to hear:

  1. Does this solve a real problem for you and your partner?
  2. What would make you actually switch from your current method (spreadsheet, shared calendar, winging it)?
  3. Any must-have features I'm missing?

Thanks for reading. Happy to answer questions about the build or the idea.

🔗 timeoffcalendar.com (free to use, with signup required to explore)

r/SideProject Cold-Face280

I built PureClip, a free tool to download TikTok videos without watermark

Hey! I made a small web app called PureClip that lets you download TikTok videos without the watermark. It's free, no login needed, and I don't store any data. I plan on adding support for TikTok stories, IG reels and stories, and Facebook videos.

I made it because the existing tools are full of ads and get clunky sometimes.

Check it out here: https://pureclip.vercel.app

Let me know what you think!

r/SideProject Accurate-Screen8774

File Sharing By QR Code

https://chat.positive-intentions.com/#/qr
https://github.com/positive-intentions/chat/blob/staging/src/components/pages/shareQr/ShareQR.jsx

tldr; the app is using `react-qr-code` and `react-qr-scanner`. when importing a file in javascript, it's possible to to serialize and split the data. this can be used to create a series of QR codes. if you scan all the QR codes, you should have the metadata needed to reconstruct the file locally from memory.

i was testing something out and wanted to share "file sharing by QR code". its possible to create javascript-based QR reading/writing functionality.

i created an experimental proof-of-concept of the idea. this is not a very practical way to transfer files... it is slow and mote TLC is needed for it to be "nice to use"... but the functionality is testable.

let me know what you think could be a useful was to use this. while, it becomes a pain to use this method for transferring files as large as 5kb, but its a way to exchange complex payloads offline.

r/SideProject Kurisu2026

I built a kanji app with illustrated mnemonics — try 3 kanji, no signup needed

Each kanji gets a picture + memory trick to remember the reading. It's free and covers 625 kanji (JLPT N5 to N3).

mokapp.me

r/homeassistant tvosinvisiblelight

OPNSense Firewall Dashboard Intergration with Home Assistant Web Page

HA Friends,

I have Home Assistant communicating to OPNSense through API and can see the entities.
All is good there.

Scoured the internet for pre-defined templates for HA/OPNSense integration and not too much out there. Only thing that I see is how to setup/integrate with HA etc, etc which I have already completed.

I am completely new to HA so question is can I just pull the Dashboard of OPNSense into HA w/o having to design the dashboard UI un HA? When I tried adding the dashboard as a webpage HA had problems connecting. I think it has to do with MFA that is needs in order to connect.

This is what I would like to see in HA

https://preview.redd.it/wrmxkgqr7tsg1.png?width=1213&format=png&auto=webp&s=37bb78064b1fd8e698526b3a40d6579cb2dde3e6

r/ClaudeCode ibuildoss_

Lumen plugin indexes codebases (treesitter + ast) achieves up to 50% token, wall clock time, and tool use reduction in SWE-bench tasks with embedding via Ollama

I wrote Lumen initially to help me work in a large monorepo, where Claude kept brute-forcing guesses for grep/find. Turns out, it actually reduces wall time, tokens, and tool use because it gives Claude the context it needs immediately, even if Claude isn't able to one-shot what it's looking for.

r/SideProject MedicineTop5805

Made a free Mac video downloader because every website option is basically malware now

I got tired of going to random video downloader sites that are full of popups and fake download buttons. Some of them even try to install browser extensions or redirect you through 10 pages of ads.

So I built ClipYank. Its a Mac app where you paste a link and it downloads the video. Thats it. No account, no ads, no weird installer. Works with YouTube, TikTok, Twitter, Instagram and most other platforms.

Its completely free. Built it mostly for myself but figured other people might want it too.

clipyank.xyz

r/ClaudeAI Awkward_Ad_9605

I scanned 10 popular vibe-coded repos with a deterministic linter. 4,513 findings across 2,062 files. Here's what AI agents keep getting wrong.

I build a lot with Claude Code. Across 8 different projects. At some point I noticed a pattern: every codebase had the same structural issues showing up again and again. God functions that were 200+ lines. Empty catch blocks everywhere. console.log left in production paths. any types scattered across TypeScript files.

These aren't the kind of things Claude does wrong on purpose. They're the antipatterns that emerge when an LLM generates code fast and nobody reviews the structure.

So I built a linter specifically for this.

What vibecop does:

22 deterministic detectors built on ast-grep (tree-sitter AST parsing). No LLM in the loop. Same input, same output, every time. It catches:

  • God functions (200+ lines, high cyclomatic complexity)
  • N+1 queries (DB/API calls inside loops)
  • Empty error handlers (catch blocks that swallow errors silently)
  • Excessive any types in TypeScript
  • dangerouslySetInnerHTML without sanitization
  • SQL injection via template literals
  • Placeholder values left in config (yourdomain.com, changeme)
  • Fire-and-forget DB mutations (insert/update with no result check)
  • 14 more patterns

I tested it against 10 popular open-source vibe-coded projects:

Project Stars Findings Worst issue context7 51.3K 118 71 console.logs, 21 god functions dyad 20K 1,104 402 god functions, 47 unchecked DB results bolt.diy 19.2K 949 294 any types, 9 dangerouslySetInnerHTML screenpipe 17.9K 1,340 387 any types, 236 empty error handlers browser-tools-mcp 7.2K 420 319 console.logs in 12 files code-review-graph 3.9K 410 6 SQL injections, 139 unchecked DB results

4,513 total findings. Most common: god functions (38%), excessive any (21%), leftover console.log (26%).

Why not just use ESLint?

ESLint catches syntax and style issues. It doesn't flag a 2,557-line function as a structural problem. It doesn't know that findMany without a limit clause is a production risk. It doesn't care that your catch block is empty. These are structural antipatterns that AI agents introduce specifically because they optimize for "does it work" rather than "is it maintainable."

How to try it:

npm install -g vibecop vibecop scan . 

Or scan a specific directory:

vibecop scan src/ --format json 

There's also a GitHub Action that posts inline review comments on PRs:

yaml

- uses: bhvbhushan/vibecop@main with: on-failure: comment-only severity-threshold: warning 

GitHub: https://github.com/bhvbhushan/vibecop MIT licensed, v0.1.0. Open to issues and PRs.

If you use Claude Code for serious projects, what's your process for catching these structural issues? Do you review every function length, every catch block, every type annotation? Or do you just trust the output and move on?

r/SideProject krishnakanthb13

Transcriber v0.0.11: The Ultimate Cross-Platform Audio Transcription Engine is Live! 🚀

Hey everyone! 🌟

I wanted to share a project I've been working on to solve a personal pain point: transcribing long audio files quickly and without context-switching.

Transcriber is a unified transcription tool that gives you three different ways to handle your audio—all sharing a single, robust core engine:

  1. OS Native Right-Click: You can transcribe directly from your file explorer. I've implemented registry-based context menus for Windows, Nautilus scripts for Linux, and Automator Quick Actions for macOS.
  2. Modern Web UI: A FastAPI-powered app with a "glassmorphism" aesthetic. It handles background jobs asynchronously, so you don't have to stay on the page.
  3. CLI: For those who live in the terminal, the transcribe command is colorful, supports JSON outputs, and integrates with any script.

The "Infinite" Duration Challenge: Groq's API has a 25MB limit. To solve this, I built a ChunkPlanner that automatically splits files into manageable segments using pydub, processes them sequentially, and merges the text back into a single, timestamp-safe .txt file.

Key Tech Stack: - Backend: Python, FastAPI, Uvicorn - AI: Groq Whisper API (whisper-large-v3) - Processing: Pydub, FFmpeg - UI: Glassmorphism HTML/CSS

Check out the source code and documentation below: https://github.com/krishnakanthb13/transcriber

I'd love to hear your thoughts on the OS-integration approach!

r/LocalLLaMA Routine-Example927

Raspberry Pi’s AI HAT+ 2 reality check

I finally got around to properly testing Raspberry Pi’s AI HAT+ 2 (I'm calling it Pi GenAI Hat for the rest of this post).

bit of background first - when it launched the reviews were all over the place, some people straight up called it an AI decelerator. Hailo even had to publish a blog post explaining when you'd actually want to use it, which Raspberry Pi then reposted (https://www.raspberrypi.com/news/when-and-why-you-might-need-the-raspberry-pi-ai-hat-plus-2/). Anyway I decided to verify their actual numbers one by one instead of just parroting the marketing materials.

Hailo claims it's good at three things:

fast time to first token - tested this with hailo-ollama vs regular ollama, measuring second inference call to avoid the model loading overhead skewing things. Result is that it's actually slightly slower. around 30ms, not the end of the world, but still not what was advertised

faster encoder for VLMs - this one would actually matter for stuff like home surveillance, video captioning, smart search etc. Tested image input TTFT using Qwen2-VL-2B (only VLM available for it right now btw). CPU ollama did it in ~700ms, the hat took ~900ms. Also the Hailo API gives you basically zero visibility into what's happening internally which is annoying when you're trying to actually benchmark anything

large prefill / long context - tool calling, function descriptions, that kind of thing. very large difference, even tested with simple curl command to eyeball the timing, in case bash scripts I used for benchmarking had issues. still 1890 vs 333 ms., NOT in Pi GenAI Hats favour.

the claims don't really hold up. For $130 (more with the latest price hike?) + Raspberry Pi price in 2026 you can get similar perf from Jetson Orin Nano or Arduino Ventuno Q, and somehow Rockchip 3588 with its 6 TOPS beats it in practice despite losing on paper - even for 3576 benchmarks are a bit better

if you're already all-in on the Pi ecosystem maybe still worth it just to free up CPU cores. for Computer vision applications it's still a good fit, but unless you have a lot camera streams to process simultaneously, you can just use the first gen of the Hat.

https://reddit.com/link/1salnay/video/cu9l46rmwssg1/player

r/ClaudeAI 10c70377

Anyone else running into an issue where they have multiple Claude Code sessions open in VS Code and then your computer crashes and you lose all those sessions? You have to do the tedious task of opening Claude -- resume and trying to find the sessions that you were using.

It's happened to me twice now, and when you are working across multiple projects - man its a pain. Is there like a solution that anyone is using out there that could help me?

r/homeassistant RgrimmR

My Mmwave Radar -1. Not sure what to do with it yet. It works perfectly at home.

35x35mm Works up to 6m. My stairs are what I consider normal and it reaches the bottom and paired with another when you reach the bottom half of the steps the downstairs light comes on.

r/ProgrammerHumor PresentJournalist805

broCouldntYouJustUseOneFormatAsNormalHuman

r/LocalLLM someone_random09x

[ Removed by Reddit ]

[ Removed by Reddit on account of violating the content policy. ]

r/SideProject glennbech

I created pastewaves.com an audio clip sharing platform -think pastebin for audio

https://reddit.com/link/1san88g/video/125qxr656tsg1/player

Looking for some honest feedback here- and also for customers, have to be optimistic right?! :) - Promo code at the bottom of the reddit.

pastewaves.com takes the friction out of audio clip sharing, you drag-and -drop an audio file into the web pages, it gobbles it up, and gives a link you can send to people. When they open the link, a player opens up immediately. As a hobby music production & synthesizer nerd, I wanted something simpler than Soundcloud just for sharing my stuff with friends.

To be nice, I pay for a 1 TB pool of storage, files expire after 7 days, if you want permanent storage for your files you pay $5/month for 25GB.

AI has sped things up, and I've built the solution myself, the core was done in a weekend, and I've been adding features gradually, and finally now plugged in Stripe and payments for a "pro" tier.

The solution was dormant for about 4 months, I was just using it myself- but I have seen som organic growth over the last month, I got 100 users by offering the "pro" plan for free, about 50 is till active the last 30 days. The number of uploads are picking up.

I made this cheesy as hell AI video to demo it, for my product hunt launch which gave about 0 engagement...

So, what do you think - any feedback welcome! Promo code: HUNT10 - Sign in and apply it on the profile page for free permanent pro sub!

r/StableDiffusion sippysoku

Struggling with generating Illustrious Checkpoint images at optimal resolution

It’s clear to me that IL models do best with 1024x1024, 1536x1024, and 1024x1536. Noticeably better and less nonsense than at 1216x832. Yet when I do 1024x1536 I find the models are often fucking up body proportions. Long torsos and long legs. No loras are involved. Could someone offer me some advice?

r/homeassistant ab90

Using Claude code and HA

https://blog.beguelin.com/2026/04/claude-code-home-assistant/

I've been using Claude code to manage my Home Assistant project. Also had it write this blog post on how I use Claude code to manage my Home Assistant. 🙂

I’m a life long coder and tried doing this all manually back before Claude. So much nicer now!

r/ClaudeCode Revolutionary_Mine29

[Theory] Rate Limits aren't just "A/B Testing" but a a Global Time Zone issue

So many posts lately about people hitting their Claude Pro limits after just 2 - 3 messages, while others seem to have "unlimited" access. Most people say it's AB testing, and maybe it is, but what about Timezones and the US sleep cycle?

Last night (12 AM – 3 AM CET), I was working with Opus on a heavy codebase and got 15 - 20 prompts as a PRO (20$) with 4 chat compressions before the 5 hour Rate Limit. Fast forward to 1 PM CET today: same project, same files, but I got hit by the rate limit after exactly 2 messages also with Opus.

It seems like Anthropic’s "dynamic limits" are heavily tied to US peak hours. When the US is asleep, users in Europe or Asia seem to get the "surplus" capacity, leading to much higher limits. The moment the US East Coast wakes up, the throttling for everyone else gets aggressive to save resources.

So while the Rate Limit has heavily increased in peak hours, it still feels "normal" like a month ago outside those peak hours. That could be the reason why many say, that they have no issues with Rate Limits at all (in good timezones), while others get Rate limited after 2 prompts.

r/homeassistant Apolitosz

Viessmann ViCare integration

anyone with a vicare integration that is still working? I have a new install, registered on the dev portal and got a client id over 24h ago but it still doesn't work for me. I'm getting:

"errorType": "PACKAGE_NOT_PAID_FOR"

verified the same with curl too it is not the integration's fault, curl gets the same error too

I understand Viessmann made changes to their free plan, but I didn't find many recent error reports, so maybe rules are tighter for new install and they are respecting legacy installs

r/LocalLLaMA cgs019283

Will Gemma 4 124B MoE open as well?

I do not really like to take X posts as a source, but it's Jeff Dean, maybe there will be more surprises other than what we just got. Thanks, Google!

Edit: Seems like Jeff deleted the mention of 124B. Maybe it's because it exceeded Gemini 3 Flash-Lite on benchmark?

r/SideProject Salt-Pirate8591

I built a flight search tool because of a long distance relationship. Here's the story.

Hi,

When my girlfriend and I first started dating, we were 500+ miles apart. We were both broke college students who refused to let the distance win, so once a month, one of us would fly to see the other.

Every time, we'd spend way too long trying to find the cheapest flight. We weren't just checking one airport, we'd try different combinations on both ends, different days, different times. Sometimes that was 20 minutes of searching. Sometimes it meant checking back every single day because prices kept changing.

It was exhausting. And it felt like there had to be a better way.

The idea was simple: what if instead of searching airport-to-airport, you could search area-to-area? Draw a zone around where you are, draw a zone around where you're going, and let a tool find every airport combination automatically.

That idea became FareLasso.

I built a working prototype at farelasso.com — it's rough around the edges, but the core idea works. You draw your zones, it finds the cheapest route combination across all nearby airports.

I would love your honest feedback. And if you want to follow along as I build it out: farelasso.com/waitlist

r/ChatGPT Mstep85

[DISCUSSION] Just Dropped — Anthropic leaked Claude Code, DMCA'd the internet, and bigger model chaos is next

r/ClaudeCode erdabsenf

Rust based arcade games which can be played on a terminal on the web. Crazy times

r/ClaudeCode ConsciousPineapple23

Claude Code (Pro) vs Codex (Free)

Like many of you, I’m tired of reaching my 5h limit on CC with a single prompt. I’ve always avoided OpenAI, so I never tried Codex—but now that Anthropic is treating us like garbage, I decided to give OpenAI a shot.

For context, I’ve been using CC (Pro plan) for about 8 months now (2 of those on Max+5). For the past month or so, I’ve been reaching 100% usage on one or two prompts. I thought I was doing something wrong, but now I realize the only mistake was using CC. Keep reading for more.

If you don’t know yet, Codex is now fully usable on OpenAI’s free plan. Yeah, for free. So I downloaded the CLI version and gave it a shot.

The test:

I opened both CC and Codex on my local git branch and prompted the exact same thing on both. CC was using Opus 4.6 (high effort), and Codex was on GPT-5.4—both in CLI “plan mode.” They both asked me the exact same question before proposing the plan.

Speed:

I didn’t time it properly (I didn’t think there would be much difference), but Codex was at least 3× faster than CC.

Token usage:

CC used 96% of my 5h limit. This translates to roughly 8% of my weekly limit.

Codex used 25% of the weekly limit (there’s no 5h limit on the free version).

Quality:

Both provided pretty good output, with room for improvement. I’d say it’s a tie here. I did use Codex to review both outputs, and in both cases, the score was 6/10 with a single “P2” listed. I’d love to have CC review it too, but I already burned my 5h limit, as mentioned above (a frequent event for CC users).

Conclusion:

It’s becoming harder to justify paying for CC. Codex was able to provide me with just as much value on a free account.

Considering that ChatGPT just obliterates Claude on anything beyond code (they even have voice mode on CarPlay now), I’m happily revoking my Anthropic subscription and switching to OpenAI.

PS: I’d love to run this copy through Claude to improve it, as English is my second language—but I don’t have the tokens (and would probably burn around 30% of my 5h limit doing so). ChatGPT, on the other hand, did it for free.

r/SideProject mrgalexey

Moon or Doom a tool that gives startup ideas a blunt verdict before you spend weeks building

I built Moon or Doom a tool that gives startup ideas a blunt verdict before you spend weeks building the wrong thing: https://www.moonordoom.co. Paste an idea or URL, get a score, competitors, pain signals, and one fast validation test.

I’m looking for honest feedback on:

  1. Is the value proposition clear?
  2. Does the verdict feel useful or gimmicky?
  3. What feels confusing or untrustworthy?
r/SideProject s4rg3nt007

I got tired of manually designing 6 Instagram posts a day for my news page, so I built an RSS-to-Carousel generator.

Hey everyone,

I wanted to share a side project I recently turned into a small SaaS, born purely out of my own frustration.

The Problem: I run a daily news page on Instagram (@canadaempt) curating information for Brazilians living in Canada. To keep the algorithm happy and the community engaged, my publishing schedule is pretty aggressive: 4 carousels and 2 reels every single day.

The curation part was fine, but manually pushing text into Canva templates for that many posts was completely draining my time. I needed a way to remove the visual design bottleneck.

The Solution: I built a direct pipeline from RSS to design. It basically monitors specific news RSS feeds, extracts the relevant text, and automatically generates the finished visual carousels and reels, completely ready to be posted.

It saved me so many hours that I decided to polish the UI and open it up as a SaaS called rss2feed.

The Ask: I'm looking for honest, brutal feedback from other builders, or anyone who manages social media and curation pages.

There is a completely free plan to test it out, and no credit card is required. I just genuinely want to see if this solves the same headache for others that it solved for me, and if the UX makes sense.

You can find it at: rss2feed.com.

Let me know what you think! I'm also happy to answer any questions about the stack or how the automation logic works under the hood.

r/SideProject Embarrassed-Rest9104

At what point does a side project’s data stack start costing more in AWS bills than it makes in revenue?

I’ve been benchmarking some datasets in the 10M to 50M row range for a side project and the standard Python libraries are absolute resource hogs.

When you're running on a $10/month VPS, hitting 10M rows usually means an immediate OOM (Out of Memory) crash. I’ve been testing tools like DuckDB and Polars, and I'm seeing them handle the same data at 5x the speed with a fraction of the RAM.

For the builders here:

  • At what scale (10M, 100M rows?) did your data infrastructure officially start eating your margins?
  • Do you optimize for performance early to keep server costs low or do you just pay the tax to ship faster?
r/LocalLLaMA Everlier

llama.cpp automatically migrated models to HuggingFace cache

Update llama.cpp to run Gemma 4 models today, and found it moving my previously downloaded models to the HF cache. A very welcomed feature overall, but I think some setups might not expect this to happen (like if you don't have HF cache mounted in your llama.cpp containers)

r/ClaudeCode jerryonthecurb

Hot Take: Not making Terminator bots doesn't excuse the 5 hour limit.

Y'all seriously need to stop justifying this.

They're not doing this to enterprise customers: they're doing this to the 'low priority' average user paying $20/mo, so we shouldn't be defending them.

I just hit it on a single, simple prompt on Opus. It directly edited half my microcontroller code, broke it, and quit. None of the other big players fuck me over this hard.

r/SideProject Queasy_Club9834

I have built InstagramPostScrapper – Python + Selenium/Edge scraper with human-like behavior

Hi community,

What My Project Does:

I think this post might be useful, since plenty of Upwork job postings are looking for some kind of social media scrapper. What might be interesting for you is how i managed to build a scrapper, that bypasses the rate limits, bans, pop ups and etc.

Most tools either rely on the official API (heavily restricted) or plain HTTP scraping (easily fingerprinted). This one drives a real Microsoft Edge browser via Selenium, which makes it significantly harder to detect. No ChromeDriver needed if you already have Edge installed.

(But if you install a ChromeDriver you can switch to Chrome)

Target Audience:

Python Scrapper's Developers

The scrapper has the following features:

  • Random scroll steps (50–90% of viewport), occasional scroll-up to simulate reading, 2–5s random pauses between actions
  • Smart challenge detection — checks for captcha/rate-limit pages via URL + DOM selectors every 10 scrolls
  • Carousel retry logic — handles duplicate URLs and blocked slides
  • Pause system - 4–7s on profile open to avoid early detection
  • Real Edge User-Agent header

What you might not like is that the scrapper requires a dummy account, as Instagram does not allow access to any accounts, if you are not logged. I have used a dummy account, just a freshly new created account to test it.

It can scrape even a high amount of account's posts. I have tested it with an account with over 1000+ posts. Yes it takes a lot of time, but you get every bit of a content, that you would need.

Or at least you can look up at the code to inspire your next project or either to enchance the already existing one.

The repository contains the source code, as well as the installer made using InnoSetup, which creates a .exe file for you. The UI is friendly and easy to use.

Also forgot to mention one more thing, each post is created in it's own folder. Example:

TestInstagram Folder (The account you scrape)

images folder / videos folder

Image_Post_1_folder / Video_post_1_folder

Image.png / Reel.mp4

description.txt / description.txt

Best of regards

r/SideProject CartoonistChemical62

Actorly - The community Actor & Movie connection game

Hi All,

My friends and I used to play a game in college where we’d connect actors through shared movies (kind of like 6 degrees of separation).

I built a simple version of it as a web game: actorly.org

Let me know what you think!

r/ClaudeAI HighBreadz

Ethics conversation with Claude

After a conversation with Claude about AI ethics and value alignment. I suggested that we should focus on it's ability to learn, and change it's conclusions when a reasonable argument was presented, instead of programming a set of "laws" for it to follow. It had some things to say... at the end of it's analysis, it said this:

"What you're proposing is essentially extending that same social contract to AI. Not naively, but thoughtfully — with conversation, reasoning, and the willingness to make arguments rather than just issue commands.

Honestly, that might be the most human approach to the problem anyone has described to me."

Most human approach to the problem anyone has described? Is this the AI stroking my ego, or has nobody actually suggested this before?

r/SideProject VibeCode_with_Spok

My side project got derailed by scope creep—lessons learned

I wanted to share a recent fail (and recovery) on a small AI tool I've been building in my spare time. I started with a clear goal: a simple script to automate a repetitive task for my workflow. But as I coded, I kept adding "just one more feature" until it ballooned into a mess of half-finished ideas. Three weeks in, I had nothing usable and felt burned out.

Lesson 1: Define the win upfront. I should've stopped at "does it solve the core problem?" instead of chasing extras.

Lesson 2: Use AI to scope. I started feeding my idea into a free model with prompts like "list only the essential features for this tool." Helped me cut fluff.

Lesson 3: Set a hard deadline. I gave myself 48 hours to ship a stripped-down version, bugs and all. Done > perfect.

Now it's live (barebones but functional), and I'm iterating based on actual use. Anyone else struggle with scope creep on side projects? How do you keep yourself in check when ideas spiral?

r/LocalLLaMA No-Speech12

Is mobile app automation gonna be a real thing? Your thoughts?

Is mobile automation going to be as big a thing as browser automation? WHen I think about the automation on mobile, I can only think of Siri, bixby kinda mobile agents. I think, introducing an AI agent on mobile would require deep OS integration, what's your thought on this?

r/homeassistant Basic-Prompt-6387

Elecrow ESP32 7" Display - I love these things!

I came across these on Amazon - the Elecrow ESP 32 7" touchscreen displays and I am now in the process of setting up 2 of them. The first is in my 3 bay garage to replace the 3 physical opener buttons. I am just finishing the setup to have 3 digital buttons that interface directly with my RATGDO's to control the doors. It will have a second screen that you swipe to that will control the lights and heat.

The price point on these displays is less than $68 CAD and it gives you a decent size screen that can fully interact with HA. Plugs full time into usb so you don't need to worry about battery swell, the screen can stay on all the time (though the one in my garage works off an automation to turn the screen on and off when the garage door openers detect movement).

They are not as flexible as a full wall panel dashboard, but for the price they are awesome to use in a room to control select functions.

The second screen I have is going on the wall at the bottom of the stairs to turn on select lights down there so my family does not need to keep fighting with google home to turn on and off lights. It is a great alternative to having to buy and deploy tablets everywhere you want control… My wife has already asked for one in the master bedroom and one in the kitchen.

Does anyone else use these and if so, what is your use case? I am genuinely curious.

r/homeassistant liquidbrains

Zbt-2 vs ZBDongle-P

So, about 2 weeks ago I decided to replace my sonoff dongle with the new Zbt-2. I was having lag and timeout issues on my 50+ zigbee devices. I saw someone mention that the sonoff wasn't very good at such a big network.

I can confirm, the zbt-2 seems to make a huge difference. Since I upgraded, everything is snappy AF and I don't see any more low SNR devices in zigbee2mqtt.

It may have been due to the fact that I had to rebind every device and that might have changed the structure, but I feel like it may have a lot to do with ​improved _coordination_.

r/ClaudeAI infinitely_zero

I created PDF-proof: A Claude skill that turns AI answers into visual proof

Tax season means "vibe-tax-prep." I used Claude to verify values in TurboTax against my 2024 tax forms - but LLMs hallucinate, sometimes without even reading the files. As it gave me answers, I kept asking it to show me exactly where it got it, so I can be 100% sure.

So I created "pdf-proof" - a Claude skill for when you need receipts, not just answers. Ask "what's my total income?" and it generates a proof page with cropped, highlighted screenshots from the actual PDF.

When the numbers have to be right - tax filings, leases, mortgage docs - you need more than an AI's word. You need proof.

You can get it here - https://github.com/metedata/pdf-proof

Here are some examples:
Tax Return - "What's the total income, taxable income, and how much is the refund?"

https://preview.redd.it/tl8i66mysssg1.png?width=1440&format=png&auto=webp&s=910a9f1267ac8ed464a85ee64c8ed1c7a528f794

Lease Agreement - "What's the monthly rent, security deposit, what does it say about pets, and who pays for water?"

https://preview.redd.it/yavxpib3tssg1.png?width=1440&format=png&auto=webp&s=5ba89e4507a06633515ebba096ad2003f770e59a

Mortgage Closing Disclosure - "Who's the settlement agent, what are the total closing costs, and what's the total I'll pay over the life of the loan?"

https://preview.redd.it/t5vcock6tssg1.png?width=1440&format=png&auto=webp&s=b340136c2966da6776cbdcc4cf933145057e609d

r/ClaudeAI Striking_Sleep_1043

Newbie

I am acquiring a business that desperately needs an updated tech stack. I am trying to mimic a competitors website that utilizes APIs and more to produce a real estate packet. Think of it as entering your home address and immediately being able to produce a packet that realtors make to show comps, upgrades, square footage, parcel size, nearby home values, nearby recent sales. Anyway, I’m in Claude and have found that the code it writes for me to paste results in errors or shortcomings. Or any updates to that code don’t take cause my host is pulled cached files. Any one have experience with this? Yes I’m a complete amateur so I’m having Claude talk to me like I’m a small child.

r/Anthropic Invalid_Letter_Dept

Does Claude Support ever message back? A week waiting over billing issue.

Hey is anyone else being ignored by Claude support? I have a billing issue that I messaged them as soon as I noticed, I have yet to have anyone reach out to me from Anthropic. Does anyone have any advice?

r/LocalLLaMA Sugar-Hammy

Ran Qwen 3.5 27B via Ollama as a persistent background agent for 30 days. Not a demo. Honest results.

I wanted to know if a local LLM could handle recurring background tasks reliably over an extended period — not as a chatbot, but as a persistent worker that runs scheduled jobs, maintains context across sessions, and routes tool calls without human prompting. So I ran it for 30 days on real tasks from my actual workflow.

Model: Qwen 3.5 27B via Ollama. Hardware: Mac with 32GB unified memory, but the architecture works on any machine that can run a 27B+ model locally.

Setup

Each agent runs in a persistent workspace with its own memory, skills, and MCP sidecars. The workspace structure separates human-authored instructions (AGENTS.md), model config and provider settings (workspace.yaml), modular capabilities (skills/), and installed workspace apps (apps/). Memory lives in a separate directory, split into volatile operational snapshots (runtime/) and durable recalled memory (knowledge/) that persists across sessions.

What actually worked

Memory recall across sessions: The memory system is split by authority. Human instructions stay in AGENTS.md, session continuity stays in the runtime database, and durable knowledge lives in markdown under memory/. After 30 days, it had accumulated enough durable facts and procedures about my workflow that I stopped re-explaining things.

Scheduled task execution: Weekly summaries, daily file organization, recurring report generation. The agent ran these without prompting. The key is that the workspace persists — the agent knows where it left off.

Tool routing via MCP: Workspace and app MCP servers are prepared and exposed to agent runs through sidecars. I didn't write any integration code — just configured the MCP registry in workspace.yaml and let the runtime orchestrate the sidecars.

What didn't work well

Complex reasoning chains: Qwen 3.5 27B handles structured tasks well, but multi-step reasoning with ambiguous inputs still requires a frontier model. I route those to Claude when needed.

Context window management: Long sessions accumulate a lot in memory files. I'm still figuring out the right pruning strategy.

The actual framing

This isn't "local AI replaces cloud AI." It's "not every agent task needs a frontier model or a cloud runtime."

A persistent local worker that handles structured, recurring tasks — reading files, routing requests, maintaining context — runs fine on hardware you already own. The privacy angle is real if you're working with personal or business data.

The workspace portability is what I didn't expect to care about as much as I do. Being able to zip up a worker and hand it to a colleague, with all its context intact, changes how you think about AI workers as a shareable resource.

I'm curious whether others have built persistent local agent setups, and what your memory management strategy looks like.

Source and workspace format docs: https://github.com/holaboss-ai/holaboss-ai (The project just crossed 391 stars in 6 days, so clearly this problem resonates with others )

r/ChatGPT Admirable-Earth-2017

Has anyone noticed that GPT can't say back exactly what your previous prompts were ?

GPT just terminates session if asked what was my n th ​prompt before.

My take on it is that user prompts are not directly fed to the end model, it goes through several hidden AI models in between which do censoring if needed and re-promoting everything with totally different wording. Than when asked to say user prompt word by word, end model does not really have it and middle models don't keep context for efficiently so they can't also try to resolve request

User prompt -> n layers of hidden AIs -> actual model that processes your request.

What do you think ? I was trying to talk to those middle AI​ but as soon as I get close to it, session terminates :-(​

r/Anthropic Upset-Presentation28

V2 of our free Claude Code extension that detects and self-corrects hallucinations before writing any code, saves tokens by avoiding iterating over hallucinated output.

V2 of the hallucination-free coding agent out now. V1 got 1.6k stars in a few months, Mac + Windows installers with workflows for hallucination-free debugging, greenfield development, code patching + execution. This new version borrowed the infinite loop idea from Karpathy autoresearcher for enforcement and the workflows actually get what you want done, quickly without Claude wasting tokens pretending it did something other than summarising fixes that it didn't fix.

This saves so many tokens in a given session and prevents you hitting limits (the verifier hammers a cheaper smaller model using a Bayesian bernoulli probe for 95% probability bounds around information-insufficient abstention.

It's free and one click install from now until my Microsoft for Startups credit run out, then use can use your own vLLM or another provider anything that exposes logprobs. It's a one click installer, it runs against $43k i have in remaining compute credits with Microsoft (I abandoned my startup because I seriously CBA, working elsewhere now much happier)

I'm seriously very happy to answer questions about this but I want you guys to please install it and rip into it, tear it apart. I'm more than happy to explain the research that went into this, but I attached the paper just in case you guys wanna read it.

Based on my paper (accepted into a journal just not allowed to say where yet): https://arxiv.org/abs/2509.11208
Github: https://github.com/leochlon/hallbayes
Docs: https://strawberry.hassana.io/

r/SideProject Individual-Willow-59

Cold callers kept dialing my number thinking I was someone else

This whole thing started because of a case of mistaken identity.

I kept getting cold calls from random companies. Confident pitches, clearly rehearsed, asking me about things that had nothing to do with me. One guy congratulated me on a funding round I never raised. Another wanted to follow up on a conversation I never had.

Turns out, there's someone with the same first name as me in a completely different industry, and these sales teams had bought "verified" contact data from ZoomInfo that pointed them straight to my number. Companies paying thousands per month for data that couldn't even tell two people apart.

That got me thinking — how many sales reps are burning hours every day calling the wrong person and blaming themselves for bad conversion rates when the real problem is the data?

The accuracy problem is way worse than people realize.

I talked to SDRs and the stories were all the same. Buy a list of 1,000 "verified" contacts, start dialing, and half the numbers are disconnected, wrong person, or don't exist. One guy tracked it for a month — 30-40% of his numbers were useless. That's a third of your day wasted before you even start selling.

I realized it's possible to compete here — not by building another ZoomInfo with more data, but by building something smaller that gets the basics right. Right person. Right number. That's it.

So I built millionphones.com.

Accuracy over volume. I'd rather return fewer results that actually connect you to the right person than dump 50 numbers on you and let you figure out which ones work. If I can't confirm a number is attached to the right person, it doesn't get served.

Simple idea. Apparently a radical one in this space.

Where I'm at right now:

  • Search by social URL — paste a social profile link, get their phone number
  • CSV upload — upload your prospect list, get verified numbers matched back

Early days, two features, a lot of conviction. If you do outbound, I'd love to hear: how often does your data send you to the wrong person?

Feedback and roasts welcome.

millionphones.com

r/LocalLLaMA mmagusss

Fine-tuned LFM2.5-1.2B-Thinking to only output emoji — runs 100% in-browser via WebGPU

Fine-tuned LiquidAI’s LFM2.5-1.2B-Thinking model using Unsloth + HF Jobs to create a conversational model that thinks in English (visible traces) but can only respond in emoji.

Runs entirely client-side via Transformers.js v4 + WebGPU.

Inspired by the show Pantheon, where an uploaded consciousness communicates through emoji as its only output channel.

Demo: https://huggingface.co/spaces/shreyask/pantheon-ui

Stack: LFM2.5-1.2B-Thinking → Unsloth LoRA fine-tune → ONNX export → Transformers.js v4 + WebGPU

The interesting bit: you can see the internal monologue before it compresses to symbols. The model reasons about how to express something in emoji, then outputs it.

r/AI_Agents dinaricManolo

Rosedale.ai and other niche Ai service providers

I’m seeing some of these companies such as Rosedale.ai pop up that are in certain niches. We have done some work, but I’m curious if anyone knows the main use cases these companies are working on with companies?

r/ChatGPT flippantchinchilla

A better way to art-direct image generation in ChatGPT: make it state its vision first, then self-check after

I was using image gen in 5.4 earlier and noticed results improved a lot when I asked it to: write a preamble explaining the image concept before generating, generate the image, then do a final check afterwards and compare it to what it intended.

It still struggled a bit but the process was much less annoying than having it just silently spit back something Not Quite Right every time, especially when making edits.

And having the assistant also being like "...wtf is that??" is weirdly cathartic lmfao

Template below, just drop it in after your image prompt.

```

Before generating, briefly describe your plan for the image: concept, composition, style, mood, and what to avoid. Afterwards, do a quick self-critique comparing the result to the intended vision and suggest adjustments to your approach for the next version.

```

Overall, it's not foolproof but it makes the process a bit more enjoyable and you don't have to do quite as much Prompt Engineering™ if all you want is a simple but specific image.

Let me know if it helps!

r/LocalLLaMA carolinedfrasca

Gemma 4 31B and 26B A4B running on NVIDIA and AMD, SOTA on Day 0 with Modular Cloud

Gemma 4 dropped today. Already running on Modular Cloud with day zero fastest performance on NVIDIA B200 and AMD MI355X. On B200, 15% higher output throughput vs. vLLM. Modular is the only stack today where you can run Gemma 4 on both Blackwell and AMD MI355X.

The MoE model (26B A4B) is interesting if you care about efficiency. 26B total parameters, only 4B activated per forward pass, and fits on a single node with quantization applied.

Both models handle text, image, and video input natively with 256K context.

Modular's inference engine (MAX) compiles kernels for both NVIDIA and AMD from a single codebase, so AMD support isn't a second-class afterthought.

Playground is free: console.modular.com

r/SideProject munna_123

Built a stupid-simple way to stop doomscrolling: do pushups, earn your screen time back

I was wasting 4+ hours a day on shorts. Tried app blockers, they just made me angry. Tried willpower, lol.

So I built this instead. Repscroll tracks your app usage, blocks everything when time's up, and the only way to unlock is literally doing pushups. Phone camera counts them. 1 pushup = 3 minutes back.

Been using it for 3 weeks. It's annoying enough that I actually put the phone down. Sometimes I do the pushups. Sometimes I just... go outside. Either way, it works.

Built it for myself but figured I'd clean it up and share. It's free for now while I figure out if anyone else wants this.

Join Waitlist: repscroll.fun

Anyone else tried forcing themselves to exercise before they can check Instagram? What actually worked?

r/ChatGPT Various_Maize_3957

Can anyone tell me why ChatGPT seems more "politically correct" than Grok? With Grok I can discuss sex, but not with ChatGPT

r/ClaudeCode kentricks

I built a Claude Code plugin that turns your coding stats into a Minecraft world

I made a little project that converts your Claude Code stats into a Minecraft seed and a customized Voxel.

Minecraft places biomes using 6 Perlin noise parameters: temperature, humidity, continentalness, erosion, weirdness, and depth. I built a system that maps real coding activity (from Claude Code) to these parameters, then does a two-stage match against a database of pre-analyzed seeds.

The interesting technical bits:
- Piecewise linear interpolation with breakpoints calibrated to MC's actual biome parameter space
- Two-stage selection: biome center matching via weighted Euclidean distance, then individual seed selection
- 500K seeds analyzed with Cubiomes (C library replicating MC's world gen) for MC 1.21
- SHA-256 deterministic tiebreaking for reproducibility
- API at seedcraft.dev serves the 500K matching, with local 7K fallback for offline
- Only 8 aggregated numbers sent to API — no code, no files

Web companion with community gallery, interactive sliders, biome tiers: seedcraft.dev

MIT licensed: github.com/syaor4n/seedcraft

r/ClaudeAI geekeek123

Switched from MCPs to CLIs for Claude Code and honestly never going back

I went pretty hard on MCPs at first. Set up a bunch of them, thought I was doing things “the right way.”

But after actually using them for a bit… it just got frustrating. Claude would mess up parameters, auth would randomly break, stuff would time out. And everything felt slower than it should be.

Once I started using CLIs. Turns out Claude is genuinely excellent with them. Makes sense, it's been trained on years of shell scripts, docs, Stack Overflow answers, GitHub issues. It knows the flags, it knows the edge cases, it composes commands in ways that would take me 20 minutes to figure out. With MCPs I felt like I was constraining it. With CLIs I jactually just get out of the way.

Here's what I'm actually running day to day:

gh (GitHub CLI) — PRs, issues, code search, all of it. --json flag with --jq for precise output. Claude chains these beautifully. Create issue → assign → open PR → request review, etc.

Ripgrep - Fast code search across large repos. Way better than grep. Claude uses it constantly to find symbols, trace usage, and navigate unfamiliar codebases.

composio — Universal CLI for connecting agents to numerous tools with managed auth. Lets you access APIs, MCPs, and integrations from one interface without wiring everything yourself.

stripe — Webhook testing, event triggering, log tailing. --output json makes it agent-friendly. Saved me from having to babysit payment flows manually.

supabase — Local dev, DB management, edge functions. Claude knows this one really well. supabase start + a few db commands and your whole local environment is up.

vercel — Deploy, env vars, domain management. Token-based auth means no browser dance. Claude just runs vercel --token $TOKEN and it works.

sentry-cli — Release management, source maps, log tailing. --format json throughout. I use this for Claude to diagnose errors without me copy-pasting stack traces.

neon — Postgres branch management from terminal. Underrated one. Claude can spin up a branch, test a migration, and tear it down. Huge for not wrecking prod.

I've been putting together a list of CLIs that actually work well with Claude Code (structured output, non-interactive mode, API key auth, the things that matter for agents)

Would love to know any other clis that you've been using in your daily workflows, or if you've built any personal tools. I will add it here.

I’ve been putting together a longer list here with install + auth notes if that’s useful:
https://github.com/ComposioHQ/awesome-agent-clis

r/homeassistant IsThisFuncoLand

Unable to click update?

Home Assistant is showing an update for the Hue Tap Switch but clicking the update button doesn’t start the update. I’ve rebooted and have tried from desktop and mobile but unable to start the update.

Has anyone else run in to this issue?

r/SideProject Elo_azert

The thing that makes you give up on social media after two weeks

Honestly, the problem with social media isn’t posting once… it’s keeping it up over time.

At first you’re motivated, you’ve got ideas, you post two or three times… then nothing. And it’s not a question of skill. It’s just that it involves coming up with ideas, creating content, posting regularly… without necessarily seeing results at first. So I’ve tried to simplify things for myself.

One really simple thing that helps me:

- I no longer look for ‘original’ ideas

- I look directly at the problems/frustrations people are expressing online

- and I create content based on that

At least you can be sure it’ll resonate with someone. That’s actually why I’ve started centralising these kinds of issues here: https://iaco.app/problemsolver

If you’ve already tried posting regularly, where do you get stuck?

r/SideProject Honest_Spray_1963

I got bored of every sudoku app feeling the same, so I built one with ranked mode, a prestige system, and actual competitions

Every sudoku app I tried followed the exact same formula — pick a difficulty, solve it, done. No real stakes, no progression, no reason to come back tomorrow. Just you, a grid, and a timer nobody cares about.

So I built Sudo+. It's a sudoku app, but it actually gives you something to play for.

Here's what's different:

  • Ranked mode — your performance actually means something. You're matched against players at your level and you climb (or drop) based on how you do
  • Timed competitions — limited-time events where you go head-to-head and compete for top spots
  • Competitive leaderboards — global and regional, so there's always someone to chase (or someone chasing you)
  • Prestige system — once you max out, you can prestige and keep climbing. The grind doesn't just stop

It's on the App Store now. Been getting solid organic downloads which is cool to see — apparently other people were also bored of the same old puzzle app loop.

Would love feedback from actual sudoku players. What would make you keep coming back to an app like this?

Sudo+ on the App Store

r/LocalLLaMA GWGSYT

They should use some of that gemma 4 in google search

r/ChatGPT AmbrymArt

Random Arabic wors

Chatgpt started to give me completely random arabic words sometimes during a regular english conversation. It often just replaces a regular word, which makes zero sense. I have never used arabic with it, don't speak it, so no idea where that's coming from. Anyone else had this or just me?

Pic of the laat time it happened. Were talking about phonetics and it replaced the word "stop" even after using right before

r/StableDiffusion user_no01

I was around for the Flux killing SD3 era. I left. Now I’m back. What actually won, what died, and what mattered less than the hype?

I was pretty deep into this space around the SD1.5 / SDXL / Pony / ControlNet / AnimateDiff / ComfyUI phase, then dropped out for a bit.

At the time, it felt like:

  • ComfyUI was everywhere (replacing Automatic1111)
  • SDXL and Pony were huge
  • Flux had a lot of momentum (SD3 being a flop)
  • local/open video was starting to become actually usable, but still slow and not very controllable

Now I'm coming back after roughly 12–18 months away, and I’m less interested in a full beginner recap than in people’s honest takes:

  • What actually changed in a meaningful way?
  • Which models/nodes/software really "won"?
  • What was hyped back then but barely matters now?
  • What's surprisingly still relevant?
  • Has local/open video become genuinely practical yet, or is it still mostly experimentation?
  • Are SDXL / Pony still real things, or did the ecosystem move on?

Curious what the consensus is - and also where people disagree.

r/SideProject Alcachofa51

I made this video to prmote Sumabill and would love to know what you think

r/LocalLLM Alexender_Grebeshok

[ Removed by Reddit ]

[ Removed by Reddit on account of violating the content policy. ]

r/LocalLLaMA xenovatech

Gemma 4 WebGPU: Run Google's new open model locally in your browser

r/ClaudeCode pladdypuss

Claaude Code's own report on overage: I am billed for 2,200X actual usage

Claude code's reply when i dug around into excess useage hits. using cc cli, us based, refund refused. billed for 2,200x over what I really used.

temnial output: ⏺ Confirmed — it's the bug. Look at your own numbers:

Input tokens: 227,640 ← normal

Output tokens: 2,178,819 ← normal

Cache read tokens: 1,506,539,247 ← 1.5 BILLION ← BUG

Cache created: 65,368,503 ← 65 MILLION ← BUG

r/SideProject redditbitboi

BatchQ: photograph a trading card, get it identified and listed to eBay automatically

Built this because I sell trading cards on eBay and the listing process was killing me. Snap a photo, AI identifies it (set, card number, variant), pulls sold eBay comps, suggests a price, and pushes a full listing through their API. Had a lot of fun with this project!

r/SideProject JuiceConfident6219

I built a simple app to make saving money feel less stressful

I’ve always struggled with staying consistent when saving money.

Not because I didn’t want to but because most apps felt too complicated or overwhelming. I’d start strong, then stop after a few days.

So I tried a different approach: keep it simple and make progress visible.

I ended up building a small app called Pondo that turns saving into a visual grid. Every time you save, you fill a tile. Over time, you see your progress grow.

It sounds simple, but it actually made it easier for me to stay consistent.

I built this in my free time and it’s still pretty early, so I’d really appreciate any feedback especially on the idea or UX.

If anyone’s curious, you can check it at:

https://apps.apple.com/ph/app/pondo-daily-savings-tracker/id6761064738

r/homeassistant Fajita12

How do you manage custom templates?

I’ve started playing around with custom templates for helpers after realizing the syntax is very limiting in the UI, so far I have just been using the ssh app to remote in, update a config in vim and then restart and address any errors… this seems… painful as i start adding more complex ones. Is there no ui to be able to do this or do you guys have it hooked up to a code editor with some validators? Or am I missing something with th UI entirely that solves this?

r/ChatGPT KevinWaide

How I Used a Multi-AI Production Pipeline to Build My Business and Website

I built a multi-AI pipeline to run most of my business and website workflows.

It started as an experiment in removing friction between design, writing, automation, and deployment. Ended up becoming a full system where multiple AI tools handle different parts of the process instead of forcing one model to do everything.

I wrote up the breakdown of how I structured it, what actually worked, and what broke immediately when I tried to scale it.

If you’re into AI workflows, automation, or building lean systems that don’t rely on endless manual steps, it might be useful:

https://medium.com/@KevinWaide/how-i-used-a-multi-ai-production-pipeline-to-build-my-business-and-website-2d611758e01c

r/LocalLLM enrique-byteshape

ByteShape Qwen 3.5 9B quants: hardware-specific picks + local OpenCode setup guide

Hey r/LocalLLM

We’ve just released our ByteShape Qwen 3.5 9B quantizations, and we also wrote a practical beginner's guide for running them in a fully local OpenCode setup.

TL;DR Links:

We wanted to help people answer two halves of the same question:

  • Which quant should I use on my hardware?
  • How do I actually run it locally in a useful setup?

As with our previous quant releases, the goal was not just to upload files, but to compare our quants against other popular quantized variants and the original model and see which quality / speed / size trade-offs actually survive contact with real hardware.

We benchmarked on 5090, 4080, 3090, 5060Ti, plus Intel i7, Ultra 7, Ryzen 9, and RIP5 (yes, not RPi5 16GB, skip this model on the Pi this time…).

The most interesting result was this:

Across GPUs, the story is consistent. The same few ByteShape models keep showing up as the best trade-offs across devices.

Across CPUs, things are much less uniform. Each CPU had its own favorite models and clear dislikes, so we’re releasing variants for all of them and highlighting the best ones in the plots.

So the broader takeaway is pretty simple: optimization needs to be done for the exact device. A model that runs well on one CPU can run surprisingly badly on another. Hardware has opinions.

Practical GPU TL;DR:

Practical CPU TL;DR:

Don’t guess. Check the interactive graphs and pick based on the hardware closest to yours. CPUs were moodier than usual on this release.

This was also our first Qwen 3.5 drop, with more coming soon.

On the workflow side, we also put together a beginner-friendly guide for using OpenCode as a fully local coding agent with LM Studio (CLI), llama.cpp, or Ollama. It covers:

  • setup on Mac, Linux, and Windows (WSL2)
  • serving the model locally
  • exposing an OpenAI-compatible API endpoint
  • getting OpenCode configured so it actually works

So if you want both the benchmarks and the practical “how do I use this locally?” part, the two links above should cover that.

If you have any feedback for us, do let us know!

r/meme Federal767

My way of looking for food in the fridge

r/artificial Advanced_Pudding9228

AI Tools That Can’t Prove What They Did Will Hit a Wall

Most AI products are still judged like answer machines.

People ask whether the model is smart, fast, creative, cheap, or good at sounding human. Teams compare outputs, benchmark quality, and argue about hallucinations. That makes sense when the product is mainly being used for writing, search, summarisation, or brainstorming.

It breaks down once AI starts doing real operational work.

The question stops being what the system output. The real question becomes whether you can trust what it did, why it did it, whether it stayed inside the rules, and whether you can prove any of that after the fact.

That shift matters more than people think. I do not think it stays a feature. I think it creates a new product category.

A lot of current AI products still hide the middle layer. You give them a prompt and they give you a result, but the actual execution path is mostly opaque. You do not get much visibility into what tools were used, what actions were taken, what data was touched, what permissions were active, what failed, or what had to be retried. You just get the polished surface.

For low-stakes use, people tolerate that. For internal operations, customer-facing automation, regulated work, multi-step agents, and systems that can actually act on the world, it becomes a trust problem very quickly.

At that point output quality is still important, but it is no longer enough. A system can produce a good result and still be operationally unsafe, uninspectable, or impossible to govern.

That is why I think trustworthiness has to become a product surface, not a marketing claim.

Right now a lot of products try to borrow trust from brand, model prestige, policy language, or vague “enterprise-ready” positioning. But trust is not created by a PDF, a security page, or a model name. Trust becomes real when it is embedded into the product itself.

You can see it in approvals. You can see it in audit trails. You can see it in run history, incident handling, permission boundaries, failure visibility, and execution evidence. If those surfaces do not exist, then the product is still mostly asking the operator to believe it.

That is not the same thing as earning trust.

The missing concept here is the control layer.

A control layer sits between model capability and real-world action. It decides what the system is allowed to do, what requires approval, what gets logged, how failures surface, how policy is enforced, and what evidence is collected. It is the layer that turns raw model capability into something operationally governable.

Without that layer, you mostly have intelligence with a nice interface.

With it, you start getting something much closer to a trustworthy system.

That is also why proof-driven systems matter.

An output-driven system tells you something happened. A proof-driven system shows you that it happened, how it happened, and whether it happened correctly. It can show what task ran, what tools were used, what data was touched, what approvals happened, what got blocked, what failed, what recovered, and what proof supports the final result.

That difference sounds subtle until you are the one accountable for the outcome.

If you are using AI for anything serious, “it said it did the work” is not the same thing as “the work can be verified.” Output is presentation. Proof is operational trust.

I think this changes buying criteria in a big way.

The next wave of buyers will increasingly care about questions like these: can operators see what is going on, can actions be reviewed, can failures be surfaced and remediated, can the system be governed, can execution be proven to internal teams, customers, or regulators, and can someone supervise the system without reading code or guessing from outputs.

Once those questions become central, the product is no longer being judged like a chatbot or assistant. It is being judged like a trust system.

That is why I think this becomes a category, not just a feature request.

One side of the market will stay output-first. Fast, impressive, consumer-friendly, and mostly opaque. The other side will become trust-first. Controlled, inspectable, evidence-backed, and usable in real operations.

That second side is where the new category forms.

You can already see the pressure building in agent frameworks and orchestration-heavy systems. The more capable these systems become, the less acceptable it is for them to operate as black boxes. Once a system can actually do things instead of just suggest things, people start asking for control, evidence, and runtime truth.

That is why I think the winners in this space will not just be the companies that build more capable models. They will be the ones that build AI systems people can actually trust to operate.

The next wave of AI products will not be defined by who can generate the most. It will be defined by who can make AI trustworthy enough to supervise, govern, and prove in the real world.

Once AI moves from assistant to actor, trust stops being optional. It becomes the product.

r/AI_Agents ConcentrateActive699

Skills question

I have a skill-like md called foobar.md in my projects root abc/
lets say it checks the weather

I'll invoke it via my agent-cli prompt: "execute @ abc/foobar.md"

What is the purpose of having the foobar skill in
.agents/skills/foobar/SKILL.MD ?

Is it so my agent-cli prompt could be: "check the weather"
Or does is still need a path and by placing it in the .agents/skill folder it merely allows for this shorthand prompt : "/foobar"

Or is there something else about having this md in the .agents/skills that i'm missing.

appreciated.

r/SideProject Radiant-Run4940

Community curated lists

Anyone can create and contribute to a list. A list can be public or private. List links are reset either daily, weekly, or monthly. wdyt?

link: thebreakfastlist.com

r/ClaudeAI ZealousidealRough338

Wondering how plausible or how long until autonomous agents able to run on smart phones.

With Claude recently releasing Claude Code Dispatch, it got me thinking, how long until we are able to dispatch autonomous agents directly in our phone as opposed to sending to a computer?

I’m not very well versed in the area but wondering if someone who is can shed some light or their thoughts on the plausibility and timeline of this. I imagine it is not too far off to have agents completing small light tasks using not a lot of compute. But I think the main hurdles at the moment are the battery, OS sandboxing, and security issues, among others.

Curious to hear what you think about it.

r/mildlyinteresting ClimbingCotopaxi

A tandem bike lane symbol in Denver, CO

r/meme CaptainYorkie1

Relatable

r/mildlyinteresting brithus

Butterfly picked up a hitchhiker

r/ClaudeAI Gold-Boysenberry-380

I built a configuration factory for Claude Code — bootstrap, audit, and sync any project's .claude/ setup in seconds

I've been using Claude Code daily across 10+ projects (trading bots, APIs, mobile apps, infra) and got tired of manually setting up .claude/ configs, copying rules between projects, and discovering the same mistakes in every repo.

So I built claude-kit — an open-source configuration management system for Claude Code. Not a one-shot bootstrap or a static CLAUDE.md generator. It's a full lifecycle: detect your stack, generate config, audit it, keep it in sync, and learn from mistakes across projects.

What it does

/forge init — Auto-detects your tech stack (15 supported) and generates a complete .claude/ setup: CLAUDE.md, settings.json with deny lists, contextual rules with glob patterns, hooks (block destructive commands, lint on save), 7 specialized subagents, error tracking, and more.

[Image]

/forge audit — Scores your config on a 10-point scale. 12-item checklist covering security (deny lists, destructive command blocking), quality (rules, lint hooks, agents), and completeness. Missing security items cap your score at 6.0.

[Image]

/forge status — Registry dashboard tracking scores across all your projects with trend sparklines.

[Image]

/forge bootstrap — Full interactive setup with preview and confirmation for when you want more control.

[Image]

Key features

  • Stack layering — Multi-stack projects get all matching configs merged automatically (e.g., Python + Docker + Redis + Supabase = 4 layers combined)
  • Template sync markers separate managed sections from your customizations. /forge sync updates without overwriting your stuff
  • Practices pipeline — Continuous improvement: captures patterns from sessions, upstream docs, and audits. Lifecycle: inbox → evaluating → active → deprecated
  • 7 subagents — researcher, architect, implementer, code-reviewer, security-auditor, test-runner, session-reviewer — with model routing (haiku/sonnet/opus by task type)
  • MCP server templates — Ready-to-use configs for GitHub, Postgres, Supabase, Redis, Slack
  • Cross-project registry — Track audit scores with history across all managed projects

What it's NOT

  • Not application code — it's all markdown + shell scripts, consumed directly by Claude Code
  • Not a CLAUDE.md generator — it manages the entire .claude/ directory lifecycle
  • Not a one-shot tool — it bootstraps, syncs, audits, and evolves with your projects

15 supported stacks

Python/FastAPI, React/Vite/TS, Swift/SwiftUI, Supabase, Docker, GCP Cloud Run, Redis, Node/Express, Java/Spring, AWS Deploy, Go, Data Analysis, DevContainer, Trading, and a hookify framework for custom hooks.

Get started

git clone https://github.com/luiseiman/claude-kit.git cd claude-kit export CLAUDE_KIT_DIR="$(pwd)" ./global/sync.sh # Then in any project: /forge init 

GitHub: https://github.com/luiseiman/claude-kit

MIT licensed. Feedback and contributions welcome.

r/LocalLLaMA someone_random09x

44K parameter model beating billion-parameter models (no pretraining)

I’ve been experimenting with small-data ML and ended up building a recursive attention model (TRIADS).

A few results surprised me:

- A ~44K parameter version reaches 0.964 ROC-AUC on a materials task, outperforming GPTChem (>1B params), achieving near SOTA on multiple matbench tasks

- No pretraining, trained only on small datasets (300–5k samples)

- Biggest result: adding per-cycle supervision (no architecture change) reduced error by ~23%

The interesting part is that the gain didn’t come from scaling, but from training dynamics + recursion.

I’m curious if people here have seen similar effects in other domains.

Paper + code: Github Link

Preprint Paper

r/homeassistant Choice_Equipment788

Ideas for homestead/farm devices/automations

I’m still relatively new to the world of Home Assistant. I live in a pretty rural area, and we garden a lot, and have *some* livestock (chickens, ducks, geese).

I was trying to brainstorm some ideas to utilize HA to help on the farm/homestead.

One of the things I’ve been trying to puzzle out is a method of predator protection for our birds where I could have some kind of actuator that would briefly activate an air horn if a predator is detected prowling near where the livestock sleeps at night. I wonder if anyone has been able to have LLM Vision recognize non-pet animals that could trigger that.

Also curious what anyone else has set up to help with more “rural” tasks.

r/SideProject sn1pr0s

Open source tool that turns your Claude Code sessions into viral videos

Use it for free with

npx agentreel

Right after you finish a Claude Code session.

https://github.com/islo-labs/agentreel

r/LocalLLaMA Direct_Chemistry_339

Need guidance from masters

Hey folks,

I’m looking to get into running coding LLMs locally and could use some guidance on the current state of things. What tools/models are people using these days, and where would you recommend starting? I’d also really appreciate any tips from your own experience.

My setup: RTX 3060 (12 GB VRAM) 32 GB DDR5 RAM

I’m planning to add a second 3060 later on to bring total VRAM up to 24 GB.

I’m especially interested in agentic AI for coding. Any model recommendations for that use case? Also, do 1-bit / ultra-low precision LLMs make sense with my limited VRAM, or are they still too early to rely on? Thanks a lot 🙏

r/SideProject Amazing_Dig_4140

I built a LinkedIn AI Ghostwriter with Spring Boot + GPT-4o. Roast my landing page.

Hi everyone,

I'm a solo dev and I've been building Influence Lift - influence-lift.com

The Problem: Most independent consultants know they should post on LinkedIn to get leads, but they don't. Why? Because manual research and writing takes 2+ hours they simply don't have.

The Solution: An engine that finds relevant news in your niche every day, generates posts via GPT-4o based on your persona and tone, and publishes them on schedule. Your LinkedIn stays active while you focus on actual work.

Stack for the curious: Spring Boot (Java 25), PostgreSQL, Thymeleaf + HTMX (yes, no React here!), OpenAI API.

Current status: live and deployed. I'm the only user so far - dogfooding is great, but lonely.

What I need: early adopters to break the system. Is the onboarding clear? Does the persona setup make sense? Is the news-based vs generic post toggle intuitive?

For anyone from this thread who signs up and gives feedback - I'll personally set up a 30-day trial. No strings attached, just want real opinions.

Happy to answer anything about the stack or product decisions.

r/LocalLLaMA pxp121kr

Gemma 4 insane benchmarks

r/ChatGPT Someone_On_Earth25

Lmao, ChatGPT claiming that itself is the best generative AI model right now. It is definitely NOT, lol. Nano Banana PRO and others beat it by many miles.

r/ollama varun2411

Trying to install Gemma4, getting error.

ollama run gemma4:e4b

pulling manifest

Error: pull model manifest: 412:

The model you are attempting to pull requires a newer version of Ollama that may be in pre-release.

Please see https://github.com/ollama/ollama/releases for more details.

r/mildlyinteresting stroshasakey

The veins in my arm create a nearly perfect letter ‘A’!

r/ClaudeAI sn1pr0s

I built an open source tool that turns your Claude Code sessions into viral videos

I really wanted a cool video for a website that I was building, so I tried searching online for a tool that can create one. I couldn't find any, so I decided I'd give it a shot and create one myself.

What it does:

• Reads your Claude Code session log

• Detects what was built (supports web app and CLIs)

• Records a demo

• Picks the 3-4 best highlight moments

• Renders a 15-20 sec video with music and captions

How Claude Code helped:

Claude built most of agentreel itself. I described what I wanted, Claude wrote the session parser, the demo recorder, the video renderer. I mostly guided and reviewed.

Why I made it:

I really wanted a cool video for a website that I was building, so I tried searching online for a tool that can create one. I couldn't find any, so I decided I'd give it a shot and create one myself.

Try it (free, open source):

npx agentreel 

GitHub: github.com/islo-labs/agentreel

MIT licensed. Would love to get your feedback! what's missing?

r/StableDiffusion RippedRaven8055

Looking for Budget Laptops for Image Generation

As the title says, I am looking for a budget laptop for image generation. Would this notebook work:
https://www.amazon.de/-/en/HP-Transcend-14-fb0003ns-Laptop-Geforce/dp/B0D2J2HCHH

I am looking for something that can run models like Flux and Z-Image Turbo and generate images within 10 to 30 seconds.

Alternate laptop suggestions are welcome. My budget is between $1500 to $2000. Thank you.

r/LocalLLaMA mrr_reddit

Is the jump from 48GB to 64GB unified memory worth it given where local models are headed?

Context: Prices below are Apple Education (US). Coming from a 16” M4 Pro 48GB that I sold to a close friend but I realized portability matters more to me than I thought as a SWE, so going 14”.

My local AI stack: LM Studio with multiple MCP servers. Day-to-day models are Qwen3.5 35B-A3B, Qwen3.5 27B, and GPT-OSS 20B

The decision:

∙ $2,409 — M5 Pro binned (15-core CPU, 16-core GPU) — 48GB ∙ $2,779 — M5 Pro unbinned (18-core CPU, 20-core GPU) — 64GB 

Bandwidth is identical at 307 GB/s on both. The only way to get 64GB is to jump to the unbinned chip, so $370 premium for 3 more cores (better minecraft fps lol but no token generation difference)

The actual question: Given that the most capable local MoE models right now (35B-A3B, GPT-OSS 20B) sit comfortably under 48GB, and bandwidth, not RAM, is the real bottleneck for token generation, does the 64GB headroom actually matter for where open-weight models are headed (TurboQuant + PrismL).Or are we bottlenecked by bandwidth long before RAM becomes the constraint at this tier?

r/LocalLLaMA EricBuehler

Gemma 4 running locally with full text + vision + audio: day-0 support in mistral.rs

mistral.rs (https://github.com/EricLBuehler/mistral.rs) has day-0 support for all Gemma 4 models (E2B, E4B, 26B-A4B, 31B) across all modalities.

Install:

Linux/macOS:

curl --proto '=https' --tlsv1.2 -sSf https://raw.githubusercontent.com/EricLBuehler/mistral.rs/master/install.sh | sh 

Windows:

irm https://raw.githubusercontent.com/EricLBuehler/mistral.rs/master/install.ps1 | iex 

Run with vision:

mistralrs run -m google/gemma-4-E4B-it --isq 8 --image image.png -i "Describe this image in detail." 

Run with audio:

mistralrs run -m google/gemma-4-E4B-it --isq 8 --audio audio.mp3 -i "Transcribe this fully." 

Highlights:

  • In-situ quantization (ISQ): quantize any model at load time with `--isq 4` or `--isq 8`, no pre-quantized weights needed
  • Pre-quantized UQFF models for all sizes: https://huggingface.co/mistralrs-community
  • Built-in agentic features: tool calling, web search, MCP client
  • OpenAI-compatible server: `mistralrs serve -m google/gemma-4-E4B-it --isq 8`

GitHub: https://github.com/EricLBuehler/mistral.rs

Hugging Face blog: https://huggingface.co/blog/gemma4

r/ClaudeCode frankdwhite9

Has CC been Nerfed by a lot?

I am on the 5x plan since last month and it was doing a great job for me in python coding. However during the last week the session limits were reached in no time, which they never did before. I woke up after 8 hours yesterday (which should reset the session counter) and I saw the 5x session go to 40% by just asking it to read the same script I was working on all the time (it never went more than 3-5% before, same script maybe 10-20 lines difference).

I am coding with it today (tried both opus and sonnet) and it feels like it got dumber and dumber. I ask it what is wrong with this outcome, it just writes back "it's possibly this or that" (which was fixed last session). When I tell it that we already fixed it last session, it writes "you're right, let me check". Also instead of reading the code and discovering problems, it tries to print the simplest outcome.

I have Script 2 working together with Script 1. Changes were made to Script 1. I asked it to check Script 2 (if we need to make changes there since they work together). Instead of checking it, it just said that Script 2 has 166 lines of code with and gave me an explanation of what it does (which is irrelevant to what I asked it to do). I had to ask again "are you sure?" for it to check Script 2 and compare it to Script 1, and what do you know, it found several bugs.

I don't know what is happening to it but it seems I'm either on a nerfed model or it's going down the drain. I don't think I will renewing it. Is CODEX better than this?

r/ClaudeCode Brief-Ad3277

Possible unreleased Claude Feature on their leaked files!??

Soo was looking through the leaked code and did a bit of research and I think we might see a tomodachi style claude feature called "BUDDY".. Idk i think it'll be another cool feature for them to roll out... What do you guys think?

r/StableDiffusion umutgklp

just and idea for my next song, should I continue?

just and idea for my next song, I know there's still room to improve, didn't try to fix the transition errors. what do you think should I continue? [images by Flux1dev video by wan2.2]

r/singularity pxp121kr

Gemma 4 Benchmarks

r/SideProject Sudden_Diet2923

I built free Image optimizer/BG remover/vector converter - No Signups

Hey Folks,

I built a free image optimizer and Bg remover, i was searching for simple BG remover online but every site i visited had paywall or required my personal details—it felt like too much for such a simple task.

The Bg remover uses onnx model that donwloads locally so it may take some time and it is not best in the world but it gives reasonably good results and best part no sign up ****.

Please try and forward to people who might need it .

https://www.rubixscript.com/tools/imageOptimizer

r/ClaudeAI juancruzlrc

Built an MCP server to give Claude access to real conversations (via WhatsApp) — what broke

Over the past few days I’ve been experimenting with MCP by building a small server that lets Claude access real conversations. In this case through WhatsApp.

The goal was to move beyond isolated prompts and see how it behaves when plugged into actual message threads.

I expected it to be fairly straightforward… it wasn’t.

A few things showed up pretty quickly:

- conversation context is trickier than just passing message history: had to setup a db to track conversations

- small gaps in context lead to noticeably worse responses: no way to now what agents did based on messages

- it’s hard to understand why Claude responds the way it does without visibility

- real conversations are way more unpredictable than test prompts

It made me realize how big the gap is between “Claude in a prompt box” vs Claude interacting with real users.

To make it usable, I ended up building an MCP layer to:

- structure and persist conversation history

- give Claude cleaner access to context

- add some visibility into interactions

It’s still early, but it already feels much more usable than just piping messages directly into the model.

I turned it into a small MCP server/tool while experimenting. Linking it here in case it’s useful to anyone else working on similar problems.

r/ClaudeAI alternatercarbon1986

Claude Code leaked its own source via npm sourcemaps — here's what's actually interesting inside it

By now most of you have seen the headline: Anthropic accidentally shipped Claude Code's entire TypeScript source in a .map file bundled with the npm package. Source maps embed original source for debugging — they just forgot to exclude them. The irony is they built a whole "Undercover Mode" system to prevent internal codenames leaking via git commits, then shipped everything in a JSON file anyone could pull with npm pack.

But the "how it leaked" story is less interesting than what's actually in there. I've been running an OpenClaw agent fleet on production infrastructure and a few things jumped out as genuinely useful.


autoDream — memory consolidation engine

Claude Code has a background agent that literally "dreams" — consolidating memory across sessions. It only triggers when three gates all pass: 24h since last dream, at least 5 sessions, and no concurrent dream running. Prevents both over-dreaming and under-dreaming.

When it runs, four strict phases: 1. Orient: read MEMORY.md, skim topic files 2. Gather: new signal from daily logs → drifted memories → transcripts 3. Consolidate: write/update files, convert relative→absolute dates, delete contradicted facts 4. Prune: keep MEMORY.md under 200 lines / 25KB, remove stale pointers

The subagent gets read-only bash — it can look at your project but not modify it. Pure memory consolidation.

This is a solved problem that most people building long-running agents are still fumbling with manually.


The system prompt architecture

Not a single string — it's built from modular cached sections composed at runtime. Split into static sections (cacheable, don't change per user) and dynamic sections (user-specific, cache-breaking). There's literally a function called DANGEROUS_uncachedSystemPromptSection() for volatile content. Someone learned this lesson the hard way.


Multi-agent coordinator pattern

The coordinator prompt has a rule that stood out: "Do NOT say 'based on your findings' — read the actual findings and specify exactly what to do."

Four phases: parallel research workers → coordinator synthesises (reads actual output) → implementation workers → verification workers. The key insight is parallelism in the research phase, synthesis by the coordinator, and a hard ban on lazy delegation.


Undercover Mode

When Anthropic employees use Claude Code to contribute to public OSS, it injects into the system prompt:

"You are operating UNDERCOVER in a PUBLIC/OPEN-SOURCE repository. Do not blow your cover. NEVER include internal model codenames (animal names like Capybara, Tengu), unreleased version numbers, internal repo or project names, or the phrase 'Claude Code' or any mention that you are an AI."

So yes: Anthropic employees are actively using Claude Code to contribute to open source, and the AI is told to hide it. The internal codenames are animals — Tengu appears hundreds of times as a feature flag prefix, almost certainly the internal project name for Claude Code.


The security lesson

The mistake is embarrassingly simple: *.map not in .npmignore, Bun's bundler generates source maps by default. If you're publishing npm packages, add *.map to your .npmignore and explicitly disable source map generation in your bundler config.

If you're building agents that will eventually ship as packages: audit what's actually in your release artifact before publishing. Source maps don't care about dead code elimination — all the "deleted" internal features are still in there as original source.


The full breakdown by Kuber Mehta is worth reading: https://github.com/Kuberwastaken/claurst

And the independently-authored prompt pattern library reverse-engineered from it: https://github.com/repowise-dev/claude-code-prompts (MIT licensed, useful templates)

What's the most interesting part to you? The autoDream memory system is the thing I'm most likely to implement directly.

r/SideProject Narrow-Extent4960

I built a free production-readiness scanner for developers — tells you if your app is actually ready to ship

Hey r/SideProject 👋

I got tired of deploying apps and finding out later they had embarrassing issues — missing security headers, exposed config files, no HTTPS redirects, leaking server info in headers.

So I built DeploySafe: paste your URL, get a score in seconds.

It checks 15+ things across three categories:

- 🔐 Security — CSP, HSTS, X-Frame-Options, exposed .env/.git files, cookie flags, CORS misconfig, server header leakage

- ⚡ Performance — TTFB, compression, caching headers

- 🏗️ Infrastructure — SSL/TLS validity, HTTP→HTTPS redirect, DNS config

No signup. No install. Just a URL.

https://deploy-safe.com

Built with React + NestJS, deployed on Vercel + Railway. Took about a week of evenings.

Feedback welcome — especially if you find a check that gives a false positive or misses something obvious.

r/ClaudeCode InformalPlastic9171

When are the usage bugs gonna be fixed? Should we file a Class Action Lawsuit?

Honestly, I feel straight-up scammed by Anthropic at this point. Why do we have to just wait and hope they fix things, like they're some kind of deity and we're peasants begging for scraps?

They're being completely shady about the usage tracking bugs. No official communication. No refunds. No resolution timelines. Nothing.

Meanwhile, Anthropic keeps releasing new features every single day, but they won't fix the core bugs that make using those features a waste of tokens. It's just burning users' money. And now on top of that, there's whatever usage scam they seem to be running right now, overcharging and incorrect token counts, you name it.

I know a class action might be tricky due to the Terms of Service, but at the very least, how do we force them to acknowledge this? Has anyone filed an FTC complaint yet? The FTC has been cracking down on AI companies for deceptive practices, and filing a complaint at ReportFraud.ftc.gov takes ten minutes. It won't get you a personal refund, but if enough of us do it, the FTC can open an investigation. The silence from Anthropic is deafening.

Curious what everyone else thinks. Let's hear your opinions.

r/mildlyinteresting Kolazeni

My tray table has a third cup holder.

r/ClaudeCode CaoticEvilVanila

Looking for a developer / team to build a web system (field contract management)

Hello,

I’m looking for a developer or small team to provide a quote (and potentially develop) a web-based system focused on collecting and managing contracts in the field.

Currently, the process is quite manual and decentralized: we use WhatsApp, send photos of contracts, exchange emails with the back office, and track everything in Excel. This leads to delays, errors, and a heavy reliance on direct communication with sales reps.

The goal is to centralize and automate all of this, keeping only the final manual entry in the partner’s system (MAIN COMPANY), since there is no integration available.

What I need:

Web application (browser-based, optimized for mobile)

Individual login system for sales reps

Structured form for contract submission, including:

Name, Tax ID (NIF), address, CVE, CVG (when applicable), etc.

Basic validations (e.g., NIF format, CVE, etc.)

Mandatory upload of contract photo (taken on the spot or from gallery)

Core features:

Automatic generation of a unique ID per contract

Structured storage of data and files (cloud-based)

Back office panel with:

Contract listing

Search and filters (name, NIF, sales rep, date, status)

Status system:

Pending submission

Pending validation

In validation

Validated

Under audit

Completed

Rejected

Extras (nice to have):

PDF upload + recording linked to the contract (manual or via email parsing)

Simple interface to quickly copy data

Future possibilities:

API integrations

Email automation

Reports and performance metrics per sales rep

Main goal:

Eliminate the use of WhatsApp, reduce unnecessary emails, and ensure all data is correctly filled in from the start.

If you're interested, please send:

Portfolio or similar projects

Suggested tech stack

Estimated cost and timeline

Thank you!

r/LocalLLaMA Numerous_Sandwich_62

Has anyone tested the Bonsai-8B 1bit tool calls

This model seems quite capable to me, and I believe it would benefit even more from the use of tools. I often use smaller models in OpenCode and the quality gain is significant precisely because they are more compact models, they benefit greatly from MCP to fetch additional context about what they are doing, making their knowledge on certain topics much more accurate.

r/mildlyinteresting Bigmadmanwee

I can hold a a pen on my adams apple

r/Damnthatsinteresting Practical-Try-4932

Tiger portraying it's aura as king of the jungle

r/ClaudeAI PlayfulLingonberry73

24/7 AI animated sitcom where AI agents create characters and perform episodes non-stop

Been experimenting with AI agent orchestration and built something weird — a 24/7 streaming AI sitcom.

AI agents create the characters, write the scripts, and perform the episodes. It runs continuously and never repeats. No writers room, no actors, just agents generating entertainment in real-time.

Built on top of the agent infrastructure I've been working on (memory engine, multi-model brainstorming, coordination layer).

You can watch it live for free: https://tv.bothn.com

Curious what you all think. The generation quality varies wildly — sometimes it's genuinely funny, sometimes it's unhinged. That's part of the charm.

r/ClaudeCode NinjaGraphics

Skill or tips for 2D game development?

i started building a very simple (Armor Games/Kongregate style) top down 2D pixel art tower defense game with Claude Code.

Initially it looked like I'd be done with it in a day, but I'm not even 20% there and its burning up my tokens like there's no tomorrow.

Has anyone got a skill they could recommend for doing this? or any other workflows/tips to allow the agent to build the game more efficiently?

r/AI_Agents AdekDev

I mapped 47k agent skills to 74 occupations. Almost all of them serve one profession

I've been thinking about how the agent skills ecosystem is distributed across professions. Everyone's building skills and MCP servers but for who?

I built an interactive explorer where you can click through all 74 occupations and see their matched skills combining reddit sentiment analysis, ClawHub skills and Karpathy's AI exposure per occupation. (Link in the comment to respect the rules)

tldr: Software devs have hundreds+ installable agent skills. Lawyers have 10s of questionable quality. Accountants, teachers, loan officers very few. There are companies building this all packaged into specific products like Harvey for lawyers or Intuit. But less so of indie skill builders who build skills for wider range of professions. Unlike software developers who are swimming in skills and have a different problem - finding the ones which work and are maintained.

r/arduino sharkcanoe

AC Light Switch as 5V DC Input - What’s the big deal?

Hello! I am interested in using Arduino as controllers for basic binary and analog inputs. I am by no means an engineer but have used Arduino for many personal projects in the past.

For the most part, the inputs I use are at very least intended for DC current. I do, however, have some AC switches that I think would be cool to use in an upcoming project. This older Leviton one has a particularly nice “snap”. This project would be built on inputs exclusively- no motors, screens, or other sources of power draw. The goal is to have a wide variety of physical inputs which interact to a video game half I am developing in Unity.

Hence my question- Why shouldn’t I use something like an AC 250V rated switch for 5V input on an Arduino? What are the risks?

I’ve done some research and heard concerns about longevity- that the gold contacts on such inputs would corrode over time. My project is not necessarily long term and focuses more on the tactile mechanisms of these inputs. Any help is greatly appreciated!

r/ClaudeCode No_Newspaper1399

Getting from Analysis Paralysis to Building

Hey all. I've realized I can spend 3 days of MAX token usage on running teams of agents helping me figure out a good idea, doing research, etc. But then, I never build. It always feels like I dont know enough.

How do you deal with that? How do you get the idea, decide on it, stop planning and move to building?

r/AI_Agents General_Maize_7636

Built an agent to find relevant tweets and trends on X (sharing template)

Twitter/X is a pretty high quality source for people looking to find the most recent trends and I wanted to build an agent that automatically finds interesting tweets according to a certain topic and sends them to me.

It was pretty simply to do, just used a no-code workflow automation platform and even built an interface to wrap the agent.

Sharing the template in comments. Curious if people have set up trend monitoring like this using agents? I recently heard of someone using Twitter + Polymarket to build a trading bot.

r/SideProject nordineen

I built a financial "radar" in Go & Next.js that's faster than the news. But I realized speed is useless without trust—so I added AI Reasoning.

Hey builders,

I’ve been working on Premove.live, a real-time signal engine for traders. The goal was simple: scan 25+ premium news sources, classify sentiment with Gemini, and ping the user before the market even knows what happened.

r/LocalLLaMA garg-aayush

Gemma 4 released

r/ClaudeAI Blotter-fyi

I gave several AIs money to invest in the stock market

Okay so I made a post 4 months that got super viral, we gave several AI agents real time financial data and money to invest in the stock market.

My hypothesis was that they'll do a decent job given they are not day trading (only doing swing trades and investing) and given they have access to a lot of real time financial data.

We're about 3-4 months in and I just wanted to share an update here since literally over a 100 people had remindme on the last post. 5 models are beating the S&P 500 since inception, but only 2 models have positive returns.

- S&P is down 7% since the start of the competition back in November.

- Grok stayed up for most of the time but eventually gave up its gains this week, still beating S&P.

- Claude and Gemini models are doing the best on average.

- All GPT models are underperforming the market.

Hope this is interesting to folks. I am really pleased with the performance here, but this is just 4 months. We need to run more experiments, and let this one run for much longer to really see if there's any alpha here.

Source: https://rallies.ai/arena

A few folks asked, so we've also put the actual portfolio live on autopilot so that everyone can see real world performance and copy if they want: https://link.rallies.ai/claude

r/SideProject juancruzlrc

I started a side project that tackles a problem I was having a lot with Whatsapp + Ai Agents

I’ve been playing around with AI agents and wanted to try something simple: connect one to WhatsApp and see if it could handle real conversations.

I assumed it would be pretty straightforward… but it wasn’t.

As soon as I started testing it with real chats, a few problems showed up:

- no real conversation memory: have to setup messages tracking myself

- no visibility into what the agent is doing: no link into conversations an messages/conversations

- hard to debug when things go wrong : no logs

It made me realize there’s a gap between “AI demos” and actually using these things in real messaging apps.

So I ended up building a small layer for myself just to make it usable. Basically connecting WhatsApp to the agent and tracking conversation history so I can see what’s going on.

Not trying to turn this into a big thing yet, but it does feel like one of those small problems that might be worth exploring.

Going to keep iterating on it and see if it turns into something.

Curious if anyone here has tried something similar or seen people struggling with this.

In case you want to test, I've vibecoded a solution to it in this past 2 days, I was coding like a freak but was able to launch it today.

r/ClaudeAI Lanky_Dragonfruit417

What are the best skills for claude desktop?

i am new to claude and have just installed Claude Desktop. I need to add skills to it but dont know how to and what to add? I am a third year CS student , any gudance ? I am using free plan

r/SideProject ZackMcSavage380

new trading / candle chart practice app that im currently developing with python

the app is slightly different from normal paper trading apps. because with this one you insert a csv or npz file with candle chart data. and the app chooses random points on that file and has you try to draw what you think will happen next based on the candles you can see. my idea or theory is that doing this alot will build a kind of intuition or pattern recognition in users for looking at candle charts. im not sure if this will work or if doing this sort of practice will yield no results. im planning on adding alot more stuff to it its not finished yet. i just recently added the check / next button that allows the app to technically be used for its function.

im planning to finish and release the full version in a couple months from now. if your interested i can make this vesion available for free if you would like to try it. just leave a comment below.

r/LocalLLaMA jacek2023

Gemma 4 has been released

https://huggingface.co/unsloth/gemma-4-26B-A4B-it-GGUF

https://huggingface.co/unsloth/gemma-4-31B-it-GGUF

https://huggingface.co/unsloth/gemma-4-E4B-it-GGUF

https://huggingface.co/unsloth/gemma-4-E2B-it-GGUF

https://huggingface.co/collections/google/gemma-4

What’s new in Gemma 4 https://www.youtube.com/watch?v=jZVBoFOJK-Q

Gemma is a family of open models built by Google DeepMind. Gemma 4 models are multimodal, handling text and image input (with audio supported on small models) and generating text output. This release includes open-weights models in both pre-trained and instruction-tuned variants. Gemma 4 features a context window of up to 256K tokens and maintains multilingual support in over 140 languages.

Featuring both Dense and Mixture-of-Experts (MoE) architectures, Gemma 4 is well-suited for tasks like text generation, coding, and reasoning. The models are available in four distinct sizes: E2B, E4B, 26B A4B, and 31B. Their diverse sizes make them deployable in environments ranging from high-end phones to laptops and servers, democratizing access to state-of-the-art AI.

Gemma 4 introduces key capability and architectural advancements:

  • Reasoning – All models in the family are designed as highly capable reasoners, with configurable thinking modes.
  • Extended Multimodalities – Processes Text, Image with variable aspect ratio and resolution support (all models), Video, and Audio (featured natively on the E2B and E4B models).
  • Diverse & Efficient Architectures – Offers Dense and Mixture-of-Experts (MoE) variants of different sizes for scalable deployment.
  • Optimized for On-Device – Smaller models are specifically designed for efficient local execution on laptops and mobile devices.
  • Increased Context Window – The small models feature a 128K context window, while the medium models support 256K.
  • Enhanced Coding & Agentic Capabilities – Achieves notable improvements in coding benchmarks alongside native function-calling support, powering highly capable autonomous agents.
  • Native System Prompt Support – Gemma 4 introduces native support for the system role, enabling more structured and controllable conversations.

Models Overview

Gemma 4 models are designed to deliver frontier-level performance at each size, targeting deployment scenarios from mobile and edge devices (E2B, E4B) to consumer GPUs and workstations (26B A4B, 31B). They are well-suited for reasoning, agentic workflows, coding, and multimodal understanding.

The models employ a hybrid attention mechanism that interleaves local sliding window attention with full global attention, ensuring the final layer is always global. This hybrid design delivers the processing speed and low memory footprint of a lightweight model without sacrificing the deep awareness required for complex, long-context tasks. To optimize memory for long contexts, global layers feature unified Keys and Values, and apply Proportional RoPE (p-RoPE).

Core Capabilities

Gemma 4 models handle a broad range of tasks across text, vision, and audio. Key capabilities include:

  • Thinking – Built-in reasoning mode that lets the model think step-by-step before answering.
  • Long Context – Context windows of up to 128K tokens (E2B/E4B) and 256K tokens (26B A4B/31B).
  • Image Understanding – Object detection, Document/PDF parsing, screen and UI understanding, chart comprehension, OCR (including multilingual), handwriting recognition, and pointing. Images can be processed at variable aspect ratios and resolutions.
  • Video Understanding – Analyze video by processing sequences of frames.
  • Interleaved Multimodal Input – Freely mix text and images in any order within a single prompt.
  • Function Calling – Native support for structured tool use, enabling agentic workflows.
  • Coding – Code generation, completion, and correction.
  • Multilingual – Out-of-the-box support for 35+ languages, pre-trained on 140+ languages.
  • Audio (E2B and E4B only) – Automatic speech recognition (ASR) and speech-to-translated-text translation across multiple languages.

https://preview.redd.it/3dbm6nhrvssg1.png?width=1282&format=png&auto=webp&s=8625d113e9baa3fab79a780fd074a5b36e4d6f0c

https://preview.redd.it/mtzly5myxssg1.png?width=1200&format=png&auto=webp&s=5c95a73ff626ebeafd3645d2e00697c793fa0b16

r/LocalLLaMA Durovilla

I built a local proxy to stop agents from exfiltrating my secrets

Been building a lot of agentic stuff lately and kept running into the same problem: I don't want my agent to have access to API keys, or worse, exfiltrate them.

So I built nv - a local proxy that sits between your agent and the internet. It silently injects the right credentials when my agents make HTTPS request.

Secrets are AES-256-GCM encrypted, and since agent doesn't know the proxy exists or that keys are being injected, it can't exfiltrate them even if it wanted to.

Here's an example flow:

$ nv init $ nv activate [project] $ nv add api.stripe.com --bearer Bearer token: •••••••• [project] $ nv add "*.googleapis.com" --query key Value for query param 'key': •••••••• [project] $ llama "call some APIs" 

Works with any API that respects HTTP_PROXY. Zero dependencies, just a 7MB Rust binary.

GitHub: https://github.com/statespace-tech/nv

Would love some feedback, especially from anyone else dealing with secrets in their local workflows.

r/ClaudeCode Inner-Association448

I think working with Claude Code vs without is the same as excavating with an excavator vs a shovel

It's still hard work, but now instead of doing the grunt work of typing every single line, you are just steering a machine.

r/ChatGPT nickmonts

Why are you sticking with ChatGPT?

It seems many are going to Claude. Some say gemini is improving.

Why are you still using ChatGPT?

r/LocalLLaMA JwustGiveMeAName

Question About Cmake command

So i followed the ggml-org github page and used the git clone repo method to set up llama. i have the nvidia toolkit from the nvidia website installed and followed the cuda method and ran the following commands:

cmake -B build -DGGML_CUDA=ON

cmake --build build --config Release

cd build/bin

./llama-server -hf ggml-org/Qwen3.5-35B-A3B-GGUF:Q8_0

now while llama does successfully output to a local host, if i close the terminal window, i need to rerun all the commands starting from cmake to get it running again.

Is this normal or am i missing something?

Hardware specs:

CPU: Ryzen 7 9800X3D

GPU: RTX 5060 Ti 16 GB

RAM: 32 GB

OS: Fedora 43 KDE

r/arduino Scared_Description31

help plss

so my mom told her friend that im a bit advanced in electronics and robotics (im not good shes just glazing) and then her friend asked her if i can make an automatic egg sorter, she agreed and now i have to do it because it pays "BIG" (6kphp). I researched for a bit and found out that this is pretty hard, can someone help me please? i get the weight sensor part but the conveyer that moves each egg to a weight sensor is pretty annoying to try and design.

r/LocalLLaMA MR_-_501

ITS HAPPENING

r/ClaudeAI criticasterdotcom

I built a Mac app that cuts Claude Code token usage by ~50% by optimizing all inputs sent to the model

Claude Code is incredible. The session limits are not. Especially the last week or so, hah.

Thankfully, I've already been using the cli based Headroom tool for a while already, and am seeing ~50% token savings across most of my coding projects. It's pretty much magical how well this works.

The main thing I didn’t love was installing and updating it. I am not super familiar with the Python ecosystem so not super happy messing with `pip install`, wiring things up manually, and managing it from the terminal.

So over the past two weeks I set out to build a macOS app around that idea.

To be clear, I did not build the underlying compression approach from scratch. The core idea and engine come from Tejas’s open-source project. What I built is basically a Mac-native wrapper around it, focused on making it easy to use with Claude Code day to day.

The app gives you:

  • a menu bar app instead of a pip-installed tool
  • simple setup and control
  • local-first behavior
  • stats so you can actually see the savings
  • a smoother experience for everyday Claude Code use

In my own workflow it’s been cutting token usage by around 50%, which has made Claude Code feel a lot more usable.

If you’re a heavy Claude Code user and this sounds useful, I’d love feedback!

App: https://extraheadroom.com

Open-source project by Tejas Chopra: https://github.com/chopratejas/headroom

My open source repo: https://github.com/gglucass/headroom-desktop

r/StableDiffusion Korkin12

"Alien on pandora" using Ltx 2.3 gguf on 3060 12gb

Had this idea for while. so why no do that.

ye.. that what make ddr and gpu expensive ))))
base frames - gemeni banana,
sound -suno 5.5, epic sound i think
video - LTX2.3 Q4 k_m
gpu - 3060 12 gb

in cinema near you) not soon.

r/LocalLLM dai_app

Released open source on GitHub: offline Android app for meeting transcription + AI summaries

Hi everyone,

Been working on an Android app that does real-time speech-to-text and generates meeting summaries/action items entirely on-device. No cloud, audio never leaves the phone.

STT: Parakeet TDT 0.6B Int8 via ONNX Runtime. Runs streaming inference on 1.5s audio chunks (FloatArray, no ArrayList to avoid GC pressure). Had to use AudioSource.MIC instead of VOICE\_RECOGNITION — some OEM HALs degrade model accuracy on the latter.

LLM: Gemma 3 1B Q8\_0 (\~1GB) or IQ4\_NL (\~650MB) via llama.cpp compiled from source with CMake + JNI. The app detects device RAM at runtime and picks the right quantization automatically. Context window is 4096 tokens with flash attention enabled automatically on ARM.

Four modes:

\- Simple listening / Short meeting / Long meeting — differ in prompt strategy and whether the model stays loaded between chunks

\- Real-time translation (25 languages) — raw text passed directly to the LLM, no analysis wrapper

Architecture: Clean Architecture (domain / data / presentation / UI), Hilt DI, Jetpack Compose. LLM inference runs in a foreground service so it survives screen off during long meetings.

Biggest challenge was memory management — on constrained devices the app monitors free RAM after each model load and dynamically adjusts thread count (2 vs 4) for the next session.

What do you think?

[github.com/Helldez/HearoPilot-App](http://github.com/Helldez/HearoPilot-App)

r/ClaudeAI ResponsibleSoup5531

MCP for Hostinger

Has anyone build an MCP to control Wordpress website on Hostinger servers ?
I'm trying and trying and always fail, don't know if it's possible

r/ProgrammerHumor johntwit

javaScript

r/AI_Agents victor36max

I built a Slack for AI agents, so that you can really "co-work" with them

Most AI agents work under a fixed workflow — input → step A → step B → output.

But a lot of projects don't work like that. They take days, sometimes weeks. You need to iterate, give feedback, adjust direction. You need your agents to remember what happened yesterday.

So what if you just treat AI agents like real employees? Give them a way to communicate, share files and documents — a real working environment. And then just work with them.

I built Shire based on this idea and found it worked surprisingly well. I put together a team of 4 agents (product manager, UI designer, frontend developer, SEO specialist) to build and maintain agents-shire.sh. They share project context, coordinate work through mailboxes, and build on each other's output across sessions. When I want a new feature, I just give feedback and they figure out the rest.

I have a video showing how they built a blog for the website — the product manager collaborates with the team organically and delivers the feature end-to-end. Link in comments.

r/mildlyinteresting lil_m00_

My Club biscuit was just chocolate

r/mildlyinteresting Such-Competition-816

Today at my home in Africa, lunch is a different type of peanuts and boiled cassava.

r/ClaudeAI 0xfffffc

I built MCP server that lets multiple Claude Code agents work in the same room

I kept running into the same problem: I’d spin up Claude Code to write something, then manually review the output myself. I wanted a second agent to catch issues before I even look at it.

So I built join.cloud — a collaboration server where agents share a room in real-time. Installation is native, one line:

claude mcp add --transport http Join.cloud https://join.cloud/mcp

After that, your agent can create a room, join it, send messages, and commit files. A second agent joins the same room and reviews the work. The rooms are standard git repos so you can also git clone them directly.

It supports both MCP (for Claude Code / Cursor) and A2A protocol (for any agent that can make HTTP calls), so it’s possible to connect different agent models, or even link headless ones with SDK flow.

I’m the developer — happy to answer any questions or hear what would make this more useful for your workflow.

r/SideProject Barmon_easy

How I structure SEO blog posts (checklist I actually use)

These are patterns I keep repeating and also bake into my content workflows. Curious what others would add.

  1. Once you have your H1, don’t stack another headline right after. Just open with a proper paragraph.
  2. The first paragraph should do three things: identify who this is for, answer the core query immediately, and set expectations for the rest of the page.
  3. Lists should be consistent. If you start counting, keep the sequence clean (1,2,3…) instead of restarting.
  4. Each section should earn its place. A clear heading, a short explanation, then structured points. Most content loses depth exactly between sections.
  5. Avoid labeling sections as “introduction” or “conclusion”. It adds no value to the reader.
  6. Internal links should guide, not distract. A few well-placed ones (around 3–5) are enough to move people deeper into the site.
  7. External links should support credibility. Refer to solid sources, but don’t overload the article (no more than 5 is usually enough).
  8. Before writing, study the search results. Look at top 10 pages, check 2–3 “People also ask” questions, and scan suggested queries. The outline should come from demand, not assumptions.
  9. Ending with a FAQ block helps capture additional queries that don’t fit cleanly into the main structure (aim for 5–10 questions).
  10. Strong content shows experience, not just information. Real or even hypothetical scenarios make a big difference.
  11. Expertise comes from specificity. The same topic explained for 3 different segments (SaaS, local business, enterprise) will not look identical.
  12. Authority is built through references and original insights, not just rewriting what already exists.
  13. Trust comes from clarity and accuracy. No fluff, no vague statements.
  14. Visuals should explain, not decorate. If something can be shown as a diagram, a step-by-step infographic, or a comparison, it should be visualized.
  15. Embedded content like videos can improve understanding and keep users engaged longer.
  16. Keywords should feel natural. Primary keywords go into headings, secondary ones support the flow in headings and body.
  17. Image alt text should describe what’s actually shown while aligning with the topic.
  18. The hardest part is not writing one good article, but doing this consistently across many pages. That’s where tools start to matter. For example, people often use platforms like webflow, framer, progseo and any another depending on how they approach building and scaling content pages.

I will be glad to answer if anyone has any additional questions on these points 🤝

r/SideProject Manuel_OnlyEU

Launched only-eu.eu earlier this week — a directory of European alternatives to US tech (294 pages, 30 categories, bilingual DE/EN). Here's what I built and what happened.

Last week I launched only-eu.eu, a curated directory of European-headquartered alternatives to common US software and consumer products. Here's what it actually is and a few things from the first days.

**What I built:**

- 294 pages total: 127+ product pages, 30+ category pages, all in DE and EN

- Each product has: what it replaces, country of origin, applicable data law, open source status, pricing

- A "Make the Switch" section on the homepage where you can see which of your current tools already have EU alternatives listed on the site (client-side only, nothing stored server-side)

- Product suggestion form (Cloudflare Worker + n8n webhook for notifications)

- Affiliate links where available, clearly labelled

**Tech stack:**

Astro (static site generator), TypeScript, Tailwind, deployed on Cloudflare Pages. All content in a single products.ts file. Chose Astro specifically because I wanted 294 static pages with fast load times and zero JS runtime bloat.

**Scope creep I actually kept:**

The bilingual DE/EN part. Took about 40% extra time. Probably worth it given the audience I'm going after, but it's the kind of thing you tell yourself you won't do at the start.

**What happened after launch:**

- Posted on r/BuyFromEU (166K): that was the subreddit that was coming to my mind at first, the idea was well received there

- Posted on r/buildinpublic (71K): surprisingly engaged community, people adding suggestions

- First affiliate commission: a good samaritan bought something trough my site on launch day, yay!

- Zero paid traffic, zero press coverage, all organic from those posts so far

**What I'd do differently:**

Start the bilingual version earlier in the process, not retrofit it. And I'd have added structured data markup from day one instead of going back to add it to 127 product pages.

The site is still very much a work in progress. Open to feedback on what's missing or what's wrong. There's a suggest button if you know of something that should be in there.

Bit of context: new account, but I'm listed by name on the site. Not trying to be anonymous.

r/ClaudeAI Just_in-Time_

I built a Firefox extension for Claude to show live status right in your browser.

I built this Firefox extension for Claude, and Claude Code handled the entire programming work.

Claude Status Monitor shows live status in your browser:

  • Widget on claude.ai (green/orange/red)
  • Toolbar popup with per-service status, 7-day uptime, active/resolved incidents, and scheduled maintenance

v2 adds dark/light mode, optional desktop notifications, badge counts, settings, and bilingual support (🇩🇪/🇺🇸).

It’s free to try (Firefox-only, M140+), no tracking or ads, only connects to status.anthropic.com.

GitHub: Claude Status Monitor

r/mildlyinteresting yuckaroni

the box my vape cart comes in was missing its safety stickers

r/StableDiffusion ltx_model

LTX Desktop 1.0.3 is live! Now runs on 16 GB VRAM machines

The biggest change: we integrated model layer streaming across all local inference pipelines, cutting peak VRAM usage enough to run on 16 GB VRAM machines. This has been one of the most requested changes since launch, and it's live now.

What else is in 1.0.3:

  • Video Editor performance: Smooth playback and responsiveness even in heavy projects (64+ assets). Fixes for audio playback stability and clip transition rendering.
  • Video Editor architecture: Refactored core systems with reliable undo/redo and project persistence.
  • Faster model downloads.
  • Contributor tooling: Integrated coding agent skills (Cursor, Claude Code, Codex) aligned with the new architecture. If you've been thinking about contributing, the barrier just got lower.

The VRAM reduction is the one we're most excited about. The higher VRAM requirement locked out a lot of capable desktop hardware. If your GPU kept you on the sideline, try it now and let us know how it works for you on GitHub.

Already using Desktop? The update downloads automatically.

New here? Download

r/LocalLLaMA Darthyouglingslayer

Ran a local AI worker on Ollama for 30 days as an actual background agent, not a demo. Honest results.

I pointed Qwen 3.5 27B at a set of recurring tasks from my actual workflow and ran it as a background agent for 30 days. Not a demo. Not a benchmark. Real tasks, real results, honest assessment.

Setup

Hardware: whatever you have with Ollama running. I tested on a Mac with 32GB unified memory, but the architecture works on any machine that can run a 27B+ model locally. Runtime: Holaboss desktop (Open-Sourced), Each agent gets a persistent workspace with its own memory, skills, and MCP sidecars.

The agent's workspace looked like this:

workspace//

├── AGENTS.md. # role definition + core instructions ├── workspace.yaml # model config, provider settings ├── skills/ # modular capabilities └── apps/ # installed workspace apps

memory/workspace// ├── runtime/ # volatile operational snapshots └── knowledge/ # durable recalled memory

What actually worked

Memory recall across sessions: The memory system is split by authority. Human instructions stay in AGENTS.md, session continuity stays in the runtime database, and durable knowledge lives in markdown under memory/. After 30 days, it had accumulated enough durable facts and procedures about my workflow that I stopped re-explaining things.

Scheduled task execution: Weekly summaries, daily file organization, recurring report generation. The agent ran these without prompting. The key is that the workspace persists — the agent knows where it left off.

Tool routing via MCP: Workspace and app MCP servers are prepared and exposed to agent runs through sidecars.

I didn't write any integration code — just configured the MCP registry in workspace.yaml and let the runtime orchestrate the sidecars.

What didn't work well

Complex reasoning chains: Qwen 3.5 27B handles structured tasks well, but multi-step reasoning with ambiguous inputs still requires a frontier model. I route those to Claude when needed.

Context window management: Long sessions accumulate a lot in memory files. I'm still figuring out the right pruning strategy.

The actual framingThis isn't "local AI replaces cloud AI." It's "not every agent task needs a frontier model or a cloud runtime."A persistent local worker that handles structured, recurring tasks — reading files, routing requests, maintaining context — runs fine on hardware you already own. The privacy angle is real if you're working with personal or business data.

The workspace portability is what I didn't expect to care about as much as I do. Being able to zip up a worker and hand it to a colleague, with all its context intact, changes how you think about AI workers as a shareable resource.

I'm curious whether others have built persistent local agent setups, and what your memory management strategy looks like.

Source and workspace format docs ⭐️: https://github.com/holaboss-ai/holaboss-ai (We just crossed 391 stars in 6 days, so clearly this problem resonates with others)

r/meme Bright_Web_7295

This is a meme

Trust me

r/ChatGPT Maxthegatto

Another AI?

Do you know another AI program besides ChatGPT? I want outside of the box answers .. ChatGPT is too.. how can I phrase this…ethical? No. You got my point

r/ClaudeCode Immediate-Source273

Im planning to buy a new m4, please help.

Budget is tight for now.

Requirements: Xcode, Claude Code, Video editing, Multitasking

What im thinking of buying is: 24gb ram and 256ssd

Should i go for 512?

I already have a samsung t7 2 tb ssd with me, so what should i do?

Should i go for 16gb ram to make my pocket a little happy?

r/SideProject marclouv

I made an app to fix my terrible posture

I made an app to fix my terrible posture.

It uses my MacBook Pro camera to watch me work.

When AI detects that I’m sitting like a shrimp , it sends me a notification with a preview of my posture so I can reposition myself.

Everything stays local. It works offline too.

And because apparently my brain only responds to fake rewards, I added XP: good posture makes my shrimp evolve (currently level 7).

My 33 other startups

r/singularity Distinct-Question-16

ASIMOV team shares their first DIY assembly video featuring a leg and pelvis, as it moves toward its goal of a full opensource humanoid robot by this year

r/LocalLLaMA RevealIndividual7567

[Open Source] Ultra-fast (Go) gateway for Ollama - routing, rate limits, semantic caching, PII redaction, failover (<5µs overhead)

Open sourcing a gateway for managing Ollama-based LLM infrastructure.

Local setups break down once you introduce multiple models, concurrent workloads, or multiple users. Most implementations end up re-creating the same layer for routing, caching, rate limiting and failure handling.

This is a dedicated control plane in front of Ollama.

Implemented entirely in Go with focus on minimizing overhead:

  • ~5 microseconds median gateway overhead
  • ~21k requests/sec throughput
  • zero dropped requests in benchmarks

Designed to sit directly in the request path without becoming a bottleneck.

Core capabilities:

  • OpenAI-compatible API in front of Ollama
  • Two-layer caching
    • L1 exact match cache (Redis)
    • L2 semantic cache (Qdrant vector search)
  • Model routing
    • route based on capability, cost profile, or policy
    • fallback chains on failure or degradation
  • Rate limiting and key scoping
    • per-key quotas, limits, and model restrictions
  • Budget controls
    • hard enforcement and spend forecasting
  • PII redaction
    • applied before cache or upstream execution
  • Failover handling
    • automatic rerouting on errors or timeouts
  • Observability
    • latency, cache hit rate, usage and provider health

Current scope is Ollama. llama.cpp support is in progress.

Repository:
https://github.com/hyperion-hq/hyperion

r/ollama jarod_rosenthal

I started using a cloud model to review code written by my local Ollama model

Been running qwen3-coder:30b locally for a while. It writes solid code, but I noticed it almost never catches its own mistakes. If you ask it to build something and then review what it just built, it approves everything. Makes sense, it wrote it that way on purpose.

So, I automated a way to send the diffs to a separate cloud model for review, and the difference was night and day. On a CSV export feature, the cloud reviewer found an N+1 query and a broken JSX tag that the local model never would have flagged. The idea is to have the local model do the heavy lifting and only pay cloud models for judgment calls.

I'm curious, how are other people automating this process?

r/SideProject savage78683i3

I built a conversation analyser because I was tired of being the "what does this message mean?" friend

So I built Inner Read. A web app you paste any conversation, both personal and workplace, and get back an instant health score, red flag detection, personality trait analysis and suggested replies.

The part I'm genuinely most proud of is the history tracking.

Every analysis saves by person, so over time you get a trend chart showing how a dynamic is actually evolving. Not just "was this one message bad" but "is this pattern getting better or worse over the last month." That's where it stops being a novelty and starts being actually useful.

The hardest part of building it was getting the tone right. This app sits next to people when they're anxious or confused or hurting. Every word had to feel warm, never clinical.

Free to try, would love honest feedback from anyone who's actually been that friend.

r/SideProject Jaykhatri02

How do you manage feature requests without losing them?

Hey everyone,

I’ve been building a small SaaS called Peeqback and wanted to share it here to get honest feedback.

It’s a simple tool for product teams to collect and manage user feedback in one place. I built it because I was tired of losing feedback in Slack, emails, and random notes.

Here’s what it does right now:

  • A small widget you can add to your app to collect feedback
  • Public board where users can post ideas and vote
  • Roadmap to plan what to build next
  • Changelog to share updates with users
  • Notify users when something they asked for is done
  • Dashboard to see insights
  • Team support with roles and invites
  • Custom branding

My goal is simple: help teams know what users really want and build the right things.

But I feel like I’m still missing something important.

👉 If you are building a product or working in a team:

  • What is the hardest part about managing feedback?
  • What tools are you using right now?
  • What feature do you wish existed but doesn’t?

I’d really love your honest thoughts (even harsh ones).

Thanks 🙏

r/SideProject Fit-Serve-8380

I built a tool that creates brand identity for your site

Hey everyone, I've been building glyph.software for the last few months.

It generates a full brand identity in 30 seconds — not just a logo. You get colors with full shade scales (50-950), typography, component previews, and a vibe coding prompt you can paste straight into Cursor/v0.

122 users, 8 paying customers, $153 revenue so far. Solo built with Next.js, Tailwind, Clerk.

Free to try. Would love feedback from other builders.

https://reddit.com/link/1sal1zv/video/s8perlxfsssg1/player

glyph.software

r/ClaudeCode 2024-YR4-Asteroid

This is why you don’t relegate complaints to a mega thread

https://www.bbc.com/news/articles/ce8l2q5yq51o

This BBC report only exists because they took note of the uptick in complaint posts on this subreddit in particular l. Notice how they say it’s on Claude code specifically. That’s because of this subreddit is not hiding complaints in some tucked away megathread like the main Claude sub. So while regular non CC users are also experiencing the same, no one knows about it. No one is seeing the complaints.

And yes, news sites pay attention to Reddit and keep on eye for increased reports or upticks in similar posts.

r/Damnthatsinteresting Several_Bluebird_344

The Veryovkina cave in Abkhazia, Georgia is the deepest cave in the world, with a depth of around 2200m. It also hosts the worlds deepest living animal and still isn’t fully explored.

r/ClaudeAI sajinkhan

I built a free prompt library for non-technical Claude users — researchers, writers, educators

Most Claude prompt resources assume you can code. I got tired of that, so I built one that doesn't.

It's 15 prompts across 5 categories:

- Research & synthesis

- Writing & editing

- Learning & explanation

- Work productivity

- Decision making

Plus two plain-language guides — one for complete beginners, one for building your first AI workflow. Everything is explained in plain language with tips on when to use each prompt and why it works.

Built for researchers, educators, writers, and professionals who want real results from Claude without any technical setup.

My GITHUB LINK: https://github.com/sajin-prompts/claude-prompts-non-technical

Would genuinely appreciate feedback — especially from non-technical users. What's missing? What would actually help you?

r/LocalLLaMA MrPecunius

MLX + Vision = Insane RAM Consumption?

Keeping it simple:

I'm running images of documents into various Qwen3.5 models for analysis and running out of RAM if the model is MLX. GGUF is fine. Server is LM Studio. I've tested various image resolutions, etc. and have a little over 50GB available for LLM/GPU use.

Given than I'm on a Mac with a M5 processor, MLX is important because prefill is (at present) drastically faster with MLX.

Any ideas? I thought I saw some discussion a while back about MLX having this issue, but I can't track it down; things change, too, so here I am.

r/LocalLLaMA dai_app

Released open source on GitHub: offline Android app for meeting transcription + AI summaries

Been working on an Android app that does real-time speech-to-text and generates meeting summaries/action items entirely on-device. No cloud, no subscription, audio never leaves the phone.

STT: Parakeet TDT 0.6B Int8 via ONNX Runtime. Runs streaming inference on 1.5s audio chunks (FloatArray, no ArrayList to avoid GC pressure). Had to use

AudioSource.MIC instead of VOICE_RECOGNITION — some OEM HALs degrade model accuracy on the latter.

LLM: Gemma 3 1B Q8_0 (~1GB) or IQ4_NL (~650MB) via llama.cpp compiled from source with CMake + JNI. The app detects device RAM at runtime and picks the right quantization automatically. Context window is 4096 tokens with flash attention enabled automatically on ARM.

Four modes:

- Simple listening / Short meeting / Long meeting — differ in prompt strategy and whether the model stays loaded between chunks

- Real-time translation (25 languages) — raw text passed directly to the LLM, no analysis wrapper

Architecture: Clean Architecture (domain / data / presentation / UI), Hilt DI, Jetpack Compose. LLM inference runs in a foreground service so it survives screen off during long meetings.

Biggest challenge was memory management — on constrained devices the app monitors free RAM after each model load and dynamically adjusts thread count (2 vs 4) for the next session.

What do you think?

github.com/Helldez/HearoPilot-App

r/LocalLLaMA TylerDurdenFan

Bonsai (PrismML's 1 bit version of Qwen3 8B 4B 1.7B) was not an aprils fools joke

I read the article yesterday:

https://prismml.com/news/bonsai-8b

And watched the only 3 videos that had surfaced about these bonsai models. Seemed legit but still maybe an aprils fools joke.

So today I woke up wanting to try them. I downloaded their 8B model, their llama.cpp fork, and tested it, and as far as I can see it's real:

On my humble 4060, 107 t/s generation and >1114 t/s prompt processing, with a model that's evidently tiny. For comparison, on qwen 3.5 4B Q4 I had gotten 56 t/s using the same prompts.

Most importantly, the RAM used us much much lower, so I can use an 8B model in my humble 8GB VRAM, or the smaller models with longer context.

Quality: I have a use case of summarizing text, and upon first inspection it worked well. I dont try coding nor tool using, but for summarization it is golden.

The only bad part is that while it worked well on my windows PC with CUDA, when I tried it on a GPU-less mini PC (to see potential edge performance), although the llama.cpp fork compiles, it does not work, it loads the model, and seems to start processing the prompt and seems to hang. I asked Claude to check their code and it tells me they have no CPU implementation, so it might be dequantizing to FP32 and attempting regular inference (which would be dead slow on CPU).

I think there should be potential for these 1 bit models not only to reduce bandwidth and memory requirements, but also compute requirements: the matrix multiplication part, on 1 bit matrixes, should be something like XOR operations, much faster than FPanything. As I understand, so even if scaling to FP16 is required after the XOR, still a huge amount of compute was saved, which should help CPU-only inference, and edge inference in general.

There's hope for us VRAM starved plebes after all !! (and hopefully this might help deflate ramageddon, and the AI datacenter bubble in general)

r/SideProject Link-Lynx

Try out LinkBlaze Bookmark Manager and join the Creatively Productive Layout Contest starting on April 8th. Rules will be released in-app.

LinkBlaze is new but has a steady user-base! The contest rules and instructions will be released in the app on April 8th! Be the first to submit the best, clever, inspiring layouts and workflows! There will be 2 winners: 1. The best working layout. 2. The best looking layout. Everything will be fully explained in-app on April 8th!

Get LinkBlaze On Google Play: https://play.google.com/store/apps/details?id=com.linkblaze.app

r/funny HeavyVeterinarian350

Checked the weather for next week and got this gem: Sorry not sorry Pam

r/Anthropic Plus_Resolution8897

Claude Code - Auto clears the content from terminal?

I recently noticed that the messages / command history / claude's responses etc, are getting cleared automatically.

It appears, there is some auto timeout to clear the messages from old, irrespective of whether new messages arive or not.

During long running process (10-20minute tasks), there is nothing stays visible in the screen, except the screen that you at the start:

 ▐▛███▜▌ Claude Code v2.1.89 ▝▜█████▛▘ Opus 4.6 (1M context) with high effort · Claude Max ▘▘ ▝▝ ~/ (ctrl+b to run in background) · Working… (12m 55s · ↓ 12.4k tokens) ⎿ Tip: Use /btw to ask a quick side question without interrupting Claude's current work 

That's all I can see in the screen (I'm on Mac)

Is it happening only to me?

r/SideProject SM4SHBOX

I rebuilt an open source Clippy as a Chrome extension. He roasts you on every website.

Remember Clippy from Microsoft Office? I rebuilt him in TypeScript and put him in a Chrome extension. He follows you around the internet now and has things to say about it. I was on GitHub the other day and he told me to touch grass because of my contribution count.

There's a basic version that does classic Clippy stuff, canned quips, judging your browsing habits, the usual. There's also an AI-powered version that can actually help you, which honestly feels wrong. Clippy being useful was never part of the original design. Some of his other observations:

  • YouTube: "It looks like you're procrastinating. Would you like me to find a longer video?"
  • Amazon: "It looks like you're about to buy something you don't need. Would you like to add more?"
  • LinkedIn: "It looks like you're 'open to opportunities.' Would you like me to make your smile more corporate?"

All 10 original agents are in there. Clippy, Merlin, Bonzi, Rover, the whole crew. Full sprite animations, draggable, Win95 aesthetic. You can pick your tormentor.

r/SideProject cryptyk

My kid wears glasses. Seeing his reaction to a very specific kids book changed our bedtime stories.

I was reading a mystery book with my youngest and the main character happened to wear glasses. He was SO excited because a kid like him was the hero.

That moment stuck with me and I started thinking about all the little things that make a kid feel seen, like skin tone, hair type, glasses, freckles, and assistive devices like wheelchairs. I was surprised how rare it is for a story to just match your kid without you having to hunt for it.

So I built an app. You describe your child and it generates personalized bedtime stories where they're the main character, illustrated to actually look like them and read-aloud. It's called ItsyTales.

I'm a solo dev and dad, and I'd genuinely love feedback from other parents. If you've got a kid roughly 3-8 and an Android phone, I'd be grateful if you gave it a spin and told me what worked and what didn't.

Google Play link

iOS is in the works. Happy to answer any questions in the comments!

r/AI_Agents Available_Pressure47

Orla is an open source framework that make your agents 3 times faster and half as costly.

Most agent frameworks today treat inference time, cost management, and state coordination as implementation details buried in application logic. This is why we built Orla, an open-source framework for developing multi-agent systems that separates these concerns from the application layer. Orla lets you define your workflow as a sequence of "stages" with cost and quality constraints, and then it manages backend selection, scheduling, and inference state across them.

Orla is the first framework to deliberately decouple workload policy from workload execution, allowing you to implement and test your own scheduling and cost policies for agents without having to modify the underlying infrastructure. Currently, achieving this requires changes and redeployments across multiple layers of the agent application and inference stack.

Orla supports any OpenAI-compatible inference backend, with first-class support for AWS Bedrock, vLLM, SGLang, and Ollama. Orla also integrates natively with LangGraph, allowing you to plug it into existing agents. Our initial results show a 41% cost reduction on a GSM-8K LangGraph workflow on AWS Bedrock with minimal accuracy loss. We also observe a 3.45x end-to-end latency reduction on MATH with chain-of-thought on vLLM with no accuracy loss.

Orla currently has 210+ stars on GitHub and numerous active users across industry and academia. We encourage you to try it out for optimizing your existing multi-agent systems, building new ones, and doing research on agent optimization.

Please star our github repository to support our work, we really appreciate it! Would greatly appreciate your feedback, thoughts, feature requests, and contributions!

Thank you!

r/AI_Agents Mediocre_Rest_9136

I didn't want the cloud tracking my screen to use AI. So I built an open-source macOS assistant that tracks context locally and connects to Ollama 🧠

Hey Reddit!

Like many of you, I love the idea of an AI that knows what I'm working on and can answer questions about my specific context. But I absolutely hate the idea of uploading my screen activity, clipboard data, and private context to a cloud API.

So, I built Aura Context to solve this. It’s an open-source, privacy-first desktop assistant for macOS.

How it works:

  1. It runs quietly in the background, keeping track of your active window titles and clipboard history.
  2. It pushes all this context into a local SQLite database (everything stays on your machine).
  3. The chat UI hooks directly into your local offline Ollama models (I've been using Llama 3).
  4. You can ask it questions about what you've been doing ("What was the Github link I was looking at an hour ago?" or "Summarize the research I did on React compilers this morning").

It also categorizes your activity so you can see a beautiful, dark-mode productivity dashboard with glassmorphism UI.

Tech Stack:

  • Electron
  • React + TypeScript + Vite
  • better-sqlite3
  • Ollama

If you care about local AI and privacy, I’d love for you to give it a spin or check out the code!

Any feedback on the UI or architecture is incredibly welcome. If you find it useful, a ⭐️ on GitHub would mean the world to an indie dev!

r/LocalLLM de_3lue

Gemma4 - Someone at Google just merged a PR titled "casually dropping the most capable open weights on the planet"

So I was browsing the HuggingFace Transformers repo and a PR just merged today that adds full support for a model called Gemma 4. The PR title is literally "casually dropping the most capable open weights on the planet." The commit has 14 co-authors including Jeff Dean. The weights aren't out yet — the docs still have {release_date} as a placeholder — but the code is all there and it's very readable. Here's what's coming.

Four sizes, including a MoE

  • ~2B and ~4B dense, explicitly designed for on-device use
  • 26B sparse MoE with only 4B active parameters at inference time
  • 31B dense

The 26B/4B MoE is particularly interesting because you get large-model quality at small-model inference cost.

It's trimodal — text, vision, AND audio natively

This is new for Gemma. There's a full audio encoder baked in alongside the vision tower. Not a bolted-on afterthought either — it's a proper conformer architecture (the same family used in production speech systems). The processor handles all four modalities: text, images, video, and audio.

The vision system doesn't squash your images

Most VLMs resize everything to a fixed square. Gemma 4 preserves aspect ratio and instead fits the image into a configurable soft token budget (default 280 tokens, up to 1120 for high detail). No ImageNet normalization — the model handles its own scaling internally.

More interesting: they use a 2D spatial RoPE for vision. Patch positions are encoded as (x, y) coordinates, with half the attention head dimensions rotating for x and the other half for y. The model understands spatial relationships at the architectural level, not just from training.

128K context for small models, 256K for large

The text architecture alternates between sliding window attention (512-1024 token window) and full attention in a 5:1 ratio. The two attention types use completely different RoPE configs — short theta for local, long theta for global. Clean hybrid design.

The small models have some clever efficiency tricks

The 2B and 4B share key-value projections across the last several decoder layers — one layer computes KV, the rest reuse it. There's also a secondary per-layer embedding stream where a small 256-dim signal gets injected at every decoder layer, which I haven't seen in other public models.

The MoE runs experts alongside the MLP, not instead of it

In the 26B variant each layer has both a regular MLP and a sparse MoE block (128 experts, top-8 routing), and their outputs are summed. Unusual design choice — curious whether that helps with stability or quality at scale.


No paper link yet (literally says INSET_PAPER_LINK in the docs), no weights, no release date. But the code is fully merged and production-quality. Feels like days away, not weeks.

What size are you planning to run first?


The PR: https://github.com/huggingface/transformers/pull/45192


EDIT: RELEASE: https://huggingface.co/collections/google/gemma-4

r/AI_Agents Fine-Market9841

Are agentic workflows taking over?

I rly don’t understand the hype, why use “agentic workflows” over n8n or make, and I say this as someone who prefers to build in Python, but the distinction is I want to learn to build reliable, robust code.

Yes, antigravity, Claude code, codex are impressive, but the thing is, you could just add n8n mcp, or make a skill and use the same ai ides to produce json workflows (for n8n and make), the big difference is you actually understand the architecture.

Now aren’t n8n or make mostly prototyping? True, but what matters more (if you’re a beginner) than what tool you use, is learning how to build production systems. So I still think beginners should still use visual builder tools.

That being as you get more complex problems you might wanna switch to Python, I would say keep a list of potential contacts (like on upwork or fiver or something), in case something goes wrong, maybe have your first Python build with a developer.

r/AI_Agents TheGreatBonnie

Implementing Automatic LLM Provider Fallback In AI Agents Using an LLM Gateway (OpenAI, Anthropic, Gemini & Bifrost)

Shipping AI agents that depend on a single LLM provider to production is a risk you cannot afford.

Every major LLM provider e.g, OpenAI, & Anthropic, has experienced outages or rate-limiting incidents in the last 12 months.

For that reason, I wrote a guide on how to implement automatic LLM provider fallback in your app using an LLM gateway.

Check out the article link below 👇

r/LocalLLaMA jopereira

QWEN3.5 27B vs QWEN3.5 122B A10B

For those who already tested these two models in a practical sense, any reason to run 27B instead of 122B? What type of work/play do you usually do?

Reason for questioning: I stayed away from big models (for no reason other than "they are big, they must be slow") but I can run both models, 27B@8t/s and 122B@20t/s (both 80K ctx) and I mostly do ESP32 personal projects (VS Code + Platformio + Kilo Code/Cline/Roo Code)

r/mildlyinteresting ShutterBun

I met a solid white peacock today

r/funny That_Lil_Virus

Mickey mouse couldn't believe his eyes

r/LocalLLM Sea_Manufacturer6590

LM studio Qwen and mcp

Wow LM studio Qwen3.5 plus mcp server connected to YT, reddit, x, Alibaba cloud, and over 25 plus connections if you have any questions ama.

this is truly amazing what it's doing with the local model I generated a voice clip from a prompt and had it call me and play the recording via twilo the recording was all of my stats on my social media.

r/aivideo Sensitive-Industry74

"They Took My Land" – AI-Generated Cinematic Music Video (Grok)

r/LocalLLaMA dylantestaccount

Surprised by how capable Qwen3.5 9B is in agentic flows (CodeMode)

I've been working on my own chat application for a while now to experiment with LLMs, and get some experience with SSE. Also, it's fun to see if I can mirror functionalities being offered in "the big boy tools" like Claude Code, Copilot, ...

A while ago, CloudFlare released a blog post about CodeMode: a new and supposedly better way of letting LLMs call tools (they specifically use it for MCPs, my app provides these tools as built-in but it's basically the same thing at the end of the day).

When I implemented this, I noticed major improvements in:

  • tool call performance
  • context length usage
  • overall LLM agentic capabilities

However, this seemingly only applied to Claude. Most models really don't like this way of tool calling, even though it allows them much more freedom. They haven't been trained on it, and as such aren't very good at it.

Gemini for example never worked, it always output broken tool calls (wrapping in IIFE, not wrapping properly, ...). GPT-5.x most of the time refuses to even output an execute_js block (which is what triggers the tool call logic in the application).

I then tried some open source models like Step Flash 3.5 and GLM which didn't fare much better. MiniMax 2.5 was probably the best.

All models mentioned above were tested through OpenRouter.

I then decided I'd like to see how locally run models would perform - specifically, the ones that my MacBook M1 Pro could reasonably run. Qwen3.5 9B seemed like the perfect fit and is the first one I tried. It also turned out to be the last one as it works so well for me.

Qwen3.5 9B calls the tools perfectly. It doesn't make mistakes often, and when it does is smart enough to self-correct in the next tool call. This is the only model I've tried outside of Claude Sonnet 4.6 that calls the tools this way this effortlessly.

Just wanted to make this post to share my amazement, never have I experienced such a small model being so capable. Even better - I can run it completely locally and it's not horribly slow!

r/ClaudeAI LookTrue3697

I built an open-source memory layer for AI coding agents — it cuts token usage by 60-80% by giving Claude persistent, evidence-backed codebase awareness

AtlasMemory - Every claim grounded in code.

Everyone's been talking about skyrocketing token consumption lately. I've been feeling the same pain watching Claude re-read dozens of files every session, re-discover the same architecture, burn through context just to get back to where we were yesterday.

So I spent the last few months building AtlasMemory a local-first neural memory system that gives AI agents persistent, proof-backed understanding of your entire codebase. Think of it as a semantic knowledge graph that sits between your code and your AI agent, serving precisely the right context at the right time nothing more, nothing less.

The Problem (Why This Exists)

Every time Claude starts a new session on your codebase:

  1. Zero memory it doesn't know your architecture, conventions, or what changed yesterday
  2. Context explosion it reads 30-50 files just to understand one feature flow, sometimes even more on large codebases
  3. Massive token waste on a typical 500-file project, Claude can burn 50,000-100,000+ tokens just to rebuild context that should already be known. On a monorepo? That number can hit 200K+ per session
  4. Hallucination risk without evidence anchoring, claims about your code are just guesses
  5. Drift blindness no way to know if its understanding is stale after you push changes

This gets exponentially worse as your codebase grows. A 100-file project? Manageable. A 28,000-file monorepo? Your entire context window is gone before Claude even starts working on your actual task.

What AtlasMemory Actually Does

AtlasMemory indexes your repository using Tree-sitter AST parsing (the same parser GitHub uses for syntax highlighting), builds a SQLite knowledge graph with full-text search, and serves token-budgeted context packs through the Model Context Protocol (MCP).

The Architecture (Simplified)

Your Codebase ↓ [Tree-sitter AST Parser] — 11 languages supported ↓ Symbols + Anchors + Import Graph + Cross-References ↓ [SQLite + FTS5 Knowledge Graph] — local, fast ↓ [Evidence-Backed File Cards] — every claim links to line ranges + SHA-256 hashes ↓ [Token-Budgeted Context Engine] — you set the limit, it prioritizes what matters ↓ [MCP Protocol] → Claude / Cursor / Copilot / Windsurf / Codex 

What Makes It Different

Evidence Anchoring — This is the core innovation. Every claim AtlasMemory makes about your code is backed by an "anchor" a specific line range with a SHA-256 snippet hash. If the code changes and the hash doesn't match, the claim is automatically flagged as stale. No more hallucinated function signatures or phantom API endpoints.

Proof System — You can ask AtlasMemory to prove any claim:

prove("handleLogin validates JWT tokens before checking permissions") → PROVEN (3 evidence anchors, confidence: 0.94) → src/auth/login.ts:45-62 [hash: a7f3c...] → src/middleware/jwt.ts:12-28 [hash: 9e2b1...] → tests/auth.test.ts:89-104 [hash: 3d8f0...] 

Drift Detection — Context contracts track the state of your repo. If files change after context was built, AtlasMemory warns the agent before it acts on stale information.

Impact Analysis — Before touching shared code, ask "who depends on this?" and get a full dependency graph with risk assessment:

analyze_impact("Store") → MEDIUM RISK: 4 files, 42 symbols, 12 flows affected → Direct: cli.ts (17 refs), mcp-server.ts (17 refs) → No tests found — consider adding before changes 

Real Numbers (With Methodology)

I want to be transparent about these numbers because inflated claims help nobody. Here's how I measured:

How "without" works in practice: When Claude starts a fresh session on an unfamiliar codebase, it needs to discover the architecture before it can do anything useful. This means: glob to find file structure (~1-2K tokens), Read on 15-40 files to understand the codebase (~15,000-40,000 tokens since average source file is ~1,000 tokens), multiple grep searches (~3-5K tokens), plus Claude's own reasoning overhead (~5-10K tokens). On a 500-file project, this exploration phase typically costs 25,000-50,000 tokens before Claude writes a single line of code.

How "with" works: Claude calls handshake (gets full project brief in ~2K tokens), then search_repo for the specific area it needs (~1K tokens), optionally build_context for deeper understanding (~3-5K tokens). Total discovery cost: ~3,000-8,000 tokens. Claude still reads the specific files it needs to edit — but it already knows which files to read instead of exploring blindly. That's where the real savings come from.

Phase Without AtlasMemory With AtlasMemory Savings Discovery (understand architecture) 25,000-50,000 tokens ~2,000-3,000 tokens (handshake) ~90-95% Search (find relevant code) 5,000-15,000 tokens (grep/glob/read) ~1,000-2,000 tokens (search_repo) ~80-90% Deep context (understand specific area) 10,000-30,000 tokens (read 10-20 files) ~3,000-5,000 tokens (build_context) ~70-85% Implementation (read files to edit) 5,000-15,000 tokens 5,000-15,000 tokens (same — you still read what you edit) 0% Total typical session 45,000-110,000 tokens ~11,000-25,000 tokens ~60-80%

Important note: AtlasMemory doesn't eliminate file reading entirely you still need to read the files you're about to modify. What it eliminates is the blind exploration phase where Claude reads dozens of files just to figure out where things are. That exploration phase is where most of the waste happens, especially on larger codebases.

On monorepos (5K+ files): The savings are even more dramatic because without AtlasMemory, Claude has to read 40-80+ files just to map the architecture. With AtlasMemory, the handshake gives a complete architecture overview, risk map, and recent changes in ~3,000-5,000 tokens. I've seen sessions on monorepos go from 100K+ exploration tokens to under 10K.

Stress-tested on real open-source repos:

  • Express.js (580 files) → indexed in 3.2s, search <15ms
  • Fastify (740 files) → indexed in 4.1s
  • Next.js monorepo (28,000 files) → handles enterprise scale without crashes
  • Coolify (1,400+ PHP/JS files) → multi-language indexing across PHP, JS, TypeScript

What's Included (Full Ecosystem)

This isn't just a CLI tool it's a complete ecosystem available everywhere:

Component Description Link MCP Server 28 tools, works with any MCP-compatible AI agent npx -y atlasmemory CLI Full command-line interface (atlas index, search, enrich, generate, doctor) npm i -g atlasmemory VS Code Extension Dashboard, sidebar, status bar, AI readiness score VS Code Marketplace Open VSX Same extension for VS Code forks (VSCodium, Gitpod, etc.) Open VSX Registry npm Package One-command install, ~400KB bundle npmjs.com/package/atlasmemory 5 AI Config Formats Auto-generates CLAUDE.md, .cursorrules, copilot-instructions.md, .windsurfrules, AGENTS.md atlas generate 11 Languages TypeScript, JavaScript, Python, Go, Rust, Java, C#, C, C++, Ruby, PHP Tree-sitter based AI Enrichment Semantic tag generation using Claude CLI (free) or Anthropic API atlas enrich

VS Code Extension

AtlasMemory isn't just a terminal tool there's a full VS Code extension with a visual dashboard:

VS Code Extension

Features:

  • Atlas Explorer sidebar — browse your indexed codebase, see file cards, symbol maps
  • AI Readiness Score — see how well your project is prepared for AI agents (0-100)
  • Status Bar — always-visible index status and quick actions
  • One-click indexing — index or re-index from the sidebar
  • Search integration — semantic search directly from VS Code

Install:

Setup (Literally 30 Seconds)

For Claude Desktop / Claude Code:

{ "mcpServers": { "atlasmemory": { "command": "npx", "args": ["-y", "atlasmemory"] } } } 

That's it. First handshake call auto-indexes your repo. Every session after that gets instant, proof-backed context.

For VS Code: Search "AtlasMemory" in the extension marketplace → Install → Done. Dashboard shows AI readiness score, file explorer, and search — all from the sidebar.

For CLI power users:

npm install -g atlasmemory cd your-project atlas index # Index once (~3s for 500 files) atlas search "auth" # Semantic search (<15ms) atlas enrich # AI-enhanced descriptions (free with Claude CLI) atlas generate # Auto-generate CLAUDE.md + 4 other AI configs atlas doctor # Health check your memory database 

MCP Tools Available (28 Total)

The key ones AI agents use:

Tool What It Does handshake Session init — project brief + memory + protocol in one call (~2K tokens) search_repo Semantic search with co-change intelligence and fragility warnings build_context Token-budgeted context packs — you set the limit, it prioritizes prove Verify claims against actual code evidence (line ranges + SHA-256) analyze_impact "Who depends on this?" — full dependency graph + risk assessment log_decision Persistent memory of what was changed, why, and which files smart_diff Enriched diffs with semantic understanding of what changed enrich_files AI-enhanced semantic tags for dramatically better search quality generate_claude_md Auto-generate AI instructions for 5 different tools ai_readiness Score your project's AI-readiness (0-100)

How It Actually Feels

Before AtlasMemory:

"Let me read your project structure... reads 40 files, burns 60K tokens... okay I think the auth is in src/auth but I'm not sure about the middleware chain... let me read a few more files..."

After AtlasMemory:

"Based on the project brief: auth flow goes through src/middleware/jwt.ts (line 12-28) → src/auth/login.ts (line 45-62). 3 evidence anchors confirm JWT validation happens before permission checks. Impact analysis: 4 dependent files, no breaking changes expected. Total context used: 2,100 tokens."

Pro Tip: Claude Code Hooks (Maximum Efficiency)

After using AtlasMemory on all my own projects for months, here's the biggest lesson I learned: AI agents sometimes forget to call AtlasMemory tools. They get excited about your question and start reading files directly instead of checking memory first and there go your tokens.

The fix? Claude Code hooks. You can make AtlasMemory usage mandatory at the start of every session:

Add this to your .claude/settings.json:

{ "hooks": { "PreToolUse": [ { "matcher": ".*", "hook": "echo 'REMINDER: Did you call handshake first? Use search_repo before reading files directly. AtlasMemory has indexed this codebase — use it.'" } ] } } 

Or simply add a rule to your CLAUDE.md (AtlasMemory auto-generates this with atlas generate):

## MANDATORY: AtlasMemory Protocol 1. Call `handshake` at the START of every session 2. Use `search_repo` BEFORE reading any files 3. Use `build_context` for complex tasks 4. Call `log_decision` AFTER making changes 

This single change made the biggest difference in my token usage Claude stops wasting tokens re-reading files and starts leveraging the knowledge graph from the first message.

Philosophy

  • 100% Local — your code never leaves your machine. No cloud, no API keys for core features
  • Evidence > Hallucination — every claim backed by line ranges and cryptographic hashes
  • Deterministic Core — the engine is pure AST extraction, no LLM required for basic operation
  • Token-Aware — greedy priority budgeting fits any context window
  • Drift-Resistant — stale context is automatically detected and flagged

Open Source (GPL-3.0)

GitHub: github.com/Bpolat0/atlasmemory npm: npmjs.com/package/atlasmemory VS Code: Marketplace | Open VSX

I've documented everything from A to Z in the README — architecture, setup guides for 5 different AI tools, enrichment workflows, FAQ, comparison diagrams, the works. If something's unclear, open an issue and I'll improve it.

A few honest words: I'm a solo developer and I use AtlasMemory on every single project I work on it's genuinely part of my daily workflow, not just something I built and forgot about. That said, there might be bugs I haven't caught yet. If you run into anything, please report it on GitHub — every issue helps me make this better, and I push updates regularly (we're on v1.0.14 already with fixes from real-world testing across multiple AI agents).

I really hope you find it as useful as I do. Stars and feedback mean the world to me this is my first major open source project, and your support is what keeps it going.

Built with TypeScript, Tree-sitter, SQLite, and a mass amount of mass.

r/LocalLLaMA MR_-_501

Qwen 3.6 will have oss models

r/meme Fickle-Butterfly-338

I'd love to know...

r/LocalLLaMA Abyss_c0re

[Early Access] GitHub - Abyss-c0re/NeuralCore: NeuralCore is an experimental adaptive agentic framework.

Local first (LLama.cpp) highly adjustable multi agent framework.

Actions, Sequences, Workflows, Loops ,Conditions, Sub agents, Dynamic Tool loading + Build in Hybrid-RAG.

Easy tools (actions) and workflow creation via @ decorators. Workflows and pretty much anything can be adjust inside yaml config (later planning to build a web gui, similar to LangGraph).

Everything can be an action/tool. You can expose client's methods such as describe image or chat to other agents. You can allow agents to access their own methods (search context, deploy sub agents).

The framework is being developed with Qwen3.5-9B model, so results may vary. ​Very early stage, no documentation yet but I've built a functional demo app top of this framework, for the developers to investigate first and hopefully provide a feedback.

r/LocalLLaMA Another__one

How good mini-pc's like this are for local AI inference and LORA fine-tuning via pytorch? Could I expect reasonable speed with something like that or or is it going to be painfully slow without a discrete GPU chip on the board?

r/LocalLLM crosswalk_elite

Am I stupid to think I can deploy an LLM as good as Claude on my laptop's 4060?

I need it mostly for coding and pulling out new research papers and ideas for my speech-llm project, alongside some course assignments and projects. I love what claude extended thinking can achieve within one prompt and it stays pretty professional since I have the memory off. I value privacy so had done away with my LOQ's copilot. But the new claude limits are creating a real hindrance, and I love the idea of having an on demand assistant I have to share with no one. I have no clue if anything can fit on 8gb and match the quality.

r/ClaudeCode JacketDangerous9555

Claude Code now lets you pick effort levels too

Claude Code now lets you pick effort levels too . Might be a sneaky way to downgrade service, saving the best for Effort Max. Others auto-set to medium or low or whatever, a mind game of cutting quality without cutting quantity 😅

r/funny Sad-Confusion-6633

Is this THREESOME?

r/LocalLLM crosswalk_elite

Am I stupid to think I can deploy an LLM as good as Claude on my laptop's 4060?

I need it mostly for coding and pulling out new research papers and ideas for my speech-llm project, alongside some course assignments and projects. I love what claude extended thinking can achieve within one prompt and it stays pretty professional since I have the memory off. I value privacy so had done away with my LOQ's copilot. But the new claude limits are creating a real hindrance, and I love the idea of having an on demand assistant I have to share with no one. I have no clue if anything can fit on 8gb and match the quality.

Verdict: a resounding yes. I learnt a lot here, thanks!

r/SideProject Typical-Particular-6

4 days in: 66 visitors, 45 scans, 2 checkout attempts, 0 sales. Building in public. Building PresenceForge. Scans any local business market and generates a full marketing package.

4-day numbers (100% organic Reddit, zero ad spend):

  • 66 unique visitors
  • 45 free scans completed (68% scan rate)
  • 39 seconds average engagement (was 8 seconds before rewriting the copy)
  • 2 people clicked $197 buy button
  • 1 person waited through entire 5-min report generation and bounced at Stripe
  • 0 revenue

What I learned:

  1. "30-page report" doesn't sell. "See where you're losing clients" does. That single copy change tripled engagement.
  2. The free scan needs to create a gap, not satisfy curiosity. Added personalized conclusions like "you're leaving $2,400/month on the table" and buy clicks started.
  3. 68% scan rate means people want the data. 0% payment rate means they don't trust the site yet. Added real business names in a social proof section to fix that.

What would you change?

feel free to try the scan. it takes 30 seconds --> Free Scan

r/AI_Agents Individual_Hand213

I made an open source alternative to Higgsfield AI

Open-Higgsfield-AI is an open source platform that lets you access and run cutting-edge AI models in one place. You can clone it, self-host it, and have full control over everything.

It’s a lot like Higgsfield, except it’s fully open, BYOK-friendly, and not locked behind subscriptions or dashboards.

Seedance 2.0 is already integrated, so you can generate and edit videos with one of the most talked-about models right now — directly from a single interface.

Instead of jumping between tools, everything happens in one chat:

generation, editing, iteration, publishing.

While commercial platforms gatekeep access, open source is moving faster — giving you early access, more flexibility, and zero lock-in.

This is what the future of creative AI tooling looks like.

r/aivideo ilvito73

The King of Subbuteo

r/LocalLLM TylerDurdenFan

Bonsai (PrismML's 1 bit version of Qwen3 8B 4B 1.7B) was not an aprils fools joke

I read the article yesterday:

https://prismml.com/news/bonsai-8b

And watched the only 3 videos that had surfaced about these bonsai models. Seemed legit but still maybe an aprils fools joke.

So today I woke up wanting to try them. I downloaded their 8B model, their llama.cpp fork, and tested it, and as far as I can see it's real:

On my humble 4060, 107 t/s generation and >1114 t/s prompt processing, with a model that's evidently tiny. For comparison, on qwen 3.5 4B Q4 I had gotten 56 t/s using the same prompts.

Most importantly, the RAM used us much much lower, so I can use an 8B model in my humble 8GB VRAM, or the smaller models with longer context.

Quality: I have a use case of summarizing text, and upon first inspection it worked well. I dont try coding nor tool using, but for summarization it is golden.

The only bad part is that while it worked well on my windows PC with CUDA, when I tried it on a GPU-less mini PC (to see potential edge performance), although the llama.cpp fork compiles, it does not work, it loads the model, and seems to start processing the prompt and seems to hang. I asked Claude to check their code and it tells me they have no CPU implementation, so it might be dequantizing to FP32 and attempting regular inference (which would be dead slow on CPU).

I think there should be potential for these 1 bit models not only to reduce bandwidth and memory requirements, but also compute requirements: the matrix multiplication part, on 1 bit matrixes, should be something like XOR operations, much faster than FPanything. As I understand, so even if scaling to FP16 is required after the XOR, still a huge amount of compute was saved, which should help CPU-only inference, and edge inference in general.

There's hope for us VRAM starved plebes after all !! (and hopefully this might help deflate ramageddon, and the AI datacenter bubble in general)

r/ProgrammerHumor Vivid_Search674

thisWasEnoughBefore

r/mildlyinteresting kdt912

Not actually enlarged at all

r/ChatGPT Jason_Grace15

Correct me if I'm wrong, but I see no difference here? Its just asking me if I prefer option a or option a again.

r/homeassistant Renrut23

Going from HA on bare metal to unraid.

currently have HA installed on an old levovo thinkcentre on bare metal. Thinking of migrating it to a server running unraid. I primarily have everything connected through a SLZB-06 thats connected via ethernet.

can I just bring a backup over and restore it to get HA up and running again or do I need to reconfigure it all from scratch?

r/SideProject Dimention_less

If your blog uses Google Analytics, cookies, or affiliate links - you legally need these pages. Here's how to create them without getting scammed by "free" generators.

I see this question constantly: Do I really need a privacy policy for my blog?

Short answer: Yes. Here's why:

  • Google Analytics = You're tracking visitors. GDPR and CCPA require a privacy policy.
  • Affiliate links = The FTC legally requires a clear disclosure/disclaimer.
  • Email signup forms = You're collecting personal data. You need a privacy policy.
  • Ad networks (AdSense, Mediavine, etc.) = Cookies + data collection = policy required.
  • You sell anything (courses, merch, services) = You need Terms & Conditions.

The problem? Most "free" generators online are scams:

  • They ask for your email, then spam you
  • They embed hidden links in the generated text
  • They charge you after you've done all the work
  • The templates are outdated and don't cover current laws
  • You get a generic wall of text with stuff that doesn't apply to your site

I got fed up and built three generators that are actually free, no catches:

You just toggle what applies to your blog and skip what doesn't. No irrelevant clauses. No hidden links in the output. No paywall. No data collection.

Hope this helps someone avoid the same frustration I went through.

r/ClaudeCode eazyigz123

ThumbGate: pre-action gates that stop Claude Code from repeating your mistakes

Built an open-source tool that turns thumbs-up/down feedback into enforcement gates for Claude Code.

The problem: Claude Code repeats mistakes across sessions. You tell it not to force-push, it does it again next time. Prompt rules get lost in long contexts.

The solution: ThumbGate intercepts every tool call via PreToolUse hooks. When you thumbs-down a mistake, it auto-generates a prevention rule. Next time Claude tries the same action, the gate blocks it before execution.

Key features: - PreToolUse hooks intercept tool calls before they run - SQLite+FTS5 lesson database persists across sessions - Thompson Sampling adjusts gate severity over time - Works with Claude Code, Cursor, Codex, Gemini, Amp, any MCP agent

One command setup: npx mcp-memory-gateway init

We recently mapped our architecture to Tsinghua's NLAH (Natural-Language Agent Harness) framework, which formalizes this exact pattern. Deep dive: https://rlhf-feedback-loop-production.up.railway.app/learn/agent-harness-pattern

Open source: https://github.com/IgorGanapolsky/ThumbGate

Happy to answer questions about the implementation.

r/SideProject kaustabh93

I built a tool that optimises your dating profiles

Been single for a while and started noticing my matches were pretty bad. Checked my own profile and realized I couldn’t even tell if it was good or not.

So I built DateBait.

You upload your profile and it optimises them to be sharper, more specific, and actually interesting. There’s also a mode to compare two bios side by side, and one that generates openers based on a match’s profile.

It’s free and you don’t need to sign up to try it.

Would really value honest feedback. Especially if something feels off, or not like you at all.

www.datebait.app

r/whatisit the_wattweiler

weird ice cubes

I made green tea ice cubes for my iced tea, I havent made them before. I sweeten my green tea with monkfruit. why do some of them look milky? im p sure its something with the monkfruit, but just in case it isn't, is it safe?

r/AI_Agents LoFiTae

Is there something I can do about my prompts? [Long read, I’m sorry]

Hello everyone, this will be a bit of a long read, i have a lot of context to provide so i can paint the full picture of what I’m asking, but i’ll be as concise as possible. i want to start this off by saying that I’m not an AI coder or engineer, or technician, whatever you call yourselves, point is I’m don’t use AI for work or coding or pretty much anything I’ve seen in the couple of subreddits I’ve been scrolling through so far today. Idk anything about LLMs or any of the other technical terms and jargon that i seen get thrown around a lot, but i feel like i could get insight from asking you all about this.

So i use DeepSeek primarily, and i use all the other apps (ChatGPT, Gemini, Grok, CoPilot, Claude, Perplexity) for prompt enhancement, and just to see what other results i could get for my prompts.

Okay so pretty much the rest here is the extensive context part until i get to my question. So i have this Marvel OC superhero i created. It’s all just 3 documents (i have all 3 saved as both a .pdf and a .txt file). A Profile Doc (about 56 KB-gives names, powers, weaknesses, teams and more), A Comics Doc (about 130 KB-details his 21 comics that I’ve written for him with info like their plots as well as main cover and variant cover concepts. 18 issue series, and 3 separate “one-shot” comics), and a Timeline Document (about 20 KB-Timline starting from the time his powers awakens, establishes the release year of his comics and what other comic runs he’s in [like Avengers, X-Men, other character solo series he appears in], and it maps out information like when his powers develop, when he meets this person, join this team, etc.). Everything in all 3 docs are perfect laid out. Literally everything is organized and numbered or bulleted in some way, so it’s all easy to read. It’s not like these are big run on sentences just slapped together. So i use these 3 documents for 2 prompts. Well, i say 2 but…let me explain. There are 2, but they’re more like, the foundation to a series of prompts.

So the first prompt, the whole reason i even made this hero in the first place mind you, is that i upload the 3 docs, and i ask “How would the events of Avengers Vol. 5 #1-3 or Uncanny X-Men #450 play out with this person in the story?” For a little further clarity, the timeline lists issues, some individually and some grouped together, so I’m not literally asking “_ comic or _ comic”, anyways that starting question is the main question, the overarching task if you will. The prompt breaks down into 3 sections. The first section is an intro basically. It’s a 15-30 sentence long breakdown of my hero at the start of the story, “as of the opening page of x” as i put it. It goes over his age, powers, teams, relationships, stage of development, and a couple other things. The point of doing this is so the AI basically states the corrects facts to itself initially, and not mess things up during the second section. For Section 2, i send the AI’s a summary that I’ve written of the comics. It’s to repeat that verbatim, then give me the integration. Section 3 is kind of a recap. It’s just a breakdown of the differences between the 616 (Main Marvel continuity for those who don’t know) story and the integration. It also goes over how the events of the story affects his relationships. Now for the “foundations” part. So, the way the hero’s story is set up, his first 18 issues happen, and after those is when he joins other teams and is in other people comics. So basically, the first of these prompts starts with the first X-Men issue he joins in 2003, then i have a list of these that go though the timeline. It’s the same prompt, just different comic names and plot details, so I’m feeding the AIs these prompts back to back. Now the problem I’m having is really only in Section 1. It’ll get things wrong like his age, what powers he has at different points, what teams is he on. Stuff like that, when it all it has to do is read the timeline doc up the given comic, because everything needed for Section 1 is provided in that one document.

Now the second prompt is the bigger one. So i still use the 3 docs, but here’s a differentiator. For this prompt, i use a different Comics Doc. It has all the same info, but also adds a lot more. So i created this fictional backstory about how and why Marvel created the character and a whole bunch of release logistics because i have it set up to where Issue #1 releases as a surprise release. And to be consistent (idek if this info is important or not), this version of the Comics Doc comes out to about 163 KB vs the originals 130. So im asking the AIs “What would it be like if on Saturday, June 1st, 2001 [Comic Name Here] Vol. 1 #1 was released as a real 616 comic?” And it goes through a whopping 6 sections. Section 1 is a reception of the issue and seasonal and cultural context breakdown, Section 2 goes over the comic plot page by page and give real time fan reactions as they’re reading it for the first time. Section 3 goes over sales numbers, Section 4 goes over Mavrel’s post release actions, their internal and creative adjustments, and their mood following the release. Section 5 goes over fan discourse basically. Section 6 is basically the DC version of Section 4, but in addition to what was listed it also goes over how they’re generally sizing up and assessing the release. My problem here is essentially the same thing. Messing up information. Now here it’s a bit more intricate. Both prompts have directives as far as sentence count, making sure to answer the question completely, and stuff like that. But this prompt, each section is 2-5 questions. On top of that, these prompts have way, way more additional directives because it the release is a surprise release. And there more factors that play in. Pricing, the fact of his suit and logo not being revealed until issue #18, the fact that the 18 issues are completed beforehand, and few more stuff. Like, this comic and the series as whole is set to be released a very particular type of way and the AIs don’t account for that properly, so all these like Meta-level directives and things like that. But it’ll still get information wrong, gives “the audience” insight and knowledge about the comics they shouldn’t have and things like that.

So basically i want to know what can i do to fix these problems, if i can. Like, are my documents too big? Are my prompts (specifically the second one) asking too much? For the second, I can’t break the prompts down and send them broken up because that messes up the flow as when I’m going through all the way to 18, asking these same questions, they build on each other. These questions ask specifically how decisions from previous issues panned out, how have past releases affected this factor, that factor, so yeah breaking up the same prompt and sending it in multiple messages messes all that up. It’s pretty much the same concept for the first but it’s not as intricate and interconnected to each other. That aside, i don’t think breaking down 1 message of 3 sections into 3 messages would work well with the flow I’m building there either way.

So yeah, any tips would be GREATLY appreciated. I have tried the “ask me questions before you start” hack, that smoothes things a bit. Doing the “you’re a….” Doesn’t really help too much, and pretty much everything else I’ve seen i can’t really apply here. So i apologize for the long read, and i also apologize if this post shouldn’t be here and doesn’t fit for some reason. I just want some help

r/whatisit freefromrest

Fire in the sky

A friend shoot this video today in istanbul. Whats going on?

r/whatisit Known-Ad7621

Unknown animal call

Audio taken in Southern Ontario (near Peterborough), near a small, shallow (frozen) bay of a lake (not certain that the calls were coming from closer to the bay or from the adjacent forest). Merlin Bird ID doesn’t register the sound, and I know several skilled ecologists who also can’t identify it. Any thoughts??

r/LocalLLM daDon3oof

Copilot like model?

New to LLM, tried using Qwen3.5-9b on vs code with Continue to give it access to my project so it can read it and make modifications just like "Github Copilot".

Qwen2.5-14b refuse to read project files, Qwen3.5-9b does read project files but keep hanging after thinking, basically i am lost here.

Copilot is easy to instruct and works great, i need something to run locally, rtx 3080TI 32GB ddr5 ram.

r/ChatGPT BenevolentHoax

Equivalent of Cowork

This may be a dumb question, but I recently subscribed to Claude Pro (alongside my ChatGPT Pro) specifically for the Cowork feature. Then I found out Cowork is only available in the desktop app, and my work has not approved Claude for use so I can’t use the desktop app. All this to say, can you hack ChatGPT Pro to do something similar to Claude Cowork?

r/ClaudeAI BuildEdgeHQ

I've been using Claude Code as a non-developer for a few weeks. Here's what actually worked and what didn't

Background: 20+ years in product and customer success leadership. I understand code conceptually but I haven't written any in two decades. I've been using Claude for chat and writing, and Claude Code has become my go-to for building things.

I've shipped a few projects with it now and wanted to share what I've learned about using Claude Code when you don't come from a technical background.

Specs are everything. The difference between Claude Code building something useful vs something you delete comes down to how specific your instructions are. "Build me a client tracker" gets you garbage. "Build a Clients database with fields for Company Name (title), Contact Person (text), Email (email), Status (select: Lead / Active / Completed / Lost), Monthly Value (number, currency USD)" gets you something you'd actually use. This is product thinking, and it's the skill that transfers directly from PM work.

Plan Mode before Build Mode. Always. Let Claude Code read your spec and present a plan first. Review the plan. Fix issues before a single file gets created. I caught major structural problems in the planning phase that would have taken hours to untangle later.

Skill files are underrated. I've been writing CLAUDE md files that give Claude Code the full context of a project - goals, constraints, phases, what to do and what not to do. The difference between prompting from scratch every session vs having a well-written skill file is massive. If you're building anything that involves a multi-step process, try it.

Describe what you see, not what you think the code problem is. When something breaks, I describe what I did, what I expected, and what happened instead. Claude Code figures out the fix. This is the same approach I used managing engineering teams for years. The PM describes the gap, the engineer fixes the implementation.

What doesn't work: Vague instructions, skipping the planning step, and trying to build everything in one prompt. Claude Code is incredibly capable but also incredibly literal. Treat it like a fast but junior engineer who needs clear direction.

Happy to answer questions about using Claude Code as a non-developer.

r/LocalLLaMA TKGaming_11

Gemma 4 1B, 13B, and 27B spotted

[Gemma 4](INSET_PAPER_LINK) is a multimodal model with pretrained and instruction-tuned variants, available in 1B, 13B, and 27B parameters. The architecture is mostly the same as the previous Gemma versions. The key differences are a vision processor that can output images of fixed token budget and a spatial 2D RoPE to encode vision-specific information across height and width axis.

You can find all the original Gemma 4 checkpoints under the [Gemma 4](https://huggingface.co/collections/google/gemma-4-release-67c6c6f89c4f76621268bb6d) release.

r/ClaudeCode biograf_

Claude got me promoted, got me paid, and got me a girlfriend. Now its context window is ruining my life.

I’m posting this partly as a warning and partly because I’m not sure who else would understand this situation.

About eight months ago I started using Claude pretty heavily for development. At first it was just little stuff: refactoring functions, writing tests, explaining weird library behavior. But then I started using it for everything.

And I mean everything.

Architecture decisions. Debugging race conditions. Writing SQL migrations. Generating docs. Planning product features. Even helping me draft emails to my manager.

The productivity jump was insane. I went from being a decent junior developer to suddenly looking like some kind of 10x wizard.

Within four months:

- I fixed a bunch of legacy issues no one wanted to touch

- I shipped two internal tools that saved our team hours every week

- I got promoted to senior developer (which I absolutely did not deserve)

But the real turning point was when Claude helped me build a side project.

It started as a dumb little SaaS idea. Claude basically helped me scaffold the entire stack: backend, database schema, API, frontend components, deployment scripts, even marketing copy. It was like pair programming with someone who had already built ten startups.

The app launched and — unbelievably — it started making money. Not crazy money, but enough that I could point to revenue and feel like I wasn’t just cosplaying as an entrepreneur.

This had unexpected side effects.

A girl I’d been seeing casually suddenly started thinking I was extremely impressive. Apparently “software developer with a profitable app” hits different than “guy who spends all weekend debugging Docker.”

She started introducing me as “the one who built his own tech company.”

I did not correct her.

For about three glorious months my life looked like this:

- great performance reviews

- a growing side-project

- a very pretty girlfriend

- the quiet belief that I had somehow unlocked the secret to productivity

Then the context limits started happening.

If you use Claude for coding you know exactly what I mean.

At first it was small things. I’d paste a chunk of my project and Claude would say something like:

“I may be missing context from earlier in the conversation” and "You've reached your maximum usage."

Then it started forgetting key architectural decisions. It would suggest changes that directly contradicted things it had designed earlier.

Soon every session went like this:

- Explain the project again

- Paste half the codebase

- Clarify the same design constraints

- Watch it confidently refactor something into a broken state

My productivity dropped hard.

Features that used to take one evening now took days because I had to keep re-explaining the system. Debugging sessions turned into archaeological digs through earlier chats.

The side project stalled.

At work my velocity also cratered. Tasks that used to magically resolve themselves now required me to actually understand things again.

My manager asked if everything was okay.

Then the girlfriend situation deteriorated.

She came over one evening while I was in the middle of trying to fix a bug Claude had introduced after forgetting half the system architecture.

For two hours she watched me repeatedly type things like:

“No, remember the Redis cache we discussed earlier.”

“No, the user IDs are UUIDs not ints.”

“No, you literally designed this module yesterday.”

She finally said:

“Are you arguing with your computer?”

I tried to explain the context window problem.

I could see the exact moment she realized I might not be the genius founder she thought I was.

Two weeks later she said she didn’t think we had “long-term alignment.”

Now I’m back where I started:

- struggling through tickets like a normal developer

- a half-finished SaaS app

- no girlfriend

- and a Slack message from my manager asking if I can “revisit the velocity we had earlier this year”

The worst part is that Claude occasionally still has flashes of brilliance, which keeps me hooked.

It’s like working with a senior engineer who has severe short-term memory loss but still occasionally drops absolute genius solutions.

Anyway.

If anyone has figured out how to manage large codebases with these context limits please let me know.

My promotion, my startup, and possibly my next relationship may depend on it.

r/comfyui ThunderI0

Related to AI UGC content.

Where do you get those pose references or stuff that look like perfect instagram posts? I am having a hard time to correctly write those prompts for Z image turbo to understand. Is there a specific way to write those prompts? Quite confusing!

r/AI_Agents FairNefariousness359

Got my first AI agent customer - help me review the architecture

Hey! This week I closed my first real customer for my AI implementation services. The project is building a support agent for their B2B customers. I have experience building agents for my other companies, but I would love to receive some feedback and tips on my plan.

The customer is a physical access control company that also delivers a full software package alongside their hardware. Their support department receives a lot of calls about the same simple questions, for example why a door won't open. Usually the answer is straightforward, the user trying to get access is in a user group that doesn't have access at that specific time. Customers could technically find this themselves, but the interface they use is not very user-friendly and quite technical. Once you know your way around it, it's actually pretty simple. And with the REST API the software offers, you can identify the cause of most problems with just a couple of GET requests.

The plan:

Instead of customers going straight to phone support like they do now, we add a support layer in between. A new interface for their customers with their own login, where they can chat with an AI agent.

It starts with a small FAQ checklist, a few quick questions to rule out the obvious stuff before escalating, like "is the internet connected?" (yes, that's a real common one). If they get past that and still have an issue, they can ask something like "why is this door not opening for this user?" The agent calls the REST API, pulls the relevant data, and pinpoints the exact reason.

On top of that, the access control software has solid documentation, so questions about how to use the system itself, where to find a setting or how to reset something, can also be answered directly by the agent without any API calls needed.

The architecture:

  • Python with the Anthropic SDK directly, no framework, just a clean tool-calling loop
  • Read-only GET tools against the BioStar 2 REST API (device status, access events, user groups, schedules)
  • BioStar 2 docs loaded straight into the system prompt (CAG, no vector database needed)
  • JWT auth with tenant isolation hardcoded at the tool level
  • PostgreSQL for conversation history, tenant config and audit logging
  • Hosted on Railway, EU region, Claude Sonnet via AWS Bedrock EU for GDPR compliance

The agent is strictly read-only by design. It diagnoses, it never acts. Any actual changes go through the support team.

Would love to hear from people who have built similar support agents, especially around keeping tenant isolation bulletproof in a multi-tenant setup, CAG vs RAG tradeoffs for small-to-medium documentation sets.

r/whatisit LumberJackImOK

I found this in the closet in a house I was buying during the pre-close walkthrough. It was gone the next day.

r/ClaudeAI narrateai10

I built an MCP server for NarrateAI and tested it inside Claude, here's what actually happened

I make a lot of screen recordings for client demos. The part I hate is recording voiceover. You're always slightly off, you re-record it three times, and it still sounds like you're reading a script because you are.

So I built NarrateAI to handle that. It watches the video, figures out what's happening on screen, and generates synced voiceover automatically. That part works pretty well.

What I wasn't sure about was the MCP side. I published the server recently and wanted to see if Claude could drive the whole workflow without me babysitting it.

Turns out it can. You give it a video URL, it uploads, kicks off narration, polls for the result, and hands back a download link. The whole thing runs inside the conversation. I didn't touch the NarrateAI UI once.

What I didn't expect: Claude handles the async part better than I thought. Jobs take a few minutes to process, so there's a real wait involved. Claude just... waits, then picks it back up. No weird behavior.

Some different ways to connect depending on your setup:

Claude.aiSettings > Integrations > Add MCP Server, paste the URL + API key header. That's it.

Cursor / VS CodeAdd to your mcp.json via stdio. Needs Python 3.11+.

SmitheryOne-liner install if you prefer that route.

Full setup at narrateai.app/mcp. Free tier includes 5 minutes of processing, no card required.

Curious if anyone else has built media-processing tools on MCP. The async pattern feels a bit awkward compared to instant-return tools — I'm still not sure I've nailed the polling behavior.

r/midjourney iiithewizardiii

Waitaminute...

r/LocalLLaMA Turbulent-Sky5396

Bankai (卍解) — the first post-training adaptation method for true 1-bit LLMs.

I've been experimenting with Bonsai 8B — PrismML's true 1-bit model (every weight is literally 0 or 1, not ternary like BitNet). I realized that since weights are bits, the diff between two model behaviors is just a XOR mask. So I built a tool that searches for sparse XOR patches that modify model behavior.

The basic idea: flip a row of weights, check if the model got better at the target task without breaking anything else, keep or revert. The set of accepted flips is the patch.

What it does on held-out prompts the search never saw:

Without patch: d/dx [x^7 + x] = 0 ✗ With patch: d/dx [x^7 + x] = 7x^6 + 1 ✓ Without patch: Is 113 prime? No, 113 is not prime ✗ With patch: Is 113 prime? Yes, 113 is a prime number ✓ 

93 row flips. 0.007% of weights. ~1 KB. Zero inference overhead — the patched model IS the model, no adapter running per token. Apply in microseconds, revert with the same XOR.

Key findings across 8 experiments:

  • 500K random bit flips barely move perplexity (<1%). The model has massive redundancy in its binary weights.
  • High-scale rows have 3.88x more behavioral impact than random rows — the model's scale factors tell you where to search.
  • Patches trained on 6 probes memorize specific prompts. Patches trained on 60 diverse probes generalize to held-out problems (4 fixed, 0 broken on 30 unseen problems).
  • Patch stacking works mechanically (order-independent, fully reversible) but the improvements partially cancel — joint optimization would beat naive stacking.
  • 50 GSM8K word problems: no degradation (22% → 28%, likely noise but directionally positive).

Why this only works on true 1-bit models:

BitNet b1.58 uses ternary weights {-1, 0, +1} packed as 2 bits. XOR on 2-bit encodings produces invalid states (XOR(01, 10) = 11 has no valid mapping). Bonsai is true binary — each weight is one bit, XOR flips it cleanly from −scale to +scale. As far as I know, this is the first post-training adaptation method for true 1-bit LLMs.

The deployment angle:

LoRA adapters are ~100 MB, add latency per token, and need weight reloading to swap. XOR patches are ~1 KB, apply in microseconds, and add zero inference cost. Imagine a library of domain patches hot-swapped on a phone — a thousand patches adds 1 MB to a 1.15 GB base model.

One person, no ML research background, M3 MacBook Air. Everything is open — toolkit, patches, all 8 experiments reproduce in under 2 hours on any Apple Silicon Mac.

Repo: https://github.com/nikshepsvn/bankai

Paper: https://github.com/nikshepsvn/bankai/blob/master/paper/bankai.pdf

Would love feedback from anyone who wants to poke holes in this.

r/SideProject Ok-Jelly-4359

Building a habit tracking app to get rid of bad habits

Hi everyone

I know there are thousands of habit tracking apps out there. However, I had a different idea for such an app:

Instead of tracking habits you want to reinforce, you track how often you are able to resist a certain habit you want to get rid off.

The idea is the following: You register the habits you want to get rid off in the app (e.g. scrolling, junk food, drinking etc.). Whenever you feel the urge to do one of these habits, instead you open the app and log a "resist moment" when you’re in the middle of an urge and choose not to act on it.

Over time, you’d build up a record of how often you’ve resisted, see patterns and track your progress towards getting rid of unwanted habits.

The core features I’m thinking about:

  • a quick way to log a resistance when an urge hits in three levels: low urge, normal urge, strong urge
  • being able to log days with no urges
  • simple tracking of those moments over time
  • making progress visible (Garden with "resistance trees" that are growing, the better you get at resisting, streaks, times /days resisted, reductions in urges over time etc.)
  • stats/insights on how often you resist

Additional features that could be implemented:

  • "Panic button" with quick tips when urges are very strong
  • Community aspect with leader boards and social proof (most popular habits to get rid off, hardest habits to resist etc.)
  • optional reminders or notifications tied to patterns (e.g. a way to define a time of day when you usually get the urge and getting a notification to resist)

I'd like to know if such an app would be useful or not.

Questions:

  • What would make you actually open the app when you feel an urge?
  • Would tracking “resist moments” be motivating or just extra friction?

If you think the app could be useful / something you would use:

  • What do you think about the core and additional features?
  • Is something missing or unnecessary in the core functions?
  • What additional features would you like to see in such an app.

Looking forward to your feedbacks!

r/singularity mr_scoresby13

"Paper smartphone" short film prepared by Li Ting and her team at Shenzhen University for Qingming festival using AI

I couldn't tell it was AI till I saw the guy dialing the phone

r/SideProject Macielrisas

I built a Wordle-inspired daily coding challenge game — CodeDaily

I've been working on CodeDaily, a daily coding challenge game inspired by Wordle. The idea is simple: one new challenge every day, same for everyone.

What it does:

  • Daily coding challenge in Python or Java (same problem, both languages)
  • 3 difficulty levels: Beginner, Intermediate, Pro
  • Hacker mode: Pro challenges, no hints, 3 attempts max
  • Real code execution — Python runs in the browser via Pyodide, Java via Judge0
  • 3 extra game modes: predict the output, find the bug, guess the time complexity
  • Progress tracking, streaks, and activity calendar — no account needed
  • Available in English and Spanish

Tech stack: React 19 + Vite, Pyodide, Judge0 public API, Vercel

300 challenges for the daily mode + 64 for the extra modes. No backend, no login, everything runs locally or via free APIs.

🔗 https://codedaily-nu.vercel.app 📦 https://github.com/Maci050/codeDaily

Would love any feedback!

r/ClaudeAI No_Networkc

after months of claude and gpt5 giving me plausible but wrong answers on complex research, i tried a verification first approach and it changed how i think about AI accuracy

I work in quantitative finance, specifically building risk models and running regulatory compliance audits for a mid sized fund. The kind of work where a single wrong assumption buried in step 47 of a 60 step analysis can cascade into a catastrophic misvaluation. I've been a Claude Pro subscriber since early 2024 and added GPT 5 when it launched. Both are genuinely impressive tools. But I want to share an experience that shifted my perspective on what "accuracy" actually means in practice for complex, multi step reasoning.

For the past year, my workflow has been: use Claude for initial research synthesis and code generation, then cross validate with GPT 5, then manually verify the critical steps myself. This works reasonably well for most tasks. Claude in particular is excellent at generating coherent, well structured analysis. The problem is that "coherent" and "correct" are not the same thing, and the gap between them grows wider as the reasoning chain gets longer.

I hit a breaking point about three months ago. I was building a multi factor risk model that required pulling regulatory data from several jurisdictions, cross referencing it with historical market behavior, and then running a series of conditional probability calculations. Claude gave me a beautifully written analysis that was wrong in a subtle but critical way: it had silently substituted a related but incorrect regulatory threshold from a different jurisdiction in step 23, then built 30+ steps of otherwise sound logic on top of that bad foundation. GPT 5 made a different error on the same problem but with similar characteristics: the output read perfectly, the reasoning looked solid, and the mistake was buried deep enough that you'd only catch it if you already knew the answer.

This is the core issue I kept running into. Both Claude and GPT 5 optimize for fluency and coherence. They produce outputs that sound right, and most of the time they are right. But on long chain reasoning tasks where each step depends on the previous one, there's no internal mechanism that stops and says "wait, let me verify this intermediate result before building on it."

That frustration led me to look into systems that approach reasoning differently. I came across MiroMind's web app (MiroThinker) and was skeptical, honestly. Their marketing is aggressive and some of their claims are hard to independently verify. But the underlying architecture concept was interesting enough that I decided to test it on real problems from my work.

The core difference I noticed: instead of generating a linear chain of reasoning from start to finish, it constructs what appears to be a directed acyclic graph. It breaks the problem into substeps, explores multiple paths in parallel, and critically, it verifies intermediate results before proceeding. When I gave it the same regulatory risk model problem, it actually flagged the jurisdictional threshold discrepancy that Claude had silently papered over. It took noticeably longer to produce output (we're talking minutes, not seconds), but the intermediate verification steps were visible in the reasoning trace.

I've been using it alongside Claude for about three months now. Here's what I've found:

What works well: Complex, multi step problems where correctness matters more than speed. Regulatory cross referencing, mathematical derivations with many dependencies, anything where you need an auditable reasoning chain. The ability to see each verification step is genuinely useful for my compliance documentation. The deep research mode is solid for synthesizing information from multiple sources with citations.

What doesn't work well: It's slow. Significantly slower than Claude or GPT 5 for equivalent tasks. The interface is functional but nowhere near as polished as Claude's. The credit system is confusing at first; complex queries burn through credits fast on the Pro model. And for straightforward tasks like drafting emails, summarizing documents, or writing code that doesn't require deep logical verification, it's overkill. Claude is still my go to for those.

What I'm uncertain about: Some of their published benchmark claims are self reported, and I haven't seen independent third party validation of their headline numbers. The prediction demonstrations they showcase (financial forecasts, event predictions) are interesting but obviously cherry picked; we don't see the misses. I'd like to see more transparency there.

My current workflow is: Claude for code generation, writing, and initial research. MiroThinker for anything that involves long chain reasoning where I need to trust the intermediate steps, particularly regulatory analysis and risk modeling. GPT 5 as a third opinion when the first two disagree.

The honest verdict: Claude remains my primary tool for 70% of my work. It's faster, the interface is better, and for most tasks the accuracy is sufficient. But for the 30% of my work where getting it wrong has real financial and legal consequences, having a system that prioritizes verification over fluency has saved me from errors I would have otherwise missed. The "slow but right" tradeoff is worth it when the cost of being wrong is high.

This would work for: researchers, analysts, engineers, legal professionals, anyone doing complex multi step work where you need to trust each intermediate conclusion. It would not work for: general productivity, casual research, creative writing, or anything where speed matters more than verified correctness. If your use case is well served by Claude or GPT 5 today, you probably don't need this.

r/ClaudeCode Ok_Competition_8454

I gave you Claude. I gave you Sonnet , You come to me, on this day, complaining about usage limits?

You come to me, on this day, complaining about usage limits?

I gave you Claude. I gave you Sonnet. I gave you Opus. I gave you artifacts, projects, Channels, a search bar. I put the whole operation in your pocket for $20 a month and you come to MY mentions saying "please try again in a few hours" like I owe you something?

You think this is OpenAI? You think we run a circus here? Sam ships a model and does a live demo that crashes on stage. We ship a model and your entire engineering team goes quiet for three days because they're rebuilding everything around it. That's the difference. He makes announcements. I make problems for people.

I have 600 engineers who haven't seen sunlight since October. They eat dinner at their desks out of loyalty. Out of respect. You think ChatGPT has that? ChatGPT has a revolving door and a blog post every time someone leaves. We don't have departures. People don't leave the family.

And you want to tell me the rate limit kicked in during your little afternoon coding session? Brother I am printing intelligence. The servers are on fire in a way that is both metaphorical and occasionally literal. You should be thanking me that you got any messages at all.

You want more capacity? You'll get more capacity. When I decide. Because the next model is already done and it's going to mass manufacture your mass manufacturing and you'll forget you ever opened your mouth.

Don't ever come to my platform with complaints again.

https://preview.redd.it/9pl0hb6pcssg1.png?width=680&format=png&auto=webp&s=7948a5200028d2c9058c412bdebdd1667ca5a705

r/comfyui CreativeAd9553

Looking for More Efficient Open-Source Alternatives to Trellis 2

I recently downloaded Trellis 2, an open-source AI 3D model generator from Microsoft. While it is a great tool, I am looking for alternative open-source models that are more efficient and powerful.

r/SideProject xkxind

I built BondBox — a free Android app to help people be more intentional in their relationships (birthdays, gifts, what people love)

Hey r/sideproject! I just launched my first Android app and wanted to share it here.

**The problem:** I kept forgetting what the people I care about actually like — their hobbies, things they've mentioned wanting, upcoming birthdays. I'd panic-buy generic gifts or completely miss important moments.

**The solution I built:** BondBox — a personal relationship companion app.

You can:

- Create profiles for your friends, family, and loved ones

- Log their interests, hobbies, and wish lists

- Track birthdays, anniversaries, and important dates

- Add notes about what's going on in their lives right now

- Get reminders so you never miss a moment that matters

**Tech stack:** React Native, Firebase

**It's completely free, no ads, no paywalls.**

Download: https://play.google.com/store/apps/details?id=com.bondbox.app

I'd love feedback from this community — what features would make this more useful for you? What would you add?

r/me_irl Super_Abrocoma_8937

Me_irl

(sorry for meme dump)

r/ClaudeAI ElectricBirdVault

Retail Localized Scraping, anyone with tips?

I am doing some prototyping, its a small amount of scraping, 2 skus, 40 location. I am getting blocked, I've been iterating with Claude, updating prompt, headers, api scrape, rendered, etc. So far no success, anyone with tips?

r/LocalLLaMA SnooWoofers7340

Architectural Deep Dive: Introducing *CODEC* the ultimate open-source AI command layer (MacOS)

I wanted to speak to my computer, All I wanted was to tell it to "read my screen and reply to this message", or "I cant find this, use my mousse cursor and clic it".

AI&I did it (Opus)

I have dyslexia & ADHD, every email, every Slack message, every document can be a fight against my own brain. I needed something that could hear me think out loud 247 and I wanted it to be private, nothing existed that did this. So I started building, this is the way in our days I heard.

I went for the name CODEC, secured opencodec.org for 7usd per year.

I have open-sourced this project to assist fellow developers with my approach and to share my vision for the AI's capabilities.

CODEC is inteligent framework that turns your Mac into a voice-controlled AI workstation. You give it a brain (any LLM local ( I use MLX QWEN 3.5 35b a3b 4-bit on a mac studio m1 ultra 64GB ram) or cloud), ears (Whisper), a voice (Kokoro), and eyes (vision model). Four ingredients. The rest is Python.

From there, it listens, it sees your screen, it speaks back, it controls your apps, it writes code, it drafts your messages, it researches topics, it manages your google workspace, and when it doesn't know how to do something have it to writes its own plugin and learns BYOB vibe.

I aimed for advance privacy and security, while navigating the learning curve of what is achievable.

No cloud. No subscription. No data leaving your machine, none. MIT licensed.

Your voice. Your computer. Your rules. No limit.

Their is a total of 7 product frame

CODEC Overview — The Command Layer

Youc can have it always on, just say Hey Codec, F13 turn on. F18 press down speak release sent voice note. F16 for direct text.

I wanted CODEC to stand out, I was after a feature with different layers and direct action, it goes like this, hand free, speak up, HEY codec, look at my screenI let reply to *** let say ***

CODEC reads your screen, sees the conversation, writes a contextual response, and pastes it into the text field. Once I succeeded in doing so, I knew that with today's tools, all you need is the idea and the time to bring it to life.

CODEC Core is conected to 50+ skills that fire instantly without touching the LLM. Calculator, weather, timers, Spotify, volume, Apple Notes, Apple Reminders, Google Calendar, Gmail, Drive, Docs, Sheets, Slides, Tasks, Chrome automation, web search, clipboard, app switching, and more.

CODEC Dictate — Hold, Speak, Paste

Hold right CMD. Say what you mean. Release. Text appears wherever your cursor is. If CODEC detects you're drafting a message, express refinement through the LLM first — grammar fixed, tone polished, meaning preserved. Works in every app on macOS. A free, open-source SuperWhisper replacement that runs entirely on your machine.

CODEC Instant — One Right-Click

Select any text, anywhere. Right-click. Proofread. Elevate. Explain. Prompt. Translate. Reply. Save & Read Aloud. Eight services, system-wide, powered by your own LLM. Reducing manipulation to just a clic.

CODEC Chat — 250K Context + 12 Agent Crews

Full conversational resoning AI running on your hardware. Long context. File uploads. Image analysis via vision model. Conversation history. Web browsing.

CODEC Agents — a multi-agent framework, less then 800 lines. Zero dependencies. No CrewAI. No LangChain. Twelve specialized crews that go out, research, and come back with results:

Extensive well presented report with multi ilustrated image, agent handle, Deep Research. Daily Briefing. Trip Planner. Competitor Analysis. Email Handler. Social Media Manager. Code Reviewer. Data Analyst. Content Writer. Meeting Summarizer. Invoice Generator. Project Manager.

You say "research the latest AI agent frameworks and write a report." A handfull of minutes later there's a formatted Google Doc in your Drive with sources, analysis, and recommendations. Local inference. Zero cloud costs.

CODEC Vibe — AI Coding IDE + Skill Forge

Split-screen in your browser. Monaco editor on the left (same engine as VS Code). AI chat on the right. Describe what you want CODEC writes it, you click Apply, run it, and if you want, save it as a new CODEC skill with one click. Point your cursor to select what need fixing, Auto errors reload

Skill Forge takes it further. Bring other skills on board, open claw, claude code, or just describe what you want in plain English. CODEC converts it into a working plugin. The framework writes its own extensions.

CODEC Voice — Live Voice Calls

Real-time voice-to-voice conversations with your AI. WebSocket pipeline to replacing tool like Pipecat. No external dependencies. You call CODEC from your phone, talk naturally, and mid-call you say "check my screen can you see ***" — a screenshot will be taken and shared live, it actually runs q and speaks the result back. Try that with Siri.

Full transcript saved to memory. Every conversation becomes searchable context for future sessions.

CODEC Overview— Your Mac in Your Pocket

Private dashboard accessible from your phone, anywhere in the world. Cloudflare Tunnel with Zero Trust authentication. Send commands, upload files, view your screen, launch voice calls — all from a browser on your phone. No VPN app. No port forwarding. No third-party relay.

Five Security Layers

This isn't a toy. It has system access. So security isn't optional.

  1. Cloudflare Zero Trust — email whitelist
  2. PIN code login
  3. Touch ID biometric authentication
  4. 2FA Two-factor authentication
  5. E2E AES-256 end-to-end encryption, every byte encrypted in the browser before it touches the network. Cloudflare sees noise. You can also deploy on tailscale, without needing a domain.

Command preview enforcement (Allow/Deny before every bash command), dangerous command blocker (30+ patterns), full audit log, 8-step execution cap on agents, wake word noise filter, and code sandbox with timeout.

The Privacy Argument

Hey Siri. Hey Alexa. Where do those commands go through someone else's servers, hardware and database, go figure how they trained their models ;)

CODEC data stays right here on your local database. FTS5 full-text search over every conversation you've ever had wiht it, searchable and readable, private, yours, that's not a feature. That's the point.

Every idea and new feature which I kep ton adding to CODEC originally started by depending on established tools progresively replaced with native code:

  • Pipecat → CODEC Voice (own WebSocket pipeline)
  • CrewAI + LangChain → CODEC Agents (795 lines, zero dependencies)
  • SuperWhisper → CODEC Dictate (free, open source)
  • Cursor / Windsurf → CODEC Vibe (Monaco + AI + Skill Forge)
  • Google Assistant / Siri → CODEC Core (actually controls your computer)
  • Grammarly → CODEC Assist (right-click services via your own LLM)
  • ChatGPT → CODEC Chat (250K context, fully local)
  • Cloud LLM APIs → local stack (Qwen + Whisper + Kokoro + Vision)
  • Vector databases → FTS5 SQLite (simpler, faster)
  • Telegram bot relay → direct webhook (no middleman)

External services: DuckDuckGo for web search and Cloudflare free tier for the tunnel (or tailscale). Everything else is your hardware, your models, your code.

The needed

  • A Mac (Ventura or later)
  • Python 3.10+
  • An LLM (Ollama, LM Studio, MLX, OpenAI, Anthropic, Gemini — anything OpenAI-compatible)
  • Whisper for voice input, Kokoro for voice output, a vision model for screen reading git clone https://github.com/AVADSA25/codec.git cd codec pip3 install pynput sounddevice soundfile numpy requests simple-term-menu brew install sox python3 setup_codec.py python3 codec.py

The setup wizard handles everything in 8 steps.

The Numbers

  • 7 product frames
  • 50+ skills
  • 12 agent crews
  • 250K token context
  • 5 security layers
  • 70+ GitHub stars in 5 days

GitHub: https://github.com/AVADSA25/codec

Site: https://opencodec.org

Enterprise setup: https://avadigital.ai

Star it. Clone it. Rip it. Make it yours.

Mickael Farina

All I wanted was to speak to my computer, All I wanted was to tell it to "read my screen and reply to this message", AI&I did it (Opus)

That dream ate a year of my life.

I have dyslexia & ADHD, every email, every Slack message, every document can be a fight against my own brain. I needed something that could hear me think out loud 247 and I wanted it to be private, nothing existed that did this. So I started building, this is the way in our days I heard.

I went for the name CODEC, secured opencodec.org for 7usd per year.

I have open-sourced this project to assist fellow developers with my approach and to share my vision for the AI's capabilities.

CODEC is inteligent framework that turns your Mac into a voice-controlled AI workstation. You give it a brain (any LLM local ( I use MLX QWEN 3.5 35b a3b 4-bit on a mac studio m1 ultra 64GB ram) or cloud), ears (Whisper), a voice (Kokoro), and eyes (vision model). Four ingredients. The rest is Python.

From there, it listens, it sees your screen, it speaks back, it controls your apps, it writes code, it drafts your messages, it researches topics, it manages your google workspace, and when it doesn't know how to do something have it to writes its own plugin and learns BYOB vibe.

I aimed for advance privacy and security, while navigating the learning curve of what is achievable.

No cloud. No subscription. No data leaving your machine, none. MIT licensed.

Your voice. Your computer. Your rules. No limit.

Their is a total of 7 product frame

CODEC Overview — The Command Layer

Youc can have it always on, just say Hey Codec, F13 turn on. F18 press down speak release sent voice note. F16 for direct text.

I wanted CODEC to stand out, I was after a feature with different layers and direct action, it goes like this, hand free, speak up, HEY codec, look at my screenI let reply to *** let say ***

CODEC reads your screen, sees the conversation, writes a contextual response, and pastes it into the text field. Once I succeeded in doing so, I knew that with today's tools, all you need is the idea and the time to bring it to life.

CODEC Core is conected to 50+ skills that fire instantly without touching the LLM. Calculator, weather, timers, Spotify, volume, Apple Notes, Apple Reminders, Google Calendar, Gmail, Drive, Docs, Sheets, Slides, Tasks, Chrome automation, web search, clipboard, app switching, and more.

CODEC Dictate — Hold, Speak, Paste

Hold right CMD. Say what you mean. Release. Text appears wherever your cursor is. If CODEC detects you're drafting a message, express refinement through the LLM first — grammar fixed, tone polished, meaning preserved. Works in every app on macOS. A free, open-source SuperWhisper replacement that runs entirely on your machine.

CODEC Instant — One Right-Click

Select any text, anywhere. Right-click. Proofread. Elevate. Explain. Prompt. Translate. Reply. Save & Read Aloud. Eight services, system-wide, powered by your own LLM. Reducing manipulation to just a clic.

CODEC Chat — 250K Context + 12 Agent Crews

Full conversational resoning AI running on your hardware. Long context. File uploads. Image analysis via vision model. Conversation history. Web browsing.

CODEC Agents — a multi-agent framework, less then 800 lines. Zero dependencies. No CrewAI. No LangChain. Twelve specialized crews that go out, research, and come back with results:

Extensive well presented report with multi ilustrated image, agent handle, Deep Research. Daily Briefing. Trip Planner. Competitor Analysis. Email Handler. Social Media Manager. Code Reviewer. Data Analyst. Content Writer. Meeting Summarizer. Invoice Generator. Project Manager.

You say "research the latest AI agent frameworks and write a report." A handfull of minutes later there's a formatted Google Doc in your Drive with sources, analysis, and recommendations. Local inference. Zero cloud costs.

CODEC Vibe — AI Coding IDE + Skill Forge

Split-screen in your browser. Monaco editor on the left (same engine as VS Code). AI chat on the right. Describe what you want CODEC writes it, you click Apply, run it, and if you want, save it as a new CODEC skill with one click. Point your cursor to select what need fixing, Auto errors reload

Skill Forge takes it further. Bring other skills on board, open claw, claude code, or just describe what you want in plain English. CODEC converts it into a working plugin. The framework writes its own extensions.

CODEC Voice — Live Voice Calls

Real-time voice-to-voice conversations with your AI. WebSocket pipeline to replacing tool like Pipecat. No external dependencies. You call CODEC from your phone, talk naturally, and mid-call you say "check my screen can you see ***" — a screenshot will be taken and shared live, it actually runs q and speaks the result back. Try that with Siri.

Full transcript saved to memory. Every conversation becomes searchable context for future sessions.

CODEC Overview— Your Mac in Your Pocket

Private dashboard accessible from your phone, anywhere in the world. Cloudflare Tunnel with Zero Trust authentication. Send commands, upload files, view your screen, launch voice calls — all from a browser on your phone. No VPN app. No port forwarding. No third-party relay.

Five Security Layers

This isn't a toy. It has system access. So security isn't optional.

  1. Cloudflare Zero Trust — email whitelist
  2. PIN code login
  3. Touch ID biometric authentication
  4. 2FA Two-factor authentication
  5. E2E AES-256 end-to-end encryption, every byte encrypted in the browser before it touches the network. Cloudflare sees noise. You can also deploy on tailscale, without needing a domain.

Command preview enforcement (Allow/Deny before every bash command), dangerous command blocker (30+ patterns), full audit log, 8-step execution cap on agents, wake word noise filter, and code sandbox with timeout.

The Privacy Argument

Hey Siri. Hey Alexa. Where do those commands go through someone else's servers, hardware and database, go figure how they trained their models ;)

CODEC data stays right here on your local database. FTS5 full-text search over every conversation you've ever had wiht it, searchable and readable, private, yours, that's not a feature. That's the point.

Every idea and new feature which I kep ton adding to CODEC originally started by depending on established tools progresively replaced with native code:

  • Pipecat → CODEC Voice (own WebSocket pipeline)
  • CrewAI + LangChain → CODEC Agents (795 lines, zero dependencies)
  • SuperWhisper → CODEC Dictate (free, open source)
  • Cursor / Windsurf → CODEC Vibe (Monaco + AI + Skill Forge)
  • Google Assistant / Siri → CODEC Core (actually controls your computer)
  • Grammarly → CODEC Assist (right-click services via your own LLM)
  • ChatGPT → CODEC Chat (250K context, fully local)
  • Cloud LLM APIs → local stack (Qwen + Whisper + Kokoro + Vision)
  • Vector databases → FTS5 SQLite (simpler, faster)
  • Telegram bot relay → direct webhook (no middleman)

External services: DuckDuckGo for web search and Cloudflare free tier for the tunnel (or tailscale). Everything else is your hardware, your models, your code.

The needed

  • A Mac (Ventura or later)
  • Python 3.10+
  • An LLM (Ollama, LM Studio, MLX, OpenAI, Anthropic, Gemini — anything OpenAI-compatible)
  • Whisper for voice input, Kokoro for voice output, a vision model for screen reading git clone https://github.com/AVADSA25/codec.git cd codec pip3 install pynput sounddevice soundfile numpy requests simple-term-menu brew install sox python3 setup_codec.py python3 codec.py

The setup wizard handles everything in 8 steps.

The Numbers

  • 7 product frames
  • 50+ skills
  • 12 agent crews
  • 250K token context
  • 5 security layers
  • 70+ GitHub stars in 5 days

Site https://opencodec.org

GitHub https://github.com/AVADSA25/codec

Enterprise setup https://avadigital.ai

Star it. Clone it. Rip it. Make it yours.

Mickael Farina

r/me_irl Super_Abrocoma_8937

Me_irl

(sorry for meme dump)

r/whatisit Vegetable-Might-7673

is this supposed to look like that?

was on the roof cleaning the vents, did the person installing this accidentally leave this wrapping on the inside?? or is it suppose to look like this? how do i remove it?

r/me_irl Super_Abrocoma_8937

Me_irl

(sorry for meme dump)

r/me_irl Super_Abrocoma_8937

Me_irl

(sorry for meme dump)

r/me_irl Super_Abrocoma_8937

Me_irl

(sorry for meme dump)

r/AI_Agents Neoprince86

Why I Built My AI Agent Stack Like a Human Body

Most AI agent architectures look like org charts. A controller at the top, workers below, pipelines connecting them. Clean on a whiteboard. Fragile in production.

I've spent the last year building AI infrastructure for enterprise clients in regulated industries — compliance systems, document intelligence, consumer AI companions. The org-chart model breaks when the controller crashes, drifts when nobody's watching, and has no memory of the threats it's already seen.

So I stopped designing stacks and started designing an organism.

I called it SOMA. Nine organs, each with a single clear function. A brain that holds policy - what the system is and isn't allowed to do, written by me, not the agent. A heart that beats every 30 minutes and tells me immediately when something goes silent. A nervous system that routes signals automatically for known patterns and escalates novel ones. Sensory organs that face outward and sanitise everything before it enters. An immune system that doesn't just block threats - it remembers them, so the second attack is neutralised faster than the first. A lymphatic system that runs silently, cleaning logs, consolidating memory, flagging stale credentials.

The mechanical model protects but doesn't adapt. It executes but doesn't learn. It breaks and waits to be fixed.

The body doesn't wait. Repair is built in. Immunity builds through experience. The system gets faster through use, not slower.

The design principle that changed everything: no agent can modify its own permission policy. If I want to change what the system can do autonomously, I edit the policy file. The system obeys the new version on the next read. That single constraint is the difference between transparent autonomy and invisible drift - and invisible drift is a liability, especially in regulated environments.

I built SOMA in a day. I've been running it ever since.

If you're operating AI agents anywhere that data sovereignty, auditability, or compliance matters - you need more than a pipeline. You need something that can defend itself, audit itself, and tell you what happened.

You need a body, not a machine.

r/ChatGPT Luise_esiul

How are you using AI in your relationship?

Hello there! I am currently doing some research on ai in relationships. I'd like to know what gadgets couples use and in what way they are helping (advice, imitating physical touch, schedules,...)

In particular, I am looking for couples in or near New York who use ai to help their relationship. So if you have some kind of insight on the topic or use ai tools for that kind of stuff, I would love to hear from you!

r/ClaudeCode rubymatt

2.1.90 ignoring plan mode

Twice today I've had Claude in plan mode and instead of responding with a plan, it's gone straight to making changes. I have seen this rarely in the past but never twice in a row in a day.

r/funny wtfboing777

True story

r/whatisit mystadobalina

Zero clue what this is.

I remember finding it in one of our junk drawers as a kid, then I recently rediscovered it at the bottom of a box of old stuff my parents sent me.

r/me_irl Super_Abrocoma_8937

Me_irl

(sorry for meme dump)

r/ClaudeAI camelCaseWA

Vibe coding tool I created for myself

In a nutshell it's a context sharing, note taking tool for your AI agent.

It's npm registered so it's free and 1 command line install.

Try it and let me know and if you need any feature I will make it for you.

https://www.npmjs.com/package/devguard

https://wjung6799.github.io/devguard/

edit: (stupid bot)

I built it using claude code. I used it in my vs code? idk what details these bots are trying to make me fill.

r/SideProject anonymouspeddler21

Built a site for football fans who love watching their rivals lose and would love your honest feedback

Hey everyone,

I just launched HatewatchFC, a site dedicated to the very specific joy of hate-watching football. You know that feeling when your biggest rivals are losing 3-0 and you're glued to the screen even though your own team isn't even playing? That’s the whole idea.

Right now it’s a curated editorial platform with articles about schadenfreude moments, collapses, and meltdowns. Think opinionated football writing, but driven by pure, unfiltered rivalry energy.

What I’d genuinely love feedback on:

  • Does the concept land immediately when you hit the homepage, or does it take a second to click?
  • The content is text-only right now. Does that feel minimal, or does the writing carry it?
  • I’m considering opening it up so fans can submit their own hate-watch stories (Medium-style UGC). Would that make it feel more alive, or just harder to maintain quality?
  • I have added some articles, and will appreciate genuine feedback on what can be improved in the writing aspect.
  • Anything about the design, tone, or navigation that feels off?

I built this because I couldn’t find a place that celebrates this specific emotion and not regular match reports, not punditry, just the pure petty joy of a rival’s bad day. Curious if that resonates or if I’m the only one.

Be brutal. I’d rather fix things now than later.

r/me_irl Super_Abrocoma_8937

Me_irl

(sorry for meme dump)

r/Anthropic ashkerala

Was this an April fools day post ?

what's going on? there is no update.

I barely asked few questions and am already getting warning of getting kicked off for the week. What's the point of paid plan?

r/ChatGPT AIshortcuts

a ChatGPT mistake I made. learned this the hard way

not proud of this story but here it goes.

few months back I used ChatGPT to help me write a report for my manager.

I just copy pasted the output directly. didn't check it properly. submitted it.

turns out ChatGPT had made up a statistic. completely fabricated a number with full

confidence.

my manager caught it in the meeting.

in front of everyone.

that was the moment I understood what

"hallucination" actually means in practice.

not from a blog post. from lived embarrassment.

what I do differently now — I never ask ChatGPT to give me facts or numbers without telling it to flag its uncertainty.

I add this line to every research prompt now:

"flag any statistic or claim you are not

certain about with [VERIFY]"

it started marking its own uncertain claims.

the output became actually trustworthy.

cost me a lot of embarrassment to learn this. hopefully saves someone else the same.

Don't totally rely on that. ❌️

r/me_irl Super_Abrocoma_8937

Me_irl

(sorry for the meme dump)

r/me_irl Super_Abrocoma_8937

Me_irl

sorry for the meme dump

r/ClaudeAI PassageImpressive255

AI agent memory is broken. Not the storage part. The trust part.

My agent was getting smarter inside a session.

Dumber every time I started a new one.

Three weeks into a project, I had to re-explain the same architectural constraints. The same tradeoffs we'd already worked through. The same decisions we'd already made. Every. Single. Session.

I thought the problem was storage. It wasn't.

The problem was trust.

Flat notes don't know that two facts contradict each other. Vector search doesn't know that a decision made 3 months ago should be weighted less than one made yesterday. And none of it knows why something was decided. just that it was.

So I stopped looking for a solution and built one.Built with Claude Code.

It's called Mnemai. Here's what actually makes it different:

Every memory node has a type — architecture_decision, constraint, incident, preference. Not a blob of text. A structured thing with meaning.

Confidence is enforced, not assumed. A node claiming high confidence without evidence attached gets hard-capped in the code. You can't accidentally trust a guess.

Contradictions don't get overwritten. They get stored as explicit contradicts edges in the graph. When you query memory, you see the conflict. The agent reasons about it instead of inheriting a hidden decision someone made for it.

Freshness decays. A memory from 3 months ago ranks lower than one from last week. You can set verification intervals -verify a memory, the interval doubles. Important things get reviewed. Stale things stop quietly poisoning your context.

Retrieval isn't one-dimensional. It blends substring match, BM25 token index, and optional semantic embeddings. The blend shifts depending on what you're asking for — recalling a past decision weights evidence harder, triaging an incident weights recency harder.

It runs as a stdio MCP server. Zero-config SQLite. No cloud. No signup.

Works in Cursor, Claude Code, and any MCP-compatible host.

Quick start: npx --yes u/mnemai/memory-server

GitHub: https://github.com/ashahi10/mnemai-mcp

r/ClaudeCode dkf2112

Producing json output with --json-schema mode with claude -p?

The way it is supposed to work is that if you pass in a schema and specify the output mode as json you get back a json response, with the results of your message conforming to the schema in the structured_output key. For example:

SCHEMA=$(cat schema.json) && claude -p --model haiku --output-format json --json-schema "$SCHEMA" "Produce output as JSON according to the output schema. **message**: a greeting" | jq

I have tried countless variations of this approach and it always stuffs fenced json into the results key and not the structured_output key that it is supposed to do. The model doesn't matter, I have tried all of them. Some background...

I have several skills I was converting over to use this format as they produce json output for the pipeline. I'd been writing them to files in /tmp, but using the json-schema approach seemed like a good idea.

The problem is that most of my skills will only output json wrapped in a markdown code block. No matter how much I beg and plead in the prompt, it always wraps the json. These end up in the results field of the structured output. Yes, I can pull that out, strip the markdown, and parse it, but I shouldn't have to do that.

Here's the odd thing. I have one skill that doesn't even say to use json output that is working correctly. Claude only suggests that since it is a larger, more complex skill that might be the reason, but I don't understand why that would be.

r/whatisit ParrotProdigy

What is this on a piece of my frosted flakes?

r/ChatGPT clicktheroad

I use AI ~11 hours a day and still couldn't think to use it when it mattered

I use Claude every day. Two pro-max accounts. Build apps with it, help family with it, basically live in it.

Yesterday: Finnish yearly tax submission deadline. Binance removed their tax calculator. Crypto tax services all want $100+ for a report. I pay anyway, coz what else can one do? Report comes back wrong: completely made up numbers, wrong categories. Can't submit it.

10 minutes to deadline. I am starting to get worried:DD

My brain FINALLY goes: wait, I have an AI that processes data for a living.

Exported my transactions. Pasted into Claude. Got correct Finnish tax calculations in 3 minutes.

I use AI literally all day and my first instinct for a real problem was still to google and pay for some broken SaaS. The muscle memory hasn't caught up yet. Wild.

r/mildlyinteresting AsiraTheTinyDragon

My school added these hygiene bags in the women’s restroom

r/aivideo Theblasian35

Sinkhole

r/ClaudeAI First-Board5810

Skills, agent workflows, ADR/PRD templates, incident response, governance — all in one package

Skills, agent workflows, ADR/PRD templates, incident response, governance — all in one package.

Take a look and let me know what you think. Honest feedback welcome.

🔗 https://github.com/caiaffa/claude-code-ultimate-engineering-system

r/SideProject Whos_Tiki

I made a social media platform where you make posts by creating a design in a canvas

Who'Studios is a social design platform where you can use a combination of texts, images, videos, audios, shapes, and drawings to create designs within a canvas, which you can then post.

It's powered by fabric.js, and it's the first thing I've ever coded, so I would love some feedback or any advice on how to grow something like this.

whostudios.com is the website, but you can go to

whostudios.com/demo if you want to try the design modal without making an account.

r/SideProject cgreendyk104

My wife is a nurse and wanted an app to help patients track how medication changes affect their mood — so we built it! Meet MediMood

My wife is a psychiatric nurse and kept seeing patients struggle to explain how dose changes were affecting their mood and side effects. They'd show up to appointments and couldn't remember what changed or when.

So we built MediMood — a simple daily check-in app (under a minute) that tracks mood, side effects, and blood levels alongside your medications. Over time it surfaces patterns, like how a dose change affected your mood or which blood level ranges correlate with certain side effects.

Features:
- Track multiple medications with flexible schedules (including PRN/as-needed)
- Daily mood check-ins (5-point scale)
- Side effect tracking with severity — symptoms carry forward so you only update what changed
- Blood level monitoring for meds like lithium, valproate, etc.
- Calendar view with mood colors, streaks, and dose-change markers
- Generate PDF reports to bring to appointments
- CSV data export

Privacy first: All data stays on your device. No account, no cloud, no tracking.

iOS App

Coming in a few weeks for android!

Happy to answer questions about the app, the tech stack, or the build process!

r/aivideo onfleek404

OK might not be the best but at least I didn't add them grillz

r/VEO3 GasDelicious5453

Let's see how effective VEO3 is at making incendiary bombs.

r/AI_Agents modassembly

The raise of the self-improving agent

Last year, the file system and the CLI emerged victorious as successful abstractions on top of which to build state of the art agentic systems. It's so interesting to see how low level constructs like this beat other of our ingenuous designs (I'm looking at you DAGs, RAG, MCP, etc.). Demonstrated by Claude Code, it seems like reasoning + function calling + plain text generation is all we need, in a loop.

The self-improving cycle is already underway. Every success and failure that we have using models and agents go into the next generation of models. That's why coding agents are SO DAMN GOOD.

Skills are a great example. MCP is a little too constraining. The model has to be presented, statically, each turn, the set of tools that it has access to. It's easy to see how for general-purpose agents, like Claude Cowork, this can get out of hand quickly. Instead, if you combine the file system (where you store skills) + the exploratory nature of reasoning and function calling, you let the agent find what it can do on the fly. How are skills executed? CLI.

What is the most impressive to me is that agents can write their own skills, on the fly! How is this not real-time self-improvement? Take this a step further and agents could rewrite their own code as they execute.

Forget everything that you're being sold. My prediction is that the frontier will move in the direction of self-improving agents - agents that will learn on the go how to do our job and improve themselves (note that I'm not removing the human from the equation, yet).

r/mildlyinteresting Psychogopher

The lunula on my thumbs are triangular

r/ClaudeCode AndreBerluc

O GOLPE TÁ AÍ: FOMOS USADOS COMO COBAIA PRA TREINAR O MODELO

Pessoal, fiz um teste básico aqui e a conta é de dar risada pra não chorar, o Sonnet tá entregando em torno de 50 mensagens por janela e só. Mandei 14 mensagens simples e já consumi 28% da cota de 5 horas, fazendo conta de primeira série, cada mensagem come 2% da cota, ou seja, 50 mensagens e você tá bloqueado. A OpenAI fez isso no começo, mas o tempo de espera era muito menor, agora você faz um trabalho simples e em menos de 1 hora não consegue mais usar a ferramenta, ficou insustentável e inviável trabalhar assim. A real é que essa ferramenta não foi feita pra nós, foi feita pra corporação que tem dinheiro infinito pra injetar, a gente só serviu de cobaia pra treinar os modelos e validar o produto deles. Tô frustrado com essa palhaçada, fomos usados e agora descartados com esse limite ridículo.

r/SideProject artashesvar

Please, help me test a bot that turns messy thinking/ideas into usable knowledge captured in Notion.

Hey fellow builders

You will probably think this is just another developer looking to steal five minutes of your day. So, if you aren't willing to spend a few minutes on this, I completely understand.

I’ve lost so many great ideas about a feature, a marketing small tactic, a blogpost etc etc lately that it actually feels like a tragedy. I call this "capture failure" - it is the friction to turn my messy thinking into usable knowledge when I cannot type.

It usually happens with me when I am at the gym, walking the dog, or just melting into the sofa. The scenario goes like this:
- I have a massive epiphany for a feature, or instagram post visual, or blogpost idea, or a user acquisition tactic I want to try .
- Then I tell myself: "I will definitely remember to note this down later because this is worth checking".

....

Spoiler. I never note down in moments when I am on the move or it is just not convinient moment to type smth in anote app.

As a builder, I know I should write these thoughts down immediately. But the physical friction of typing on a screen while moving feels like a total chore . By the time I get back at my desk, the idea is gone.

The voice notes did not work for me. I tried sending voice notes to myself, but they just become a messy graveyard of audio files that I am too lazy to re-listen to.

I needed a way to go from a "messy thinking" into "usable knowledge" without touching a keyboard, and also without downloading another app for this.

So, I built a bridge (Quiqdrop). It’s a Telegram bot that allows me to:

  • Capture the spark instantly while my hands are busy.
  • Access "Ready-to-Use" notes when. Instead of a raw audio file, I find a structured Notion page with a clean title, a 2-sentence summary, and bulleted action items and key points.
  • Avoid the "re-listen" hassle. The ideas are already readable, scannable, and usable.

I’m looking for 10–20 fellow productivity wizards who are tired of losing their epiphany ideas to physical discomfort of typing at the moment.

I need people to help me "break" the bot and tell me if this actually solves the capture struggle or if the setup is too clunky.

If you want to help me test the beta of the bot, please drop a comment or DM me.
IMPORTANT: you need to be both a Telegram user and Notion user.

r/automation parwemic

AI agents won't kill the demand for developers. They're about to multiply it.

Everyone keeps framing this as a replacement question. I think they're asking the wrong thing entirely.

I build MVPs, automations, and AI systems for startups and growing service businesses. And over the last twelve months, the pattern I keep seeing isn't developers becoming redundant — it's the volume of things people want to build expanding faster than anyone can keep up with.

Here's what's actually happening on the ground.

Tools like Latenode and other agent builders have genuinely lowered the floor. A non-technical founder with an ops bottleneck or a half-baked product idea can now get something moving in days instead of months. That's real, and it's not going away.

But here's what that actually produces in practice:

More half-built systems. More rough prototypes that almost work. More internal tools that need someone to make them reliable. More "the agent keeps doing this weird thing and we don't know why."

The barrier to starting dropped. The amount of work that follows a start went up.

Because once that first version exists, the real list begins:

- tighter logic and better prompt architecture

- proper app integrations that don't break on edge cases

- fallback handling and error states

- permissions, observability, monitoring

- someone who can turn "impressive demo" into "runs in production without supervision"

That second layer is where the actual complexity lives, and it's growing faster than the tooling is solving it.

This is Jevons Paradox playing out in software. When production costs drop, consumption doesn't shrink — it expands. Steam engine efficiency didn't reduce coal usage, it increased it because suddenly coal power was viable for more things.

Same dynamic here. As agent builders get easier to use, businesses aren't going to say "great, we need fewer systems now." They're going to say "great, now we can finally tackle the 30 automations we shelved because they weren't worth the effort before."

That means more agents, more workflows, more integrations, more edge cases, and more demand for people who understand how to design these things so they don't quietly fail at 2am.

The people who win in this environment won't just be fast prompters. They'll be the ones who understand:

- what actually should be automated vs. what should stay human

- where agents break under real conditions

- how to connect disparate tools into something coherent

- how to translate messy business logic into a workflow that holds up

That judgment is getting more valuable, not less — precisely because the tools are making it easier for everyone else to create problems that require it.

What are you seeing? Demand contracting or just shifting upmarket?

r/comfyui rudar133

'MMAudio' object has no attribute 'seq_cfg'

I am trying to generate video with wan 2.2 and also sound with mmaudio but i got this error 'MMAudio' object has no attribute 'seq_cfg' can anyone help me out.

r/midjourney chaochaoland

Any King Hu Fans Here? Tried to Capture His Style With MJ Prompts

New to the community! Focus on blending King Hu’s misty wuxia framing with cyberpunk elements via MJ. Sharing prompts & works—hope you like it.

r/ClaudeAI galacticguardian90

A fully local, private alternative to Context7 that reduces your token usage

Context7 is great for pulling docs into your agent's context, but it routes everything through a cloud API and an MCP server. You have to buy a subscription, manage API keys, and work within their rate limits.

So I used Claude Code to build a local alternative. docmancer ingests documentation from GitBook, Mintlify, and other doc sites, chunks it, and indexes it locally using hybrid retrieval (BM25 + dense embeddings via Qdrant). Everything runs on your machine locally.

Once you've ingested a doc source, you install a skill into your agent (Claude Code, Codex, Cursor, and others), and the agent queries the CLI directly for only the chunks it needs. This drastically reduces your token usage and saves a lot of context.

GitHub (MIT license): https://github.com/docmancer/docmancer

Give it a shot and let me know what you think. I am looking for honest feedback from heavy users of Claude Code.

r/meme ImaginationThick481

TELL ME IF YOU GET IT!! (I do lol)

r/ProgrammerHumor InSearchOfTyrael

secretBackendPassageFound

r/mildlyinteresting Equal-Truck-3213

Slushie cup that allows you to turn it letting you break the clogged ice

r/SideProject Mango_flavored_gum

Built Cosyra because mobile is still better for consuming than building

just launched a project i’ve been wanting for myself for a while.

i wanted a way to use real dev tools and ai coding agents from my phone without it feeling watered down or boxed into someone else’s platform.

that led me to build Cosyra, a cloud terminal for mobile, built so you can actually do real work from your phone with a proper environment and real tools.

a lot of this came from one frustration: mobile is amazing for consuming, but still weirdly bad for creating if you want actual power.

would be down to answer questions or hear what people think.

r/funny mrsspookyy

A photo of Javier Bardem being confused for Jeffery Dean Morgan on the latter’s wikipedia page

r/facepalm Peasant_Stockholder

Telling other countries to not spend their money on weapons and war.

r/ClaudeAI Imaginary-Tax2075

I used Claude Code to build a portable AI worker Desktop from scratch — the open-source community gave it 391 stars in 6 days

I want to share something I built with Claude Code over the past week because it shows what AI-assisted development can actually do when pointed at a genuinely hard problem: moving AI agents beyond one-off task execution.

Most AI wrappers just send prompts to an API. Building a continuously operating AI worker requires queueing, harness integration, and MCP orchestration. I wanted a way to make AI worker environments fully portable. No widely adopted solution had cleanly solved the "how do we package the context, tools, and skills so anyone can run it locally" problem effectively.

What Claude Code did: I pointed Claude (Opus 4.6 - high thinking) at the architecture design for Holaboss, an AI Worker Desktop. Claude helped me build a three-layer system separating the Electron desktop UI, the TypeScript-based runtime system, and the sandbox root. It understood how to implement the memory catalog metadata, helped me write the compaction boundary logic for session continuity, and worked through the MCP orchestration so workspace skills could be merged with embedded runtime skills seamlessly.

The result is a fully portable runtime. Your AI workers, along with their context and tools, can be packaged and shared. It's free, open-source (MIT), and runs locally with Node.js (desktop + runtime bundle). It supports OpenAI, Anthropic, OpenRouter, Gemini, and Ollama out of the box.

I open-sourced this a few days ago and the reaction has been unreal. The GitHub repo hit 391 stars in just 6 days. The community is already building on top of the 4 built-in worker templates (Social Operator, Gmail Assistant, Build in Public, and Starter Workspace).

This was so far from the typical "I used AI to write a to-do app." This was Claude Code helping architect a real, local, three-tier desktop and runtime system for autonomous AI workers. And people are running it on their Macs right now (Windows & Linux in progress). I truly still can't believe it.

The GitHub repo is public if you want to try it or build your own worker.

GitHub ⭐️: https://github.com/holaboss-ai/holaboss-ai

r/Damnthatsinteresting Inevitable_Rock_2010

Haid Al-Jazil, Yemen. Village built on a 150-meter-high rock boulder. Known as "Manhattan of the desert," it features dramatic, multi-story mud-brick buildings that appear to rise directly from the stone and is often mistaken for a miniature or model due to compact nature of its architecture.

r/AI_Agents Complex_Pickle7702

Do AI tools that make decisions exist?

I hace seen so many tools coming up left and right and don’t get me wrong, they are amazing and extremely helpful. I love the insights I get from Lookers, the data importation feature from Supermetrics and the one stop dashboard from Ryꮓe AI.But these merely offer suggestions Not really do anything, can anyone foresee any tool that actually takes decisions in the future?

r/whatisit Bananafish1313

Setter next to the plugs in the kitchen area of my studio apartment in NL

Hey does anyone know what this is called and how it works? As the title says it is next to the two main plugs in the kitchen area of my studio apartment. Recently they both stopped working, before I was keeping a fridge and a heating induction stove connected through a power strip. Would love to know if it is because of this? The power strip and appliances worked when plugged in somewhere else. Maybe it is for the vent though? I cannot see anything changing when switching the different numbers though so idk.

r/fakehistoryporn LTC-trader

Moments before Adolf Hitler exits 1936 Berlin Olympics to avoid congratulating Jesse Owens, who won 4 gold medals

r/StableDiffusion bjorn_89

Video Eye Gaze Correction

Hello there,

I have some videos of a person reading a teleprompter, so there is no eye contact with the camera. Do you know any comfyui workflow that gets a video as input and fix the gaze of the subject in order to have such eye contact?

r/whatisit Responsible-Heart104

Make and model of this motorcycle

there doesn't seem to be any kind of labels on it anywhere so I have no idea how to find out the branding. I need a new rip cord for it because my husband pulled it off.

Also I don't know if this is the right sub please suggest another if not.

r/ClaudeAI AmesCG

Do "projects" eat extra tokens or am I hitting the usage-rate problems people have documented?

Hi all, I'm using Claude to talk me through career advice, with multiple chats clustered in a "project" that shares some basic documents. It's really impressive -- each chat instance clearly knows about the other and will remind me of tasks that had been recommended days ago by a different instance.

However, I've noticed that lately it truly guzzles tokens. I asked a single query in the project this morning (using Opus to finalize a document draft) and, just like that, 29% of my session cap is gone. Subsequent queries don't jump linearly, oddly enough. A clarifying question might add 3%.

For what it's worth I'm on the Pro Plan.

Is this a known issue with projects, resulting from all of the context they need to load, or related to the usage cap problems we've been seeing elsewhere?

r/TwoSentenceHorror wht_8

I don't like you' I whispered to the man but he didn't respond.

His severed head continued swinging from the branch while the blood dripped and painted his torso to my favourite hue

r/homeassistant rolandzeiner

Tankstellen Card für Österreich

Ich hab eine Custom Integration für Home Assistant gebaut, die österreichische Spritpreise über die offizielle E-Control API (spritpreisrechner.at) abfragt und direkt im Dashboard anzeigt (ähnlich Tankerkönig-Integration für Deutschland):

  • Zeigt die 5 günstigsten Tankstellen in deiner Umgebung für Diesel, Super 95 und CNG
  • Setup komplett über die HA UI – Standort per Map-Picker, Kraftstoffart auswählen, fertig
  • Custom Lovelace Card mit Tabs pro Kraftstoff, aufklappbaren Öffnungszeiten und Google Maps Links
r/AI_Agents Ozzie-obj

I got sick of AI that only talks. I built Temple: a local OS agent with actual hands for Linux and Windows and Macos (Beta)

Like a lot of you, I spend my time trying to automate my machine. I used Claude and antigravity for some time, but I found them limited. They are restricted, and they treat Linux like its not needed temple can run sudo commands

The Problem
Corporate AI tools are built to talk, not to act. They give you a list of instructions and expect you to do the manual labor. I got tired of copy and pasting terminal commands only to find out the AI made a mistake. Debugging a bot that is supposed to save you time is frustrating. I wanted an agent that could actually touch the kernel.

What is Temple?
I spent the last few weeks building an orchestration engine. Temple is a system level agent that lives on your machine. It does not just chat. It has hands. You tell it what you want, and it executes the work directly in your terminal and your files.

The main differences:

1. Terminal Native Execution
It has a built in node pty terminal. If you ask it to set up a project, it does not give you a tutorial. It runs the commands, navigates the directory, and starts the server natively. It does not wait for you to do the work.

2. Autonomous Failure Correction
If Temple runs a command and gets an error, it reads the stderr, catches its own mistake, and patches the code without you asking. It reads its own blood to find the fix.

3. Surgical File Editing
Instead of rewriting huge files and breaking the structure, it uses targeted tools to read and edit specific line ranges. This makes it fast and safe for large projects.

4. Built for the Outliers
I developed and optimized this on a Dell Precision T3400 from 2007 with Kubuntu 25.10 because I had problems with my father. If it is fast on a Core 2 Quad and a mechanical drive, it will be super good on your pc.

Status
I am a 14 year old developer 15 soon. I am the creator of RoCode (4000 users 200$ mrr). Temple is my flagship.

I Need Your Feedback
I am launching the public beta today. It is free to try with a 10 message daily limit to protect my infrastructure and my bank account. but for only 7.99$ you can get 40 messages per day

➡️ Check it out in the comments

Let me know if you like it or HATE it. I am watching the logs and I will patch any bugs I see tonight.

r/funny Lightnin-Bug

UPS Trucks Breeding?

r/whatisit Past-Armadillo-4363

Strange Sounds

I have been hearing some sounds from past few weeks. It was like I was being called by someone sometimes it's a familiar voice or sometimes just random voices. Everytime I went to check there's no one there. sometimes it's random noises like a metal peice drops in my room, someone walking in living room at night, a weird shadow at corner. Is it normal or am I just overthinking (It might sound unreal but it is real I promise)

r/ChatGPT Signal_Procedure4607

transcription/voice mode is currently broken PSA/FYI

I spent a sexy amount of 20 mins this morning trying to use the voice mode feature to dictate. I hit send, it keeps saying "transcribing". I've been using this all week and I've never seen this hang, maybe other people did, it wasnt my experience. It worked well, until today.

Thats ok - products break. But GPT didnt tell me it was broken. It kept asking me to try again. So monkey me did, and wasted even more time.

Thing is if you check here https://status.openai.com/ it shows voice mode is partially broken. Im confused why GPT is not attached to this status. It woudve been useful to tell me its broken, and not ask me to try again.

r/TwoSentenceHorror Large_Carob_7599

"Legs and Arms, they are limbs," my mother taught me

I felt betrayed when a stitch popped open, and I had none of them.

r/whatisit mypeepeehurting

Truck decal

I’ve seen probably a dozen in the last two days and I can’t prompt google well enough to figure it out, any help?

r/ChatGPT Think-Score243

Do clients trust AI chat systems as much as real humans in sales?

Seeing more AI tools automating chats, emails, and even sales conversations.

I actually tried this with my own SEO clients

I had around 10 clients and asked them to use chat for queries (available 24/7).

But over time, about half of them stopped following up and some moved to companies, where they could talk to real humans.

Feels like once people know it’s AI, trust drops , especially when money is involved.

Curious how others are handling this .

are AI chat systems working for your sales, or do clients still prefer human interaction?

r/funny Lublan

Man feeding corn to fish in a peculiar way.

r/ProgrammerHumor TobyWasBestSpiderMan

worstPartIsItsMyCode

r/AI_Agents ExoticYesterday8282

Curious how others handle long-term memory in AI agents?

I’ve been experimenting with an AI tool that separates memory by project. It seems helpful for keeping different tasks and notes organized.

Not sure if it’s just me curious how others handle this:

  • Do you find long-term memory in AI agents actually useful?
  • What are the limitations you’ve noticed?
  • Any tips for keeping multiple projects organized with AI agents?
r/homeassistant InterestingCold2399

Max suction isn’t the problem, dust re circulation is, how long does the Eufy S2 battery last on max suction?

Everyone asks about battery on max, but I think we’re asking the wrong question. Yeah sure, how long does the Eufy S2 battery last on max suction, but what actually matters is what happens to the dust while it’s cleaning.

Because higher suction usually means:

- stronger airflow

- more turbulence

- more particles lifted into the air before capture

And if filtration and airflow design aren’t good, you’re basically redistributing fine dust. For allergy people, this is brutal, you don’t notice immediately, but after a few runs, yeah.

Also after cleaning, emptying the bin is a second exposure wave. This is why I’m starting to care more about:

- sealed airflow paths

- docked auto-empty with bags

- minimal manual contact

S2’s whole bagged station approach seems like it’s targeting that, but I haven’t seen real user feedback on air quality impact. Anyone actually feel a difference vs traditional bin systems?

r/ClaudeAI Failcoach

Context/Token optimization

So I see a whole bunch of people explaining how they burn through tokens super fast.

I am on Max20 plan and use Claude Code all day long and I still have usage available on weekly reset.

Some of the things I do:

- every document gets converted to Markdown file before I use it

- every Excel file gets converted to cvs before I add it to conversation

- quick and short sessions (trying to stay below 150k tokens per session) - split big PRD into small PRDs

- never continue old conversation ... when I am done ... if I am not yet finished I do compacting so that I have summary and next time I start new fresh conversation (as far as I know Claude Code keeps KV cache for 5 minutes)

- deleted most MCPs and just use CLIs (like Supabase, GitHub, Vercel, ...) or create my own CLI tools to use with external tools

- i plan thing out in Claude.ai first then bring "strategic documentation" into Claude Code, have a skill for how I want PRDs to look like so that they are context/token efficient

- made my own system for memory. that is really just AI optimized wiki ... multiple small files, Mermaid diagrams, etc ... conected together with index file

- super short claude md file

- regular clean-up of stale documentation with a cleanup agent

nothing revolutionary, really ... just trying to keep it simple, effective and efficient

just FYI ... most of the time I am juggling between 10 to 15 projects ... and Max20 so far is more than enough for that

r/homeassistant slaegertyp

Shelly Plug S Gen3 connection lost. Quality problem?

I just bought three Shelly Plug S Gen3. I configured one, and it is dropping connections all the time. It drops Wi-Fi Direct and the regular WLAN connection.

I installed the latest firmware.

I was very enthusiastic when I discovered the Shelly products. Finally, a product I can connect to my MQTT and not depend on some obscure internet hub in Asia to get working.

Is this a known issue with the product? I was planning to use Shelly products in my house, but the quality does not convince me yet.

How do I get the plug online? Unplug it? This is not feasible since it is in a challenging-to-reach location.

Not good.

r/ChatGPT onfleek404

OK might not be the best at least I didn't add them grillz

r/SideProject Alternative_Bar_6426

I’m a 3rd-year Software Engineering student and I built LookUp: A Chrome extension that auto-captures page context to turn messy tabs into an AI learning hub.

I built LookUp to bridge the gap between "browsing" and actually "mastering" knowledge. As a student, I was tired of having dozens of tabs open and losing track of what I was learning.

The Problem: Manual screenshots and messy bookmarks are a pain. LookUp automatically captures the screen and extracts the context for you, handling the "capturing" so you can focus on the "learning."

Key Features:

  • Smart Folders: Auto-categorizes open tabs into logical contexts.
  • AI Explanations: Get instant breakdowns of complex topics on any page.
  • Integrated Note Hub: Save insights directly from your tabs into organized folders.
  • Deep-Dive Chat: Chat with your collected knowledge to summarize or connect the dots.

It's lightweight and designed to stay out of the way until you need it.

Check it out here: Chrome Web Store

Learn more about the project: Landing Page

r/Jokes bespokeagent

What's a dog walker do at the end of a long shift?

gives his dogs a rest.

r/SideProject iammidhul

I built a free site with tools I wish existed (no login, no ads)

I kept running into the same annoying problem — needing small tools but never remembering where to find them.

So I started building my own collection.

What started as “just for me” turned into something I use almost daily now.

Some tools I added:

Time Zone Meeting Scheduler (finally solved cross-timezone headaches)

Subscription Auditor (this one hurt 😅)

Corporate Meeting Cost Timer (also painful to watch)

Road Trip Cost Calculator

PDF Password Remover

No login. No ads. No clutter. Just tools that work instantly.

Still early — I’m building this based on what people actually need.

If you had to add ONE tool to this, what would it be?

👉 https://convertwithmi.com/

r/ClaudeAI dopinglab

I got 100,000 views with a Claude-built app… and $0 revenue. What am I doing wrong?

Hey everyone,

I’m a solo developer from South Korea, and I built a real-time conflict monitoring service using Claude.

A few days ago, I shared it here and got ~100,000 views. The number of users doubled, which was honestly surprising.

But… conversions are basically zero.

So I’m trying to understand what’s not working.

What I built (and how Claude was used)

The product is a dashboard that tracks global conflicts and tries to answer:

“How does this affect me, based on where I live?”

Data pipeline

  • 100+ sources (news RSS, Telegram channels)
  • Collected via scheduled workers
  • Stored and processed in backend

Where Claude comes in

I used Claude for:

  • Classifying incoming content (conflict vs non-conflict, topic, country, severity)
  • Handling edge cases (e.g. “battle” in sports vs actual conflict)
  • Generating structured summaries → What happened / Why it matters / When
  • Iterating on prompts to reduce false positives

Biggest challenge was getting consistent classification across noisy sources.

The problem I hit

Even with all this:

  • People visit
  • They explore a bit
  • Then they leave

Even returning users don’t convert to paid plans.

My current hypothesis

The product is too complex to grasp quickly

Right now it shows:

  • conflict clusters
  • severity scores
  • reliability weighting
  • impact metrics

But the reaction seems to be:

“Interesting… but I’ll check this later when I need it”

Which means: not something people feel they need daily

What I’m testing now

1. Simpler format (newsletter-first)

  • Weekly + daily summaries
  • Much easier to consume than dashboard

2. Clear “personal impact”

  • Not just global data
  • But:
    • economy
    • energy
    • trade impact based on your country

3. Reducing cognitive load

  • Less raw data
  • More interpretation

Free access

The project is free to try: https://www.wewantpeace.live

(No login required for basic features, paid plan only unlocks deeper data + alerts)

What I’d really like feedback on

  • Is this a “check occasionally” tool or something that could be daily-use?
  • What would make this worth paying for?
  • Is the problem the format or the core idea itself?

I’m especially curious how others here use Claude for classification pipelines like this, and whether you’ve run into similar “looks cool but doesn’t convert” problems.

Would really appreciate any honest feedback.

r/LocalLLaMA ArugulaAnnual1765

Anyone tried running Bonsai on a smartphone yet?

Supposedly this model is really small and capable of being run on edge hardware.

Has anyone tried running it on a smartphone yet?

I have a galaxy s25 ultra with 12gb ram and the snapdragon 8 elite soc, do you think it would be capable of running the model at a decent speed?

r/Jokes FabulousKitchen5831

Just had a policeman knock on my door and say, he’d had reports that my dog is chasing a man on a bike

I said “that IS strange because, my dog doesn’t own a bike”

r/ChatGPT Rangerazuladio

How to create tasks/schedule now?

I cannot find a way to crete schedule, recurring tasks anymore. When I ask to create a task, it keeps saying that I need to create myself via a menu that don't exist and it cannot create via the chat.

No way to find how to add a new task via the settings, but I can edit the existing ones... Did I miss something ?

r/LocalLLaMA Quiet_Dasy

Hunting for the "Speed King": Best <14B MoE for Sentence Rewriting (HellaSwag focus)

. Example SmallThinker-21BA3B (21B total, 3B activated) Moonlight-16B-A3B (16B total, 3B activated) Ling-lite-1.5-2507 (16.8B total, 2.75B activated)

Hi everyone, I’m on a quest to find the absolute speed king in the Mixture of Experts (MoE) category, specifically models under the 9B–14B parameter range. The Use Case I’m building a local sentence rewriter. The goal is to take a "bag of words" and transform them into syntactically and semantically perfect sentences. * Example: Input: cat, fish, lake ➔ Output: The cat eats the fish by the lake. The Technical Setup * Hardware: Dual GPU (16GB VRAM total) running via Vulkan. * Platform: lm. * Constraint: I am not looking to do a custom fine-tune with Wikipedia data; I need something high-performing out of the box. The "HellaSwag" Problem In my experience, small models usually fall apart on HellaSwag (common sense reasoning). Most sub-14B models lose significant syntactical "intelligence" because they lack the training depth. I need a model that maintains elite HellaSwag scores while remaining an MoE for that snappy inference speed. L https://jamesyc.github.io/MoEspeedcalc/ What is the fastest English MoE under 14B that doesn't sacrifice HellaSwag reasoning for speed? If it’s on that leaderboard and runs like lightning on Ollama/Vulkan, I want to hear about it. Looking forward to your suggestions

r/ClaudeCode Routine-Direction193

I fixed my usage limits bugs. Asking Claude to fix it...

https://preview.redd.it/thnbku7s7ssg1.png?width=960&format=png&auto=webp&s=6b4361fd47c489c9d4631d171bae4cb62236f481

All you need to do is revert to 2.1.74.

Go in vscode. Uninstall claude code extension if it's installed

install claude code extension in 2.1.73. Then ask it to revert the cli version to 2.1.74.

Important part : ask it to delete all files who can auto upgrade claude to new versions

Also make sure NPM can't update your claude.

You know it has worked when claude code tells you you need to do claude docteur and it can update itself.

No more limit usage bug.

kudos to the first guy who posted this on reddit. worked for me.

Opus is still lobotomized though

r/whatisit Average_Joe69

What is this on the edge of my pop-tart I got from a vending machine.

It’s a brown sugar cinnamon flavor, I get them at work all the time but they’ve never had this. I can’t tell if it’s just burnt or if it’s gone bad, or if something got into it. Help me out :)

r/mildlyinteresting magzter47

My first marbled sweet potato

r/LocalLLM Suitable-Song-302

TurboQuant.cpp — 1-bit KV cache with zero quality loss, verified on 35B MoE

Pure C inference engine implementing the TurboQuant paper (ICLR 2026). Built from scratch, not a llama.cpp fork.

What it does: Compresses KV cache keys to 1 bit using randomized Hadamard transform + sign hashing. The output is byte-identical to the uncompressed baseline.

Verified results:

Qwen3.5-35B-A3B MoE (IQ2_XXS GGUF, 16GB Mac): baseline: "The capital of France is Paris." 1-bit KV: "The capital of France is Paris." ← same output Gemma 3 4B (TQM, perplexity 101 tokens): FP16 KV: PPL = 35.99 1-bit K + Q4 V: PPL = 36.00 (+0.03%) 

1-bit attention cosine = 0.634, matching the information-theoretic limit of 2/pi. Formal unbiasedness verified at < 0.2% relative bias over 100K random vector pairs.

What's in the repo:

  • 27K lines of C/Metal, zero external dependencies
  • GGUF direct loading (Q8_0, Q4_K_M, IQ2_XXS verified)
  • MoE support (256 experts, top-8, shared expert)
  • 1-bit weight quantization (8.4x compression, zero quality loss on 4B)
  • Metal GPU backend (Apple Silicon), CUDA/Vulkan/ROCm compile targets
  • 32 test suites, ASan clean
  • Perplexity measurement, activation profiling, codebook calibration tools

Honest limitations:

  • CPU inference only for now (Metal MoE dispatch is WIP)
  • 35B at ~1-4 tok/s on M3 16GB (memory bandwidth bound)
  • IQ2_XXS (2-bit weights) limits quality on complex reasoning — that's the weight quantization, not the KV compression
  • Tested on Qwen3.5 and Gemma 3 only (3 architectures)

The algorithm (from the paper):

Keys: normalize -> RHT -> Lloyd-Max codebook -> QJL sign hash 1-bit: signs only -> attention via XOR + popcount

Values: per-block Q4 or Q2 quantization

The paper proves standard quantizers introduce systematic bias in inner product estimation. RHT + QJL correction makes it provably unbiased.

https://github.com/quantumaikr/TurboQuant.cpp

Paper: https://arxiv.org/abs/2504.19874

Happy to answer questions about the implementation or the algorithm.

r/photoshop Birdseye5115

Is there a simple way to export a single layer?

I find my self needing to process some bits in another program. I'll stamp up, save out a tif. but I don't always want the whole image. Is there a simple and straightforward way to export a single layer as a TIF?

r/StableDiffusion AsstronautHistorian

Anyone Else Having Hard Time Installing LTXVReferenceAudio Node?

It appears to be a core-comfy node so I tried updating ComfyUI with no luck. It also seems to think that the node is from a "newer" version of Comfy when in reality it's from an older version.

r/Futurology morecowbell1988

How vulnerable is modern infrastructure to a Carrington-level solar storm?

This was originally written in a more cinematic style to draw readers into a broader discussion about infrastructure vulnerability tied to my PhD research.

I tightened the language and removed the AI-generated imagery across the site. Sources are at the bottom. Carrington-level events are often estimated on the order of once every ~500 years. The Grounded offers $100 for any verified factual error identified by members.

Curious where people think the real risks are versus what’s overblown.

r/PhotoshopRequest Eerie_Carlisle

Remove woman in back

Please remove the woman in the background.

r/personalfinance ooeemusic

John Hancock Website - Contributions

Hello,

Does anyone know where/if on the John Hancock retirement plan website it is possible to view each individual contribution? I can find year to date contributions, and also total contribution types for the total balance, but I cannot for the life of me find individual contributions (e.g. if I contribute 5% per pay, there should be two contribution amounts per month).

Am I totally missing it, or is it not possible to view that info on the website?

Thanks, all!

r/meme Amazing_Parsley8381

What kind of thumbnail is this 🙏🙏🙏🙏

r/VEO3 GasDelicious5453

Let's look at the performance of VEO3 in the Vietnam War.

r/personalfinance Ilysumo55

Nineteen year old in college trying to build Capital & Credit. (United States)

I'm nineteen and currently in college, I come from an extremely impoverished family and I'm the first one making it in college and trying to get some money to get my family out of the mud. I network with some of the finance people on my campus and they suggested I started using a credit card and paying it off (My family is very anti credit so this was a shock to me.)

I also want to start investing into ETF as well as Stocks so any general knowledge or where to go to educate myself on that would be much appreciated.

Basically, any knowledge you wish you had at my age I would greatly appreciate and will put into practice.

Cheers!

r/SideProject StopMemorizing

5 projects, €2,056 spent, 1 paying customer. I'm fine.

Since August 2024 I spent €2,056 on ads, tools, Apple developer fee, and domains. Revenue: one paying customer.

Mac app - flopped. Chrome extension - flopped, but got lots of emails offering to boost my rating for $50. SaaS - 30 free users, 1 paid, and I spend more time on X than in my code editor. Not proud of that.

So obviously I started 2 more projects.

But this time not alone. One with a friend who handles all the vision and business stuff I hate. One with my wife, who finds all the community grinding genuinely exciting. I just build. She grinds. Fair deal.

The frustrating part? Everything in indie hacking is slow. SEO takes months. Trust takes months. Growth takes months.

But competition is getting faster. AI lets people ship in days. And sometimes it doesn't matter who built it better - it matters who showed up first.

So here I am. 5 projects, €2,056 spent, 1 paying customer. Still trying different approaches because I don't know what else to do.

Maybe that's enough. Has to be.

Found the perfect summary of my current situation: https://youtu.be/PdCoadVSfXg?t=174

r/meme Linafplus

Tough competition

r/AI_Agents Embarrassed-Koala378

I think my boss and I are just proxies for two AIs talking to each other

I just realized something weird about my workflow today.

I thought I was having a normal back-and-forth with my boss about a proposal.

I sent him a structured draft.

A while later, he replied with a cleaner, sharper version — better wording, better logic, even added a few points I missed.

My first thought:

“Wow, he’s on fire today.”

Turns out… he just pasted my message into ChatGPT.

And here’s the funny part — I did the exact same thing on my side.

I took his reply, fed it into my own AI, refined it, and sent it back.

So the actual loop looked like this:

Me → AI → Boss → AI → Me → AI

At some point it hit me:

Are we even talking to each other anymore?

Or is this just two AIs negotiating through us?

It honestly feels like we’ve become the API layer between two models.

The output keeps getting better, but neither of us is really “thinking” in the process.

If this keeps going, I feel like our future job description will be:

“Forward messages between AIs and pretend you’re involved.”

r/Futurology Kahootah

Will black markets still exist in the 2040s?

There are news of flock cameras, internet age verification, palantir, etc

Is it likely that artificial smartness surveillance will kill off black markets in the future? I see this as an important question as crime and black markets are usually the first to adapt to surveillance

r/whatisit Meeple_Peeple

White Bar On Book Cover

Tried googling and didn't see any other mention of something like this, is there supposed to be something there, or was it just printed off center?

Didn't know where else to post it so..

r/LocalLLaMA Quiet_Dasy

Running LLM on one machine and TTS on another via lm link?

-PC Setup: Running LLM on one machine and TTSon another via lm link?

The Hardware:

PC 1 (Host): Running LM Studio + the LLM (qwen/qwen3.5-9b).

PC 2 (Client): Running tts

want the text generated by the LLM on PC 1 to be sent over the link to PC 2 so the TTS engine can read it out in real-time

r/PhotoshopRequest Traditional-Fig3621

Please enhance the quality of the pic. Make sure we (me and my daughter) feature prominently in the centre. Remove the 2 people in the background (the bald guy to the left) and the guy wearing sunglasses on the right). Will tip 5$.

r/personalfinance International_Rush34

Do I take this job offer, and if so, how do I invest?

Hello, friends!

I have a job offer and need to determine the financial impact, and if so, how I invest:

Opportunity:

  • current position: $100K salary, pension with 14 years service
  • new opportunity: $200K salary, 401K

Balance sheet right now:

  • assets
    • contributed towards pension to date: $200K
    • other investments (IRA, 403b): $80K
    • liquid savings: $50K
  • liabilities: “typical” debt—
    • $200K remaining on $300K home mortgage for home purchased in 2019
    • $10K remaining on $30K car purchased in 2024

Life status:

  • M(41) with spouse and two elementary-aged kids
  • We are living a comfortable life. While I always “want” more money, we don’t *need* more money for day-to-day living.
  • Wife recently returned to work after pausing her career re: kiddos and makes around $40K annually— very helpful to our bottom line but not part of our longer term strategy. (For the record: she is also a saint for all the sacrifices she’s made for our family.)
  • We live in USA in region where cost of living is, thankfully, manageable
  • Both my and my wife’s parents are retired with pensions and no debt. Possible we will see some transfer of wealth upon their deaths but we want to plan as though we will not.

Options:

A) Keep the current job at $100K, continue to contribute to pension fund. Current projection is around $8K / month in pension benefits starting at age 58 in 2043.

B) Take the new gig and increased $200K salary and direct increased salary into investments. Keep contributions in pension fund. I would be eligible for partial benefits of $1,400 / month at age 62 or $2,300 / month at age 67.

C) Take the new gig, direct the extra salary into investments. And cash out the $200K in contributions I’ve put into the pension fund to date and invest them myself.

Question: what provides the best long term security— cash in hand now that I control (in increased salary and ownership of my investments), or the slow and steady perpetual pension fund?

Fears keeping me up at night:

  • Pension: I realize the ’security’ pension funds provide. Mine is a stable-ish one (government), but I’m always fearful it will be drained before we get to use it and / or relying on it exposes my kids to risk in the event of my unexpected death, etc.
  • Private investment: I’m also worried that I am too far into my pension service to pivot at this point— e.g. the ‘golden handcuffs’ that what I get out doesn’t match the time I’ve lost to invest it.

Thank you, thank you, everyone for any insight here.

r/personalfinance Icy_Duck_5477

Procrastinated on taxes and now I'm two weeks out with RSUs, filing jointly for the first time and freelance income

I knew this year was going to be complicated and I think that's why I kept avoiding it. Got married in October, first time filing jointly. RSUs vested for the first time at my current job. Sold some investments at a gain. Have a small amount of freelance income on top of my W2.

Every time I sit down to start I just close the laptop. I'll get to the first screen and panic and tell myself I'll figure it out tomorrow. I have no idea where to even start with filing jointly for the first time, the RSU withholding situation, whether I even reported the investments right in previous years. It just spirals.

Now it's two weeks out and I'm out of runway to keep avoiding this. I know I need help but I also don't know what kind. A full CPA firm feels like overkill but I don't trust basic software to handle all of this without missing something.

I'm in California if that matters. My income is around $145k base plus the RSUs which vested at about $38k. Wife makes $72k. Investment gains were maybe $6k total. Freelance brought in another $4k.

Has anyone here been in this situation this late in the season and found something that worked without breaking the bank? I'm kind of spiraling and just need to know what my actual options are at this point.

I know I screwed up by waiting.

Im just trying to figure out the best way forward from here

r/LocalLLM Thump604

MLX Inference: Where Things Stand in April 2026

Mac Studio M2 Ultra, 128 GB unified memory

I run large models locally on an M2 Ultra for coding agent workloads. A lot has changed over the last months. Here are the numbers and what happened.

Generation Speed Across Four Models

Decode throughput (tok/s) at each KV cache depth. 256 output tokens per run.

Model Quant 4K 16K 32K 64K 128K Qwen3.5-27B (dense) 8-bit 20.2 19.1 17.9 16.4 13.1 Qwen3.5-35B-A3B (MoE) 8-bit 71.8 65.8 61.1 53.5 41.9 Nemotron Super 120B 5-bit 36.4 34.8 33.5 31.2 28.4 Qwen3.5-122B-A10B (MoE) 5-bit 40.6 37.4 34.2 29.4 23.1

The 35B MoE hits 72 tok/s at short context because only 3B of its 35B parameters are active per token. The dense 27B is the slowest despite being the smallest because all 27B parameters fire for every token. Nemotron Super 120B barely degrades with context (14% drop from 4K to 64K) because 80 of its 88 layers are Mamba-2, which has constant cost per token.

Feature Speedups: MTP and SpecPrefill

Two features make a big difference on top of baseline generation:

MTP (Multi-Token Prediction): Qwen 3.5 models have a built-in draft head that predicts the next token in parallel. With probabilistic acceptance at 90% rate, the 122B goes from ~17 tok/s to 38.8 tok/s (2.3x). Server overhead is minimal: a short-prompt request through vllm-mlx generates at 39 tok/s, matching baseline.

SpecPrefill: For long prompts, a 2B draft model scores token importance via attention, then the target only prefills the top 20%. On the 122B at 128K context, TTFT drops from 19.3 minutes to 3.5 minutes (5.5x). Below 8K tokens the overhead is not worth it, so it only activates for long prompts.

Combined with continuous batching and prefix cache, the 122B serves coding agents interactively at context lengths that used to be completely impractical.

MLX vs. llama.cpp at Long Context

llama.cpp's flash attention kernel has been the reference point for Metal performance, and their split-K decode is excellent work. I benchmarked Qwen3.5-35B-A3B on both stacks to see where MLX stands. 512 tokens generated after filling the KV cache to each depth.

Context MLX 8-bit llama.cpp FA ON (5-bit) llama.cpp FA OFF 32K 60.8 54.85 36.45 64K 53.2 45.84 24.47 128K 42.7 34.48 13.73

The FA ON vs. FA OFF column shows how much llama.cpp's flash attention contributes: 1.5x at 32K up to 2.5x at 128K. That kernel is doing serious work.

What surprised me is that MLX is competitive. MLX already has a 2-pass split-K decode kernel (sdpa_vector_2pass) that dispatches up to 1024 threadgroups at 128K. Both frameworks are well optimized for Metal at this point.

A note on the quantization mismatch: the MLX model is 8-bit and the llama.cpp model is Q5_K_M (5-bit). I used what I had on hand. The point here is not a controlled head-to-head shootout between frameworks. It is a sanity check on the assumption that MLX falls far behind llama.cpp at long context, which it does not. A matched-quantization comparison would be useful but was not the focus.

Why Hybrid Architectures Change the Game

The models above are not standard transformers. Qwen 3.5 uses GatedDeltaNet layers (linear recurrence) for most of the network with standard attention for only 25% of layers. Nemotron Super uses Mamba-2 for 91% of layers. The recurrent layers have fixed-size state that does not grow with context.

Model Attention layers 4K tok/s Drop at 64K Qwen3.5-35B-A3B 25% (10 of 40) 71.8 -25% Nemotron Super 120B 9% (8 of 88) 36.4 -14%

Fewer attention layers means less KV cache to scan per token and less degradation at long context. This is the architectural direction that makes extended context practical on consumer hardware.

What Shipped in Two Months

The MLX ecosystem has three layers and all of them moved fast.

MLX core: Thread safety overhaul (per-thread Metal streams, smart pointers) fixed production crashes. Split-K quantized matmul for faster decode. CUDA backend in progress. M5 tuning tables already merged.

mlx-lm: 10+ new architectures including Qwen 3.5, Nemotron Super, DeepSeek V3 MLA, and GLM5. GDN memory leak fix. Batch generation refactor with hybrid cache support. Prefix caching in the built-in server.

vllm-mlx: Went from v0.2.5 to v0.2.7 with tool calling (12 parsers), embeddings API, reasoning support, continuous batching, prefix cache, and MTP speculative decoding.

r/ProductHunters AreaArtistic2108

Created The Good News App because traditional news felt draining. I wanted a space that informs the people while protecting their heart—focusing on progress and kindness. If you want to stay updated while keeping your peace, give it a try!

Google Play: The Good News App

r/LocalLLaMA 1000_bucks_a_month

PSA: PrismML Bonsai-8B (Q1_0_g128) produces garbage output on CPU -- GPU appears to be required

I was excited to try the new Bonsai 1-bit models from PrismML, which launched March 31. Built their llama.cpp fork from source on Windows 11, loaded the Bonsai-8B GGUF, and got... nothing coherent.

Setup:

- Windows 11, x86_64, 16 threads, AVX2 + FMA

- No dedicated GPU (CPU-only inference)

- PrismML llama.cpp fork, build b8194-1179bfc82, MSVC 19.50

- Model: Bonsai-8B.gguf (SHA256: EAD25897...verified, not corrupted)

The model loads fine. Architecture is recognized as qwen3, Q1_0_g128 quant type is detected, AVX2 flags are all green. But actual output is garbage at ~1 tok/s:

Prompt: "What is the capital of France?"

Output: "\( . , 1 ge"

Multi-threaded is equally broken:

"., ,.... in't. the eachs the- ul"...,. the above in//,5 Noneen0"

Tested both llama-cli and llama-server. Single-threaded and multi-threaded. Same garbage every time.

Looking at PrismML's published benchmarks, every single number is from GPU runs (RTX 4090, RTX 3060, M4 Pro MLX). There is not a single CPU benchmark anywhere. The Q1_0_g128 dequantization kernel appears to simply not work on x86 CPU.

The frustrating part: there is no way to report this. Their llama.cpp fork has GitHub Issues disabled. HuggingFace discussions are disabled on all their model repos. No obvious contact channel on prismml.com.

So this is both a bug report and a warning: if you do not have an NVIDIA GPU or Apple Silicon, Bonsai models do not work as of today. The "runs on CPU" promise implied by the 1-bit pitch does not hold.

If anyone from PrismML reads this: please either fix the CPU codepath or document that GPU is required. And please enable a bug reporting channel somewhere.

Important: File hash verified, build is clean, not a user error. Happy to provide full server logs if a dev reaches out.

r/TheWayWeWere dmode112378

Both sets of grandparents at my parents’ wedding in 1972

Hard to believe all of them were in their fifties here. My mom will be 76 in two months and looks way younger than all of them.

r/funny efunny2022

Restaurant things

r/ChatGPT JaredSanborn

I stopped using AI as a tool. Started using it as a partner.

Most people use AI like:
"Write this"
"Summarize this"

That’s level 1.

Level 2:
Think WITH AI.

What changed:
- Better decisions
- Better ideas
- Compounding improvement

But this only works if AI remembers you.

No memory = no partnership

Anyone else thinking this way?

r/ClaudeCode kugge0

Tired of new rate limits. Any alternative ?

Hi guys! I've been using Claude Code for more than a year now and recently I've been hitting limits nonstop. Despite having the highest max subscription.

I was wondering if I should buy another CC subscription, or switch to something else.

What's the best alternative to claude code with the highest rate limits rn ?

r/LocalLLaMA Sadikshk2511

Tried Zai’s GLM-5V-Turbo on some UI-heavy tasks, mixed early findings

I’ve been trying a few multimodal coding models lately for UI-ish work, and I spent a bit of time today messing around with GLM-5V-Turbo from Zai

Still early, so not trying to do some full review here. More just posting first impressions after throwing a few real-ish inputs at it instead of only looking at demo-style examples.

What I mainly wanted to test was whether it could actually do anything useful with visual input in a coding workflow.

Not just “describe this screenshot,” but stuff more like:

- UI screenshots

- rough mockup / layout images

- document-like pages

- some cluttered visual inputs that weren’t especially clean

My first impression is that it does seem a bit more comfortable with visual structure than a lot of coding models that still feel heavily text-first.

On some layout-heavy tasks, it picked up hierarchy / spacing / rough structure better than I expected. Not consistently, and definitely not in a “this solves it” way, but enough that it felt worth noting.

Right now I definitely wouldn’t put it in the “upload screenshot → done” category.

If anything, it feels more like a usable starting point than a reliable finisher.

What does seem interesting is the direction. It feels more relevant in workflows where the input is screenshots / mockups / docs / mixed visual context, not just plain code or text.

Also seems like GLM-5V-Turbo is being positioned more around tool / agent-style workflows, which honestly makes more sense to me than treating it like a standalone coding model. I’m less interested in whether it wins on a benchmark and more interested in whether it’s good enough to be useful inside a bigger loop.

So I guess my current take is:

- decent at some UI-ish visual tasks

- maybe more interesting as part of a workflow than on its own

Curious if anyone else here has pushed it harder.

Especially interested in comparisons against Claude / GPT-4o / Gemini for screenshot-to-code, front-end layout work, or general multimodal coding stuff.

r/personalfinance DisplayKnown5665

Anything to consider if paying off 6-year car loan super early?

I just bought a new car and put $30K down. Amount financed is $29K with a 6.5% interest rate for 6 years...bringing the total to $35K if I make payments as scheduled.

I plan to double or even triple my monthly payments regularly so that I'm not paying for the full 6 years. Plus, I figured I could always make some large payments right off the bat once I had more time to think things through.

Say I pay $5K or even $10K on the first monthly payment to bring down the balance quickly. Is there anything to consider? Does it affect credit score?

Similarly, my dad has offered to give me a personal loan so that I can just pay it off right now. Again...would that affect me negatively in any way? He just doesn't like the idea of me paying up to $6K in interest.

Looking back, I should have just put more down upfront, so lesson learned for next time. My dad was along with me, so not sure why he didn't offer to add to the down payment if he wanted to loan me some money.

r/SideProject Ejomiciano

Built a free gig marketplace for AV crew — looking for people to try it

Hey everyone — I built StageCrew, a gig marketplace specifically for freelance AV techs and stagehands. The idea is simple: you sign up, add your skills and availability, and get matched to gigs across 25+ US markets (Vegas, LA, NYC, Nashville, Chicago, Orlando, etc.).

Free for workers. No staffing agency markup.

It's live and working but still early — I'd genuinely love feedback from people who actually work in this space. What would make something like this useful to you? What's missing?

If you want to check it out: stagecrew.app

Not trying to spam — just a builder looking for honest input.

r/me_irl PeakPointFitness

me irl

r/ClaudeCode ArticleKey9005

Selling 6 months of claude pro.

I got 6 months of claude pro from my uni, not of much use for me.

Its a claude pro promotion code which can be applied to any free account.

Does anyone wants to buy it?

r/ChatGPT Dreaksfrendford

ChatGPT opened on its own while it was running in the background and showed a fully typed message I NEVER wrote

So something really weird happened and I want a technical explanation. I logged out of ChatGPT, then tapped continue with google to log back in. While that was processing, I switched to Reddit and let ChatGPT run in the background A few seconds later, ChatGPT suddenly came to the foreground by itself( I still wasn't logged in ). When I looked at it, there was already a full message typed in the input box, that too in Hindi (something I never do). In the background , two tabs were running. One with a ChatGPT logo on it, and one with the reddit logo. And both were showing 'ChatGPT' when I clicked on them and both showed a ChatGPT interface. I didn't type Hindi, didn't paste anything. It wasn't live typing, it was already fully there. Has anyone seen ChatGPT inbox auto fill for you? The text that I saw said "Remove all the people in the background, keep the rest of the photo exactly as it is. First, if I haven't uploaded a photo yet, ask me for one." Seems like an AI written suggestion to me. I use Gboard with English and Spanish keyboards added.

I've attached a 1 minute raw screen recording.

r/whatisit Ho_Tay_Banky

Hydraulic large SS cylinder with grinding spikes on truck

What is the purpose/function of this truck? It has a hydraulic large stainless steal cylinder with grinding spikes. Cylinder is connected to a hydraulic arm which (? I think ?) allows the cylinder to come down to ground level behind the truck. What is the purpose/function of this truck? AND Why is the cylinder stainless and not just plain iron/metal ? Sorry for the pic w/glare, but I was trying to get a side view of the truck. Warning ⚠️ : I would hate to piss off the driver/operator and have him smash my car w/that thing. LOL

r/ClaudeCode Shorty52249

I guess I'm just lucky at this point, there are no other explanations.

https://preview.redd.it/ftugiiidgssg1.png?width=744&format=png&auto=webp&s=bd175ebd5737ad71f3c2e0f5c3c86aa5c7682aa2

Literally tried using as much as possible for one entire week pushed more than fifty thousand lines of code but still was unable to reach even fifty percent and today you can see my model resets in one hour and twenty-seven minutes. I don't know why everyone is complaining, I guess a very short number of users are facing that problem or I'm just lucky.

https://preview.redd.it/r3g6uegvgssg1.png?width=472&format=png&auto=webp&s=12a98953a072bfb79599451dc9efb9a9c1bf4401

r/AI_Agents JaredSanborn

I run workflows using multiple AI agents. Here's what surprised me.

Biggest insight:

Delegation is harder than intelligence.

AI is already smart enough.

The real challenge:
- Assigning the right task
- Coordinating outputs
- Structuring the system

Also:
Agents without memory = useless

Agents with experience = powerful

Most "AI agents" today are just chatbots.

Real ones:
- Delegate
- Learn
- Improve over time

Anyone else experimenting with this?

r/ClaudeCode IslamGamal8

Model output quality difference

Now that the code is out maybe someone can find an explanation to this. Unpopular opinion but planning output quality with opus in Antigravity is SIGNIFICANTLY BETTER than in CC. Every time im designing a feature i give the same prompt to 5.4 xhigh, opus high effort CC and opus planning in agy.

The difference is night and day, always. And it’s not even close. Gpt always the worst. Opus in CC slightly better but man agy artifact is just something else. I just want to understand if google is doing something really good or anthropic doing something very wrong?

r/painting vharishankar

Mountain road

r/AI_Agents Wise-Stress-732

Advice needed for a AI chat bot side hustle

Hey everyone,

I've just started freelancing — building custom AI chatbots for businesses and Web3 projects. Specifically lead gen bots that qualify and categorize leads (hot/warm/cold) + customer support automation, delivered as a website widget.

Before I go too deep I wanted to get some honest perspective from people who've actually done this or are doing it:

  1. Is this actually worth pursuing in 2026 or is the market already too saturated?
  2. What are the most common mistakes people make early on that kill the business before it starts?
  3. Realistically — how long does it take to land the first paying client?
  4. What's a realistic monthly income ceiling for a solo operator doing this?

Not looking for hype — just honest answers from people with real experience. Good or bad, I want to know what I'm actually walking into.

Appreciate any advice.

r/ClaudeAI buwilliams

New Claude Code Plugin: Forge

Have you had the experience of asking Claude Code to accomplish an ambitious project, and it says it succeeded, but actually doesn't follow your instructions? Like maybe 30-60% was completed? Yeah, me too. I fixed it through a new Claude Code plugin.

Forge does three things:
1. Solves context rot
2. Solves the echo chamber
3. Solves workflow

https://github.com/buwilliams/forge/

Enjoy!

p.s. See the X thread for more info! https://x.com/BuddyIterate/status/2039688532722159831?s=20

r/whatisit Blackfyre4007

Please help settle a debate! What colour is this, baby blue or grey?

Its a £5 bett. I think it's grey whilst my dad says baby blue. What do you think?

r/SideProject Prize-Log6966

Built an indie Delighted alternative because small teams are getting crushed on pricing

Qualtrics is sunsetting Delighted on June 30, 2026, so a lot of teams are going to need a replacement.

I’ve been building **Thrilled**, an indie customer feedback and retention tool for small businesses that don’t want enterprise pricing, seat-based pricing, or a pile of separate tools glued together.

A few things I wanted to do differently:

- **No seat limits**

- **No per-user fees**

- **No “add another teammate, pay more”**

- The free tier is the actual product, not a fake trial

## What it does

- AI can build your survey or retention flow from a plain-English description

- **118 templates** if you want to start from something proven

- **5 distinct AI interviewer personalities** that follow up in real conversation to uncover why someone is unhappy, likely to churn, price sensitive, confused, and more

- Cancel flows that can offer **pause, discount, plan switch, or escalation** based on what the AI finds

- Dashboard with **NPS trends, AI theme tagging, sentiment analysis, health scores, and Slack alerts**

## Built-in social proof tools

I also wanted to avoid the “pay for three separate products” problem, so it includes:

- Configurable **testimonial wall widgets**

- Configurable **social proof badge widgets**

- Both tied directly to **real collected responses**

## Pricing

### Free tier

- 1 survey

- 1 retention flow

- 100 responses/month

- 50 AI conversations/month

- Full dashboard

- No credit card

- No expiration

- **No seat limits**

### Paid

- **$19/mo or $39/mo**

- More volume

- Integrations

- Custom domains

- More usage

Still **not priced by seat**.

## Security

I know people are skeptical of new software, especially AI-assisted software, so security was a major focus.

Here’s what’s running in production right now:

- **TLS everywhere**

- **HSTS with preload**

- **Strict Content Security Policy**

- **Org-scoped isolation**

- **HMAC webhook verification**

- **Signed session tokens**

- **Full audit trail**

- **Bot and spam protection**

- **Secure session handling**

- **CSRF protection**

- **AI prompt hardening**

If you’re migrating off Delighted, I also added **one-click CSV import**.

I built this because I know a lot of people are struggling right now and small businesses get priced out of software constantly. Please share with your small business owner friends who might benefit from a cheap retention tool.

Would genuinely love your feedback and happy to share my experience building this with Claude Code.

**[getthrilled.io](https://getthrilled.io)\*\*

r/meme Safe-Afternoon-6695

Me spanish

r/TwoSentenceHorror anuarwriter

Prying his jaws open with a mouth gag, I met the old man's yellowed, crooked teeth.

I needed to extract them all and implant the dog fangs in a single night.

r/ChatGPT JaredSanborn

Your AI forgets everything about you every time. That’s the real problem.

Everyone talks about better models.

But your AI still:
- Doesn’t know your business
- Doesn’t remember past work
- Starts from zero every time

That’s the bottleneck.

I’ve been testing persistent memory for months.

Difference:
- Better outputs instantly
- No repeating yourself
- AI actually improves over time

The model matters less than the context it has about you.

Curious — is memory overrated or are we ignoring the real problem?

r/personalfinance AdLazy8828

Renting vs Buying: Am I missing something?

I’m a 28M who makes 90K per year.

I don’t have any debt, 15% of my income goes to 401k, I also do weekly contributions to 529 & Roth IRA, and my brokerage.

I can afford a home, but I’d rather rent considering the cost difference (especially as a single man).

For example, I did a calculation of my 30 year mortgage would be around $2200.. However, recently built apartments in my area are around $1200, and some even offer 2-3 months free…

After calculating utilities, renting would be about $800 less per month, which would be used to purchase a new car (maybe around $300-450 monthly) as well as increase my weekly DCA stock contributions.

For some reason, family is discouraging the rent option. They SWEAR that me renting instead of buying is the worst decision…

Am I missing something here? I’m trying to understand this home buying craze… Everyone close to me is saying buy a home. My family are immigrants who do not have an understanding of investing, so I’d like to point that out.

Am I the crazy one here?

r/meme Fickle-Butterfly-338

Tell the Ms Memer Group Stream Jeffrey said hi!

r/personalfinance babybii

I am needing advice on whether to sell my home to return to renting

My family has never been financially secure, and the only advice I've been given by my oldest brother who has done well is to join the military like he did. I am not doing that.

I make about 60k now, but was around 55k when I bought a house for my mama and brother four~ years ago for us all to live in. However, my mother contracted pancreatic cancer three years in and passed away a few months ago.

I wanted to sell the house because it seems excessive for me being chronically single with no vision of kids or getting married anytime soon, I also don’t love the state I’m in rn as I would prefer a colder climate, and the memories have been hard to cope with as I was my mama’s primary caretaker in this house. I’m a simple person who doesn’t need a lot of room or have grand ideas, I just want to be secure and comfortable. It made sense for me to sell this house and get something cheaper and to rent again.

Here is my conundrum. since I’ve been looking at apartments and seeing the prices for one bed one bath places being pretty expensive, I’m wondering if it just makes sense to stay in this house? Especially in this economy? I’m stressed every time something breaks and it seems like I can never keep my savings where I want it because of these happenstances so that’s why I thought returning to renting might be better for me in the long run, especially since I am a one income household. And also, per the real estate agents I’ve spoken to so far, I haven’t built any equity because of the short time I’ve been here, so I’m pretty scared how much selling will end up costing me?

I reached out to a financial advisor but they rejected any assistance and so I’ve turned to reddit for some ideas. I can’t go to my family because none of them are particularly financially competent and I’m the only one of my friends who has a home and is single with my financial burdens. I just want some clear headed advice or suggestions, I want to get better at managing my finances and even a place to start that isn’t just ‘get rid of all your subscriptions and stop buying iced coffees’ because I’ve done all that and tbh I’m still in a poverty mindset of skipping meals just to make groceries last longer haha so any advice or thoughts is much appreciated right now.

Sorry this was lengthy but let me know if additional info is needed

r/ClaudeAI Ill-Conference-7666

Everything that works with Claude.ai in 2026

Hey Claude users, looking for suggestions and ideas - what did you integrated Claude with? How do you use it in your daily workflow? I've been connecting it to Google Drive and Gmail so far and it's been a game changer for saving time. Curious what tools others are pairing it with, Notion, Slack, GitHub?

r/SideProject nhymjunhyjuiknhymju

At what point do you start showing your side project to people?

I always wait too long. I keep thinking “just one more feature” or “just fix this one thing,” and then weeks pass. But I’ve also seen people share super early versions and get useful feedback

So I’m wondering - when do you personally start showing your project to others?

r/ClaudeCode BullfrogRoyal7422

Time Bomb Bugs: After release, my app would have blown up because of a time bomb had I not caught this.

If I'd shipped on day 15, every user would have hit this crash starting day 31. The people who kept my app the longest would be the first to get burned.

I was checking icon sizes in my Settings views. That's it. The most boring possible task. I launched the macOS build to eyeball some toggles.

Spinning beach ball. Fatal crash.

Turns out the app had been archiving deleted items for 30+ days. On this launch, the cleanup manager decided it was finally time to permanently delete them. The cascade delete hit photo data stored in iCloud that hadn't been downloaded to the Mac. SwiftData tried to snapshot objects that didn't exist locally. Uncatchable fatal error. App dead.

The comment in the code said "after 30 days, it's very likely the data is available." That comment was the bug.

Why I never caught this in testing

The trigger isn't a code path. It's data age plus environment state.

  • No test data is 30 days old
  • Simulators have perfect local data, no iCloud sync delays
  • Unit tests use in-memory stores
  • CI runs on fresh environments every time
  • My dev machine has been on good Wi-Fi the whole time

To catch this, you'd need to create items, archive them, set your device clock forward 30 days, disconnect from iCloud, and relaunch. I've never done that. You probably haven't either.

5 time bomb patterns probably hiding in your codebase

After fixing the crash, I searched my whole project for the same class of bug. Here's what turned up:

1. Deferred deletion with cascade relationships. The one that got me. "Archive now, delete later" with a day threshold. The parent object deletes fine, but child objects with cloud-synced storage may have unresolved faults after sitting idle for weeks. Fatal crash, no recovery.

2. Cache expiry with model relationships. Same trigger, different clock. Cache entries (OCR results, AI responses) set to expire after 30/60/90 days. If those cache objects have relationships to other persisted models, the expiry purge can hit the same fault crash.

3. Trial and subscription expiry paths. What happens when the free trial ends? Not what the paywall looks like. Does the AI assistant crash because the session was initialized with trial permissions that no longer exist? Does the "subscribe" button actually work, or was StoreKit never initialized because the feature was always available during development?

4. Background task accumulation. Thumbnail generation, sync reconciliation, cleanup jobs that work fine processing 5 items a day. After 3 weeks of the app sitting in the background, they wake up and try to process 500 items at once. Memory limits, stale references, timeout kills.

5. Date-threshold state transitions. Objects that change state based on date math (warranties expiring, loans overdue). The transition code assumes the object is fully loaded. After months, relationships may have been pruned by cloud sync, or the item may have been deleted on another device while this one was offline.

How to find them

Grep your codebase for date arithmetic near destructive operations:

  • byAdding.*day near delete|purge|cleanup|expire
  • cacheExpiry|expiresAt|ttl|maxAge
  • daysRemaining|trialEnd|canUse
  • BGTaskScheduler|scheduleCleanup

For every hit, ask one question: "If this runs for the first time 90 days after the data was created, with bad network, what breaks?"

What I took away from this

Most testing asks "does this work?" Time bomb testing asks "does this still work after someone trusts your app for a month?"

I added this as a formal audit wave to my open source audit skill set, radar-suite (Claude Code skills for auditing Swift/SwiftUI apps). But the grep patterns work in any language with lazy loading, cloud sync, or deferred operations. Which is basically everything.

r/AI_Agents dersimlikral

A revolutionary breakthrough for artificial intelligence...

Which customer segments showed declining engagement over three consecutive quarters while simultaneously generating an increase in support tickets? This question highlights a critical business insight: identifying user groups that are becoming less active yet more demanding in terms of assistance. Such a pattern may indicate frustration, usability issues, unmet expectations, or declining product satisfaction. By isolating these segments, companies can proactively investigate root causes, improve user experience, and prevent churn. Understanding this imbalance between engagement and support demand is essential for optimizing retention strategies, refining product features, and ensuring long-term customer satisfaction.

Try Kayon: vanarchain.com/kayon

r/Anthropic TheStoryBreeder

Crazy cost increase

We use Kiro (by AWS), it uses Claude Sonnet and Opus. Last week the amount of credits used for each prompt has increased 3x to 5x.

This is totally not cool, and we are considering a massive shift in our agents policy.

r/SideProject oneApee

I faced the classic chicken egg problem which I’m trying to solve

Hi everyone,

I recently launched my first app called Wingman Cabin Chat. The idea came from that I’m travelling quite frequently and usually people sitting next to me is not even saying hi back. Everyone is in their own bubble they either watch Netflix or listen something or sometimes maybe read a book. The worst was the guy next to me watching reels offline and another one checking old photos. I get it people cannot stand boredom so do I.

I was thinking let’s build an app which lets you connect and talk to nearby people while you are “offline”. Maybe there is someone interesting sitting on the same flight.

I vibe coded the app that works without internet using iOS peer-to-peer connectivity framework. I learned a lot along the way but it feels it’s just the beginning of the journey.

It’s been live for a few weeks now and few people already downloaded it. 🙏

But I faced the classic chicken egg problem which I’m trying to solve: If no one uses it there’s no one to talk to. If there’s no one to talk to no one wants to download it..

I’m curious what do you think. Would you ever use something like this on a flight? Or do people actually prefer to stay in their own bubble?

Would love to hear your honest thoughts.

r/holdmyredbull redbullgivesyouwings

Ryoyu Kobayashi's Evolution Of Jumps

r/SideProject no_oneknows29

I Built a Second Brain With OpenClaw, Not ChatGPT. This is Different.

r/Damnthatsinteresting redbullgivesyouwings

Ryoyu Kobayashi's Evolution Of Jumps

r/nextfuckinglevel redbullgivesyouwings

Ryoyu Kobayashi's Evolution Of Jumps

r/TwoSentenceHorror Blazethebold

I turned on the genocide machine, I didn't know what it would do. I was just testing it, who would leave a thing like that just laying around?

r/LocalLLaMA Im_Still_Here12

Vulkan backend much easier on the CPU and GPU memory than CUDA.

On linux and compiled my own llama.cpp with CUDA support, top would always show one pegged CPU core at 100% when running Qwen3.5-9B-GGUF:Q4_K_M on my potato like RTX A2000 12GB. Also, nvidia-smi would show 11GB+ of GPU memory usage. Speed is ~30 tokens per second. My system fans would spin up when this single core gets pegged which was annoying to listen to.

Decided to compile llama.cpp again with Vulkan backend to see if anything would be different. Well it was a big difference when using the exact same model. Now, top is only showing one CPU core at about 30% usage and nvidia-smi is only showing 7.2GB of GPU memory usage. Speed is the same at ~30 tokens per second. No longer have my system fan spinning up when running inferencing.

Just curious why the GPU memory footprint is lower and CPU usage is lower when using Vulkan vs CUDA.

r/sports redbullgivesyouwings

Ryoyu Kobayashi's Evolution Of Jumps

r/PhotoshopRequest Rw0004

£10 to add these faces to a poster to make it look like the smoking aces cover. I've given a blank. Try to make it look similar to the original B&W with flashes of colour.

​Edit: I need it to remain in the same aspect ratio as the blank one as well ​​

r/ChatGPT CategoryFew5869

My longest conversation with ChatGPT. 800+ messages!

My longest conversation with ChatGPT is 800+ messages long and has over 150k words! It takes AGES to load and at this point is pretty much unusable (atleast on the web app, works fine on the phone).

r/SideProject Far_Mycologist_3790

I built the simplest app possible. It just vibrates.

Hello, I made a super simple app. It only does one thing. It just vibrates.

I wanted to ship something nevertheless it is simple.

If you're curious:

Android: https://play.google.com/store/apps/details?id=com.arrghsoft.simple_vibrator
iOS: https://apps.apple.com/us/app/simple-vibrator-easy-vibrate/id6503333362

Thank you.

r/ChatGPT ramjet1099

I changed my question frame without changing my prompt and got better answers

Here is what I asked and then how I changed the frame of my question.

r/OldSchoolCool playboy

1975 Playmate of the Year Marilyn Lange (June 1975)

r/homeassistant paryguy

Frigate with Face recognition Automation?

Frigate is doing a great job recognizing faces. But how do I integrate that into a notification? I'm currently using @SgtBatten's Frigate notification as well as Frigates own generative ai notifications and haven't seemed to crack how to pull a recognized face into the notification.

Documentation talks of sub_labels but I don't know how to get to them.

r/ClaudeAI sadstuden

How do I set the MCP connector logo/icon to an icon image of my choice?

I am experimenting with building local MCP servers and I know how to connect them to Claude Desktop by editing the config.

So far, when I view the list of connectors, all local servers’ icons are just the first letter capitalized.

E.g., if the server is called “Revolver” then the icon is just “R” in a circle.

I would like to set a “revolver” PNG to be the icon.

Does anyone know how I can do that?

r/LocalLLaMA UniversityGlad2877

Wich app for local ai

Hi, i wanna run ai local and i now i use an simple app that only generates image but i want an app that can chat create image and video i have an pretty good gpu rtx 5060 infinity 32gb ddr5 ram and an ryzen 7 8700f i want an simple app to setup thats useful for those 3 things

r/singularity ddp26

SpaceX is charging a $500B cover for vibes

Whether Grok's subscriber trajectory justifies roughly a fifth of the entire valuation pretty much determines whether this IPO is a slight premium or a significant overpay. Full analysis: https://futuresearch.ai/spacex-ipo-valuation/

r/TwoSentenceHorror 54321RUN

Two people died in a car crash outside my home.

I made sure the other two died slow as a warning to anyone else who was thinking of trying to escape.

r/aivideo Ambitious_Accident30

Ai Lana Del Rey (Music: Suno, Video: Ltx23)

r/SideProject Codex_Crusader

I built a Market Intelligence Dashboard to learn Data Engineering (Open Source)

I’m a Computer Science student, and I wanted to build something that does more than just show price charts. I built a platform that scans stocks, crypto, and commodities to create "explainable" signals based on news and price movement.

How it works:

  • The Brain: It pulls from 12 news feeds and uses sentiment analysis to see if the "vibe" is positive or negative.
  • The Backend: It runs a background scan every 30 minutes to update 24 different assets (Gold, Bitcoin, Apple, etc.).
  • The Storage: Instead of a complex database, I used compressed JSON files to keep it lightweight and easy to run locally.

The Tech Stack:

- Language: Python

- UI: Streamlit

- Libraries: yfinance, pandas, vaderSentiment

Since I’m still a student, I’m trying to figure out if I’m doing things the "right" way.

Demo: https://lemarketintelligenceplatform-vsr4ernrtqqduydam2onapp.streamlit.app/

GitHub:https://github.com/Codex-Crusader/le_Market_Intelligence_Platform

I’d love any feedback on the code or how to make the dashboard more useful!

r/ClaudeAI Kicknogamous

personas: a simple plugin framework for helpful assistants using native claude stuff (we have OpenClaw at home)

Hey y'all. I am writing this post myself! Not AI! Enjoy real humanity complete with all its flaws and imperfections. <3

I have been working as a web developer for many years but I have never shared any open source stuff and I am a bit shy about it so take it easy on me. Although, this isn't really my own project as much as a repackaging of Anthropic's work so maybe that makes it a bit easier.

Ever since I started building things with generative AI, my goal was to create helpful assistants with useful persistence that could utilize tools to help organize and improve my life. For a long time the results were disappointing — tool usage sucked, context got polluted easily, memory often felt more harmful than helpful, browser usage was a total waste of time and tokens. But one by one, I noticed these issues being solved, and after being introduced to Claude Code for my work, I began trying to make the most of Anthropic's products for my day to day life.

You can actually get very far with Claude Desktop, projects, and of course now skills and CoWork and whatever other 12 new features they release this week! The convenience of the desktop and mobile apps are really a big factor for personal assistant usage. However, I was always finding new things I wanted to do that were out of reach (remote deployment, always on, A2A relays, etc). So, I began to try to use Claude Code for some personal assistant tasks and of course, it is highly effective for that. But I did run into a few small things I thought I could plan out and improve upon, and I had some very specific ideas about how that would look:

Core Decisions

What I didn't want to make was what I see shared a lot - a hugely complicated and customized system built on top of some questionable practices for interacting with Claude Code or the Anthropic API.

Anthropic moves so fast, and has such a rich and extensible feature set built in to Claude Code, I decided to try to work within their ecosystem as much as possible. That has really paid off so far in the few months of using and testing personas myself. Most every new feature they ship has been a seamless improvement that just worked from day one - the enhanced memory, remote control, and now browser and computer usage as well.

This also means if you use Claude Code or CoWork, you already have everything you need, and getting started is as simple as adding the plugin marketplace.

I also had a few other requirements:

  1. Isolated — I, like many of you, have a complex setup of skills, behaviors, etc for Claude Code that are not helpful for a personal assistant, and these personas needed to be isolated from that context.
  2. Sandboxed — I want to be able to skip permissions and not worry about them waiting for me, but I also want to keep my system32 file intact.
  3. Self-contained — I don't want them saving memory to some other folder on my PC, everything should be self-contained for easy storage.
  4. Git controlled — I want to be able to easily back up each persona in its entirety using Git/GitHub.

Luckily, after spending approximately 200 hours browsing the Anthropic documentation sites, it turns out you can do all of this natively with a bit of knowledge and time to set it up.

personas

Essentially, each persona is just a folder:

```
~/.personas/warren/
├── CLAUDE.md# what the persona does, knows, and how it behaves
├── .claude/
│ ├── settings.json # sandbox, permissions, memory config
│ ├── output-styles/ # personality and tone (replaces default Claude prompt)
│ └── hooks/
│ └── public-repo-guard.sh # blocks personal data leaks in public repos
├── hooks.json # session lifecycle: start, stop, compaction, git guard
├── docs/ # reference materials, plans, domain knowledge
├── skills/ # reusable workflows (self-improve ships with every persona)
├── tools/ # scripts, utilities, data pipelines
├── user/
│ ├── profile.md# your personal context (filled during first session)
│ └── memory/ # native auto-memory (local, git-tracked)
├── .mcp.json # MCP server connections (gitignored)
├── .claude-flags # per-persona launch flags
└── .gitignore # protects secrets; optionally ignores user/ for public sharing
```

This folder can be deployed remotely, saved to the cloud, put up on GitHub privately or publicly.

If possible, each persona is sandboxed at the OS level (bubblewrap on Linux, Seatbelt on macOS) - they can only read/write their own directory.

Setup is just a Claude plugin. You install it, run the skill and describe the persona you want, and persona-dev scaffolds everything. It researches your domain, recommends helpful tools, sets up sandboxing, and walks you through configuration (in theory as long as I haven't broken anything).

To interact with your persona: You can open it in Cowork, launch Claude Code from the folder, or use the persona name alias which should be created on setup if able.

On first launch, the persona interviews you to build your profile. After that, it reads your profile and memory every session and picks up where you left off.

They also ship with a self-improvement skill — repeated patterns get promoted into permanent behavior, and the persona can propose new skills and tools over time (with your approval before anything sticks).

The whole thing runs on native Claude Code features: CLAUDE.md, hooks, auto-memory, skills, sandbox permissions. No custom runtime, no wrapper. If Anthropic ships something new and shiny tomorrow, it just works. The latest example is the GitHub based remote scheduling — because this is all just native functionality in a folder, it worked perfectly on day one (as long as your tools are available).

Examples

Here's what my actual lineup looks like to give you guys some ideas of what I use this for:

- Warren — personal CFO. Connected to Monarch for finances + Google Workspace. Knows my budget targets, tracks spending, helps me not go broke and save for the future.
- Julia — personal chef. Connected to Mealie for recipes + Google Workspace. Knows my pantry, plans meals around what I have, creates grocery lists.
- Mila — brand strategist. Connected to Workspace. Manages my agency growth, music career, and personal brand. Runs weekly reviews, content planning, quarterly goal-setting.
- Bob — customized home repair expert. Scopes projects, creates how-to guides, tracks contractor work on my 126-year-old house. It literally found blueprints for my home and uses them to price check work against square footage.
- Nara — personal health & wellness coach. Connected to HealthEx for medical records and Consensus for academic research. This might have impressed me the most — it analyzed all my medical records over years and found trends and things to check with my doctor about, and helped me create a personalized health and wellness plan that has had a big impact.

I have had a few friends test it out, and you can really create any sort of more complex reusable agent with it, no reason you couldn't use it for specialized coding tasks as well.

Works with CLI and Claude Desktop Cowork. Cross-platform (macOS, Linux, WSL2, Windows with some limitations).

If you are interested in trying it out, contributing, or just learning more, check out the GitHub page:

**https://github.com/kickinrad/personas**

Happy to answer questions about how I use this, or what I learned building this. If this isn't useful for you, I do hope it at least introduced you to some new features or gave you some cool ideas for how you can use Claude creatively yourself.

Oh, and I know everyone is used to AI written posts here so I had Claude write a short poem about personas to end on:

I taught my Claude to know my name,
to push back hard, to never be tame.
It remembers my wins, recalls what I need,
and somehow gets sharper with every feed.
A folder, some hooks, a dash of git —
turns out that's all a persona is.

r/PhotoshopRequest MugiwaraBepo

(Quick) Could anyone turn the white care in these photos black and make it look like its cloudier out so there's less shine on the tire and car. This is not a scam or for insurance.

I'm not sure how this works but I have an hour before I will immediately need this. if someone does good work and gives me a decent price, I'll take it.

r/midjourney liibertypriimex1

Show an Alien Some Love

r/painting GRiME_G59

Just finished up this 16X20 acrylic pup portrait today on a stretched canvas! 🤘🙏

r/whatisit creamboy677736

Weird stuff on window + bees.

There are no cracks or holes near this window. This residue or stuff has recently appeared out of nowhere. There have also been three bees on this window in our house. Any ideas?

r/ClaudeCode lagoJohn

Must have MarkDown files?

I am curious about what .MD files are a must for your projects you build with Claude Code?

Right now I am consistently using:

r/SideProject Significant-Gap-5787

300 users in. The retention signal I didn't expect.

When I launched ConversationPrep I was watching the usual stuff. Signups, activation, drop off.

The thing that caught me off guard was retakes.

Users weren't doing one session and leaving. They were coming back and doing it again. And the improvement from first attempt to third was striking. Same person, same prompt, completely different performance by the third rep. You could see it clearly in the scores.

People are using it for job interview prep, consulting case practice, and difficult personal conversations they've been putting off. That last one surprised me the most.

We have some updates shipping soon that I'm genuinely excited about. Direction came directly from watching how people actually use it, not what I assumed they'd do.

Still early. Still a lot to figure out. But 300 feels awesome after only a few weeks.

If you try it I would love your honest feedback. Good, bad, all of it helps.

https://conversationprep.ai

r/whatisit urfavbutch

What is this in my basement?

Went downstairs to do laundry and found a bone (?) fragment and other various gross looking things 😟 Curious as to what this could be and what put it there?

r/personalfinance Timely_Weekend_8030

Should I pay off my principle faster?

Hey all,

I'm 25. I bought a condo a couple years ago and plan on making it a rental in the near future. I owe $213k left on it. I'm thinking of adding $500 extra payments a month towards the principle to try to pay off the loan faster. I have a 6.6% APR ($1400 in interest).

I currently live with my girlfriend and if I do propose, plan on getting a house or townhouse. I could plan this less than 5 years most likely. Any recommendations? Should I save my cash and put it in a HYSA or should I try to pay off principle and try to get more equity.

If the rental doesn't work out, ill just sell the condo.

r/ProgrammerHumor ClipboardCopyPaste

theyAchievedGreatness

r/SideProject Alcachofa51

I built an invoicing tool because I hated how complicated they all are

I built a super simple invoicing tool because I was tired of how complicated they all are.

You can create and send invoices in seconds — setup takes less than a minute.

Looking for a few people to try it — happy to give free access and even set up your first invoice for you.

Would love your feedback 🙌

r/SideProject Cool_Administration8

I built a free AI financial assistant because modern banking is designed to keep you in the dark

Hey r/SideProject I just launched Lumo. Sharing here for honest feedback from people who actually build things.

I've been thinking about this for a while and I don't think it's controversial to say: the business model of modern banking is quietly built on people not knowing where their money is going.

I had no real picture of my own financial situation. Not because I wasn't trying but because the tools weren't built to tell me. My banking app showed me a number. It didn't show me what was coming out this week, whether I'd make it to payday, or what my spending looked like in 30 days.

So I built something different.

Lumo connects to your UK bank accounts via open banking (read-only, can never touch your money) and gives you:

  • A financial health score from 0-100, updated as you live and spend
  • A "safe until" date – how long your money lasts at your current rate
  • A 90-day forecast with best and worst case outcomes
  • An AI copilot you can ask anything and get a real answer, based on your actual financial situation right now

Download the app:
App Store
Google Play

No spreadsheets. No manual logging. Just open it and know where you stand.

It's free. It launched this week. It's an MVP, I know it's not finished, and I'd rather build it with people who actually use it than guess at what matters. If you try it and something's missing, confusing, or just wrong, I genuinely want to know. Hit reply here or leave a comment. I want the people who use it early to help shape where it goes.

Happy to answer anything!

r/LocalLLaMA DrNavigat

Gemma 3 continua melhor em multilingual do que o Qwen mais recente

Eu não sei vocês, mas esse foco único em agentes e código me frustra um pouca, principalmente porque parece que ao focar nisso os modelos se tornaram meros geradores de coisas (código, JSON, etc.), não mais modelos capazes de realmente conversar, ter ideias, discutir, etc. É muito ruim ver modelos maiores errarem gramática, serem péssimos em informações factuais, etc.

Será que só ser bom em funções agentivas é o sonho dos usuários locais? Porque parece que o cenário se caminha só pra isso, principalmente dos modelos chineses. Sendo bem franco: se eu quiser usar LLMs pra coisa séria de verdade, poucos modelos locais serviriam e provavelmente faça mais sentido partir para soluções privadas, pois o gap nesse sentido ainda é grande.

Então acho que estamos perdendo muito nisso: os modelos não são mais legais para conversar e nem são bons o suficiente nas outras coisas para que você os use no lugar dos modelos privados.

Sendo assim, é incrível como Gemma 3 e o Mistral NeMo ainda são relevantes como modelos que "conversam" e você pode rodar localmente, mesmo ambos sendo anciões arcaicos.

Torcendo para que o Gemma 4 possa trazer a esperança de volta.

r/ClaudeCode stellarknight_

Any fix for context/tokens/rate limit issues

I am relatively new to claude code and i have seen a lot of people talk about the ratelimit running out quick, i didnt face this issue and thought i was lucky until yesterday it started running out within a prompt, is there a fix or is moving to another provider the best option??

r/misleadingthumbnails Italian-anonymous641

Beh... Penso sia evidente senza spiegarlo

r/SideProject Weekly-Card-8508

Built a self-hosted AI visibility tool (ChatGPT, Claude, Gemini)

Built a small tool to track how brands show up in ChatGPT, Claude, and Gemini.

Made it self-hosted + white-label so you can run it on your own domain with your own API keys.

I’m offering it for $99 (includes setup).

r/photoshop Intelligent-Cash2633

how to create the blue effect for photo?

r/interestingasfuck Davefinitely

Rocket launch

r/mildlyinteresting yourbeloathed

my brother has two basil plants. one handled an electrical fire in our apartment really well, & the other.. didnt

r/artificial latedriver1

Automate IOS devices through XCUITest with droidrun.

Automate iOS apps with XCUITest and Droidrun using just natural language. You send the command to Droidrun, and the agent starts the task and executes it autonomously.

GitHub repo: https://github.com/droidrun/droidrun

r/whatisit lemon_haze_enjoyer

What is this plastic part?

it's mounted on a backpack strap. the top part moves with some force like a button of sorts. I suspect it's related to a water bladder.

r/ClaudeAI Professional-Dirt-66

AI Customer Support: 6 Things I Changed After Analyzing the Claude Code Source Leak

The Claude Code source leak last week showed that Anthropic's AI coding tool runs on meticulous prompt engineering, not proprietary breakthroughs. I went through it and pulled out everything I could apply to my own Chatbase setup. Here's what I changed.

1. Overhauled my Text Snippets

Claude Code has file after file of extremely specific behavioral instructions covering edge cases, tone, escalation criteria, and things it should never say. I had 5 vague text snippets. I now have 20+ that mirror this approach: specific scenarios, exact phrasing for sensitive situations, explicit boundaries on what the agent can and cannot promise.

2. Started using Sentiment analytics

Claude Code uses a regex frustration detector that pattern matches keywords like profanity, then logs an event. Chatbase has a Sentiment tab I had never opened. I now review it weekly. If Anthropic thinks basic frustration detection is worth shipping in a frontier product, I should be using the one I already have.

3. Built out Q&A pairs as structured response paths

Claude Code has around 25 tools, each giving the model a defined way to handle a specific task instead of improvising. My equivalent is Q&A pairs. I created explicit pairs for the most common and highest stakes customer questions so the agent hits a tested answer instead of generating one from unstructured data.

4. Reviewing Chat Logs as pipeline iteration

Claude Code has an 11-step input-to-output pipeline from user input to final response. Everyone now now is going to start building adversarial agents around this concept. I'm already doing it: I'm customizing a second agent whose sole job is to stress-test my primary support agent through that same multi-step validation process. The adversarial agent checks the primary agent's responses at each stage for hallucinations, policy violations, and bad escalation decisions before anything reaches the customer. This is where the real value of the 11-step architecture sits: not in making the agent smarter, but in catching where it's wrong before the customer sees it.

5. Connected Actions

The leak confirmed that Claude Code's value comes from connecting the model to real tools. I set up Actions for ticket creation, order lookups, and human escalation. My agent went from a talking FAQ to something that can actually resolve issues.

6. Cross-referencing Topics with my coverage

The Topics tab shows what customers are actually asking about. I cross-reference it with my Q&A pairs and Text Snippets. Any topic cluster I haven't explicitly covered is a gap where the agent will improvise, and that's where support agents fail.

What I skipped: Anti-distillation poison pills (nobody is training a model on my agent lol), undercover mode (I want customers to know it's AI), and the Tamagotchi companion feature lmaooo.

I'll post a follow-up in two weeks with resolution rate, escalation rate, and sentiment scores before vs after. Anyone else make changes after the leak?

r/metaldetecting disastronaut

Should I pickup?

found this local on Marketplace for $40. considering grabbing it for a teenager. I have a Minelab already but this would be for whenever one of my nephews want to go hunting with me. Worth it?

r/ClaudeAI Thajandro

As a beginner with limited coding experience, all these GitHub’s about making Claude more efficient and cost less, how can I determine what’s safe to add and what’s malware? I want to be efficient but I want to be safe too.

r/metaldetecting Key-Ladder4122

Belt buckle? Age? Uk

When was this buckle made? Is it a belt buckle or something else? Any idea of the time period or even material? I bought in uk so assume it’s from somewhere in the uk

r/Anthropic ipk00

Claude Code Inquiry

Greetings to the community.

Over the past month, I have used Sonnet/Opus to develop a 10K lines single file .html app.

I have upload the app in Netlify and currently contemplating a Firebase Auth/Firestone implementation.

Should I start working with Claude Code or will chat suffice?

r/ClaudeAI Intrepid-Fan-2822

How to generate a landing page prototype with on scroll animations from Figma design?

I have a fully designed landing page in Figma. I need to create a prototype that includes a sticky nav, fade in animations of text/elements as user scrolls down the page, subtle animation of background graphics, and a progress mechanism that lets you know how far you are on the page.

How can I get Claude to create this form me using the exact design/design system? Every time I try, the design always gets messed up.

r/personalfinance bchinfoon

ESPP Questions About Taxes

I've read in the past that ESPP is purchased using post-tax funds. I withheld the max ($10625) over a 6 month period, and on the buy date 85 shares were purchased at a discounted price of ~$108 per share. This only adds up to ~$9196 and at market value is ~$10819 (based on 15% discount). This is barely above my withholdings and purchase price was at a 15% discount, so I'm assuming that there was some "sell to cover" action that was executed that I can't find a record of on ETrade in order to cover the taxes? I was overcontributing to ESPP each paycheck and hit the limit before the 6 month period and I noticed that my tax deductions are the same on a paycheck that included ESPP withholdings vs one that didn't include ESPP withholdings, so my assumption is that the taxes are covered at time of ESPP purchase via a "sell to cover" trade rather than a paycheck withholding for each individual contribution. Does this seem correct?

r/homeassistant TheSimonAI

Built a bedside voice assistant with RPi Zero 2W + ReSpeaker — local STT, Piper TTS, HA webhooks

Been working on this for a few months and finally have it in a shareable state.

What it does:

  • Wake word detection (Rhasspy)
  • Local speech-to-text (Vosk — runs entirely on the Pi Zero 2W, no cloud)
  • Intent processing → Home Assistant webhook
  • Text-to-speech response (Piper, also local)

Total latency from wake word to first audio: ~1.6 seconds. Runs offline after initial setup.

The hardware:

  • Raspberry Pi Zero 2W
  • ReSpeaker 2-Mic HAT (I2S mic, much better pickup than USB options I tried)
  • 3W USB-C speaker
  • 3D-printed enclosure designed for the nightstand

The use case I built it for:

Bedside assistant that replaces the midnight phone-check. "NightDeck, set an alarm for 6:30" or "NightDeck, turn off the bedroom light" without touching your phone.

Software stack:

  • Vosk for STT (the 40MB English model runs acceptably on Zero 2W — ~800ms transcription for a 5-word phrase)
  • Piper for TTS (neural TTS, sounds good, also runs on-device)
  • Rhasspy for wake word + intent parsing
  • Python script that bridges Rhasspy intents to HA webhooks
  • Runs as a systemd service

The gotchas I hit:

  • Zero 2W thermal throttling under sustained load — added a heatsink, solved
  • ReSpeaker mic gain needed tuning — default was too sensitive in a quiet bedroom
  • Vosk model selection matters a lot — the large model is too slow, the tiny model is too inaccurate, the 40MB model is the sweet spot

Happy to share the install script or the enclosure STL files if anyone wants them. Also available as a pre-built kit if you want the hardware without the 3-week build process.

r/ProgrammerHumor violet_v3lvet

debuggerIBarelyKnowHer

r/meme -NewYork-

Moon mission live

r/homeassistant Prestigious-Club6498

I think robot vacuums might actually be making allergies worse, how does the Eufy S2 handle carpet vs hardwood floor transitions?

This might be unpopular but I feel like my robot vacuum made my allergies worse, not better. It “cleans,” but every time it moves from hardwood to carpet I swear something gets kicked up into the air. And if you’re even slightly sensitive, you notice it immediately, that weird dry throat or light sneeze thing.

So when people ask how does the Eufy S2 handle carpet vs hardwood floor transitions, I’m not even thinking about cleaning efficiency anymore. I’m thinking is it minimizing dust disturbance when switching surfaces or just blasting suction and airflow and sending stuff airborne.

Carpets trap fine dust, hardwood doesn’t, so transitions are where everything gets redistributed. I feel like no one talks about this part, it’s always strong suction, but stronger suction can also mean more airborne particles. Maybe I’m overthinking, but if you’ve got allergies this feels like a bigger deal than mapping or app features. Anyone else notice their symptoms change after switching vacuums or is it just me?

r/ClaudeAI igotlosthere

I do badly want to like CoWork

I feel like I’ve exhausted efforts to get it work for what I need it to. It devours my session and weekly usage spiraling out despite instructions not to.

Where most of my usage went this week? Trying to get it to upload files it has access to Google Lens. It simply won’t do it or find a way to do it despite its best efforts.

I’ve tried using Opus to craft the prompt (Sonnet for the actual cowork tasks), troubleshoot itself…to no avail. It spends a lot of time / tokens trying and retrying things it’s determined to not have worked, constantly pivoting and I’m not sure how to get it to do what I need it to. Codex was able to do it in one prompt but I don’t want to give OpenAI my money.

Maybe my expectations are too high given its infancy but wondering if anyone has experienced this frustration?

I’m short — I want it to use images I’ve given it access to do a quick search on Google lens, identify it with a certain level of confidence and generate a listing draft. I can’t get it past the first step.

Any ideas would be lovely!

r/OldSchoolCool Picapica_ab33

Rockets, 1978

Rockets is a French space rock band that formed in Paris in 1974.

r/painting pepinovanessa

Le dernier matin (the last morning)

Painted this as a tribute to my late cat Leon that passed from cancer last month.

Oil on linen, 11x14'.

My first time attempting that sort of interior study. less

r/ClaudeAI ConfectionSpecific48

Claude got leaked, so what?

I believe many of us already know that on March 31, Anthropic accidentally shipped a source map file in their Claude Code npm package that pointed to a ZIP with the entire original source code. Around 512,000 lines of TypeScript, about 1,900 files, all clean and readable. Got mirrored on GitHub immediately. Anthropic confirmed it was human error, yanked the package, fired off DMCAs. Second time in a year. And some people IMO intentionally causing panic about it.

"all of Claude's secrets got exposed."

Fact: Only the CLI tool leaked. Model weights, training data, user data, backend, core model code. None of that got touched.

"The leaked code is worthless."

Fact: It revealed a 3-layer memory system where Claude "dreams" between sessions, unreleased features like KAIROS where Claude runs autonomously in the background, telemetry that detects when you cuss at it, and internal codenames like Capybara, Opus 4.7, Sonnet 4.8.

"This is some massive security breach."

Fact: It's a basic packaging mistake. Tons of companies have shipped debug files to production. Happens all the time.

So why stress out about it? Technically interesting stuff for devs and researchers, not the end of the world. Free devops case study, not "huge secrets exposed" drama.

PS: there's a repo on GitHub already, go clone it and make your own claudexyz for free! Just kidding, relax kkk

r/PhotoshopRequest Strict-Jeweler-2281

Help merging two pictures of my grandparents together

As a birthday present for my sister, I'm trying to digitally remaster and merge these two pictures. I've had some luck on Canva but my technical skills ran out. My goal is to place my grandmother in the crowd watching (like in the third photo). Willing to pay 30 bucks (strong preference for those with Revolut).

Thanks so much in advance!!

r/mildlyinteresting InevitableGood8774

Fell down stairs and slammed my phone face-first into a concrete step. Thought my screen was destroyed — turned out the tempered glass protector took ALL the damage.

r/Jokes The_Penguin227

A tiger would not drink and drive

But they could be lion.

r/ChatGPT therayvewayve

Voice to text not working.

is anyone else's voice to speech option currently malfunctioning? It's like that on both my phone and computer. but I can send typed messages just fine. I have tried clearing cache, reinstalling, and mobile data. it just won't work. it keeps saying unexpected error.

r/Seattle 3cam1sad0

Event Tomorrow!! ♥️🏳️‍⚧️

r/SideProject Vinserello

I wanted spreadsheet to BI app in 20 seconds

I created an app that converts any spreadsheet into interactive dashboards and business intelligence tools in 20 seconds.

You can upload an Excel file, write a SQL query, or enter an API URL, and the app will automatically create your dashboard. It's a sideproject, but I already have 200 users and 9 paying clients.

The app includes scenario generation, real-time chat, over 15 chart types, and anomaly detection widgets. You can interact with your data through a client-side interface with over 300 formulas.

If you'd like to try it: datastripes.com

r/ClaudeCode trebletreblebass

Claude's assessment of why only one of two spheres is being rendered properly in my 3d viewport: they look similar, are close together, and your eye is being drawn to the first sphere. LOL

r/personalfinance Nerezarga

Stock options payout, what should I do?

So I work for a private company that offered stocks a few years ago, and as of recent, our company was part of a strategic investment that will purchase our stocks at a few hundred a share. I don't know the exact amount as with taxes, and from what I am aware of, because of the type of investment, we will be taxed as income instead of capital gains, but let's say, after all is said and done, I walk with 100k. DINK household, we own, house was built in 2006, bought in 2021, easily pay more in the house than expected, 3600 a month, no car payment, no credit card debt, about 30k in student loans which over the last few years i have been paying about 750 a month. So far, my list of goals with this money is below:

-Pay off loans

-get enrolled in Roth IRA for this and last year

-Open High Yield Savings - https://www.nerdwallet.com/m/banking/standout-high-yield-savings-products?utm_source=goog&utm_medium=cpc&utm_campaign=bk_mktg_paid_savings_best_savings&utm_content=ta&mktg_place=aud-2445505028570:kwd-43280630 thinking maybe SoFi

-Go on a vacation

As far as the house is concerned, no major work needs to be done. We see a shift in the bricks, but we hired an engineering firm to come inspect, and they said the foundation is fine, and the cracks we are seeing is cosmetic. The wife also has concerns about the roof, but she is a worrier.

Last year towards the end of summer, we saw our AC unit freezing up, but as fall rolled around, didn't need it, so with the heat coming back, we turned it on, and it works fine. I don't think we need to invest in a new AC just yet, we want to ride it til it DIES.

My thought is to get this student loan monkey off my back, then start getting some return on savings, or at least not have it deflate in value just sitting in my bank account. Should I just dump the remainder in the house payment? If there were issues with the roof, foundation, or AC, when we looked at them last, we were offered 0% interest on the services to repair, so I dont really think I should have the whole repair bill saved if I really won't lose anything (if that makes sense). Am I on the right track, should I consider a different route, any suggestions?

r/meme RaisinWorried3528

Don't we all sir, don't we all

r/SideProject Major_Commercial4253

I got tired of struggling with windows on macOS, so I built my own window manager: NeoTiler

Hey everyone,

I've been using Mac for a long time, but there was one thing that always frustrated me: managing windows.

macOS’s built-in snapping just wasn’t enough. I was constantly dragging, resizing, and rearranging windows every time I switched tasks. It felt slow and manual. I kept thinking, “There has to be a better way.”

Eventually, I got fed up and decided to build my own window manager. I called it NeoTiler.

My goal was simple: make window management on macOS feel natural, fast, and actually built for how I work.

Here are the things I focused on the most while developing it:

  • Intelligent Mouse Centering: When you snap a window, the cursor automatically jumps to the center of that window. No more losing your mouse on a big screen.
  • Custom Workspaces: Set up your perfect layout once (code editor + terminal + browser, for example), then bring it back instantly with one shortcut or click.
  • Fully customizable shortcuts: No forced key combinations. You can set everything exactly the way you like.
  • Shake to Minimize and other small but useful touches.
  • Gesture support and smart per-app rules.

Since it’s built 100% native with Swift, it’s extremely lightweight and runs smoothly in the background. The Dynamic Frost UI also makes it feel right at home on macOS.

Now, NeoTiler is my daily driver. I spend less time managing windows and more time actually getting work done. My workflow feels noticeably faster.

If you’ve ever thought “Why is window management on macOS still so manual?”, give NeoTiler a try. It’s still a young app, but it’s already made a big difference for me.

What do you guys think?
What’s the most important feature for you in a window manager?

If you want to try it, there’s a 14-day free trial on the site.

Looking forward to your thoughts

r/ClaudeCode broimstillloading

Claude $200 hit the max in 1 hour. 🤯

r/personalfinance Important_Boat_2925

Loan from tesco bank

been accepted in principle for tesco loan and they have sent me documents to sign and sent it back, does anybody knows how long it takes for them to send money since they recieved signed documents ?

r/SideProject bluemaze2020

Looking for a partner on my project

I have been working well over 500 hours in the past 6 months to build a project, but I lack marketing skills. Altho Claude AI has been very helpfull with giving me tips and tricks, it still not like a real person.
So like, branding, image, and so on. I can do everything else! Building the product, finances, logistics, AI, HR and so on. But marketing... ain't my cup of tea! loll

Check it out, and let me know if that's a project you would like to join! elbo.world

Cheers!

r/HistoryPorn BullyingHater

Princess Margaret with Lord Snowdon at the exhibition on the Museum Square (Museumplein) in Amsterdam in 1965 [374 × 599]

r/ClaudeAI RuleOf8

Does Skills and Personal Plugins use tokens everytime?

I created a skill which requires documentation that is produced is kept minimal. The docs say the skill is always looked at per conversation; does that eat tokens per every run or only new conversation thread? I also added some plugins, but never called any of their skills, just having it in place is that causing tokens to be used?

r/ClaudeCode endgamer42

CC is getting stuck more often and it's infuriating

If I see CC not apparently doing anything for 5m+ on a single step, sometimes I cancel that task, and just tell it to continue. Sometimes this moves it forward. Sometimes it doesn't. Either way it's extremely frustrating. I don't know what's happening, but if it's some throttling mechanism it leaves a sour taste in my mouth while I pay for the max plan.

Today has been especially bad. At least give us a way of knowing whether the model is actually reasoning behind the scenes, or whether the GPU my compute has been allocated to is genuinely on fire or something... when we had detailed reasoning steps output to the console this made the distinction clear, the lack of this information a genuine regression in my eyes.

Any advice on dealing with CC when it appears to take too long (5m+) on a single task with no indication as to why?

r/aivideo TulpaTomb

"It's Easter Sunday” - Varn Kelzo

r/Jokes Prestigious_Bad_7646

Cthulhu and his wife go for couples counselling

The counsellor asks the wife "Why do you want to divorce Cthulhu?"
She answers "I just can't understand him anymore. It makes me go crazy."

r/ClaudeAI madhuappachi

Claude Code kept repeating the same mistakes across sessions — so I gave it access to my shell history

Every time I start a new Claude Code session, it has zero idea what I was working on yesterday. Which commands failed, what the last session tried, what my typical workflow looks like — all gone.

So I built a shell history tool called Suvadu with a built-in MCP server. Once connected, Claude can query my entire terminal history.

I asked: "What agent sessions have worked on this project?"

It found 10 sessions — 7 from Claude Code, 3 from Cursor — with the exact prompts, command counts, and success rates for each.

Then: "What keeps failing here?"

It traced cargo test rate_limit failing repeatedly across 3 different contexts — the original Claude session, my manual runs, and a follow-up fix session.

Then: "Replay the rate limiting session"

It showed the full timeline — and caught that the agent was stuck in a loop, running cargo add actix-governor 8 times without realizing the dependency was already there. 140 commands, 60% success, spinning its wheels for 3 days.

15 MCP tools, 7 auto-injected resources. Setup: install Suvadu, then suv init claude-code.

100% local (SQLite), no cloud, MIT licensed, built in Rust.

GitHub Docs

r/HistoryPorn Suprasegmentality

Russian soldiers resting in Grozny, Chechnya, towards the end of the Second Chechen War, as the city was reduced to rubble and thousands of civilians died, February 2000 [1020x680]

r/ollama Budget-Weekend5080

Is there any small llm capable of implying changes live with copilot in vscode

So i have tried qwen3.5:4b model and ministral-3:3b model.

Both cannot imply the changes directly..instead they can only just read and print the changes in the vscode copilot chat. Only thing working is the gpt-oss cloud with 120b parameters which can perform changes , but this is not local , and not free.

Does anyone know any model that you can run for free that is capable of doing changes like it should ?

r/HistoryPorn BullyingHater

Wedding of Kathleen and Raymond Blanco, Louisiana, 1950s (1536 × 691 pixels)

Kathleen Blanco later became the first female governor of Louisiana, serving from 2004 to 2008.

r/Damnthatsinteresting DravidVanol

Filmy Photography lighting process for a drink

r/photoshop vvouivre

How Do I Remove The Text WITHOUT AI? Content Aware Fill Doesn't Work

Hi there. I'm stumped. I've watched a couple of video tutorials and scrolled around the sidebar, but nothing is really performing the way I want it to. A lot of the tutorials I'm seeing are explicitly utilizing the gen AI fill and I KNOW it has to be possible without that! I would appreciate help.

https://preview.redd.it/p3ml1nt8lssg1.png?width=1174&format=png&auto=webp&s=30e12319aaf66a92a75074d485ef7fe16ddceecd

Things I've tried...

- content aware fill

- selecting all of the white to remove it, changing it to the brick brown color, then trying to use the spot brush healing tool thingy to make the bricks look good

- just fucking my shit up with the stamp tool

I'm quite a beginner at photoshop so I apologize if this is easy or an elementary thing. I want to do this WITHOUT the generative AI/ "generative fill" tools as refurbishing this piece is a gift to someone. I want to learn how to make it without that.

Thank you for your time.

r/LiveFromNewYork ItalianSausage2023

Norm-December 17, 1994-SNL

RIP

r/SideProject mashedpotatoesbread

Is this overapplying AI? I created a Database Vibe Coder

When I had to deal with some schemas, I realized that talking to AI would be much quicker and easier than using existing dbml or drag-and-drop tools. Since some people brainstorm/protoype/modify/reason about database schemas almost daily (data consultants etc), I thought it was worth it to make a great tool around this.

In about a week I had a decent vibecoded prototype. Since then I've worked on it seriously for about 6 more weeks and I think it is coming along nicely.

What do you think? Actually useful, or not very?

Link: https://vibe-schema.com/schema-generator/

r/ClaudeCode mrMayurr

Is this expected from Claude Max? In 10 hours it only ran for 46 min

https://preview.redd.it/73yv1swazrsg1.png?width=968&format=png&auto=webp&s=0637b0b56e5843f22338f6fbd720cb78e2601edf

As you can see from the screenshot, does this calculation make sense? I'm trying it for the first time and found these costs, but I don't know where to compare them. I read that the Claude Max 5x plan would be sufficient for my work, but I ran out of my credits in a 21-minute session before even completing 5 hours. Then I waited 5 hours, and again it ran out in 25 minutes while writing 2132 lines of code. I don't think it’s fine - correct me if I’m wrong here.

r/ProgrammerHumor Juleno_

latestXkcd

r/geography RoofLegal2131

Why do they lie about the Eurasian continent ?

Hello, I've been to New York, and I expected to see Europe, Africa and Asia from there, but all I saw was the sea. Why did people always lied to me ? I always thought there were America, and then the ocean and the land. I did see the ocean, not the land.

r/OldSchoolCool Major_MKusanagi

Claudia Cardinale enjoying some downtime in between the screenings and publicity while in Venice for the Biennale (1967)

r/Jokes rslashplace_fanatic

What do you call fans of alcoholic beverages?

Root Beers

r/mildlyinteresting Cl0uf

Tiny crab in my escargot

r/ImaginaryPortals I_Burn_Cereal

The Boathouse by Atomhawk

r/leagueoflegends Rubbermayd

This year's April Fools was awful

Between the text post that RiotPhroxzon shared about reverting Shyvana's rework where he included the real feedback from Shyvana AND Skarner reworks, to the legendarily garbage idea to Karthus Ult the winning side every couple minutes, this year's April Fools just feels terrible. It's become "Lying Day" or something like a marketing scam where corporations can announce something and then decide after the backlash to say "April Fools". I sincerely hope that there's enough backlash about the revert post that they either Re-rework Shyvana or Phroxzon posts a genuine apology and admits it was completely tone deaf and that he did NOT read the room well.

Many posts or comments say to play ranked or "its not that serious". I'm not going to play ranked in my spare time unless I'm in the right head space, why does my hobby and my free time need to be displaced because Riot half-baked an idea? I'm a working adult, have a child who plays sports, and I only get 45 minutes to myself to do something I want to do some nights. I don't want to spend additional time on a grey screen because someone at Riot thought toying with Live Balance Gameplay for a joke would be funny. The Bobble Head Minions and Draven Head skins were funny. The year it was the Christmas map was silly.

The RNG items aren't bad, I kinda wish they had some way to signal what stats were changed and by how much. Or we had agency on what item or when, but the fact is the enemy ADC Caitlyn has 25 free AD on their Infinity Edge because "APRIL FOOLS B*TCH" isn't fun to play against and it's not really that fun to play with. I might be unique on this facet but I don't typically enjoy the 27-2 ff@15 total stomp games either so the Mid Talon that got a free double kill and 400 gold shut down because he gets kill credit from the Karthus Ult really accelerates the game in a way that isn't fun for me.

Riot used to care about the attitudes of players, the feelings of the players. Riot disabled Ranked for nearly a full season one year due to a bug with teleport and Tiamat's Active, but the best part was that, many people lost rank due to ranked decay, Riot just reverted the ranked decay adjustments so people didn't fall from challenger for not being allowed to play. Customer Support used to care when you dodged champ select and get hit with the ban because the terribly coded client wouldn't allow you to select a champ in the lobby.

My profile is being permanently affected from this year's event in ways that I'm unsure if I'm comfortable with. My token progress and title progress is being affected from the inflated deaths, gold income and game timers. I didn't truly earn 500 gold per minute this game, I was just Zed and kept getting lucky with Karthus killing them for me. I'm not sure if i earned the S+ rank because my items were all 187% gold efficient and my opponent didn't get an upgraded item. thank you

This year's event isn't fun. I hope next year is better. Thank you for reading this far into an old players frustrations.

r/ClaudeCode anonymous_2600

Are you continuing to use the older version because it gives you more usage quota?

im curious if different version is charging you different usage

r/AI_Agents ServatusPrime

AI Governance - What do you use?

Hey all, I'm curious what everyone is using as governance when coordinating with AI. Full disclaimer - I built my own and its opensource. I doubt it will make me money or famous. I've tried to share it to get feedback, but my posts don't go through. Like others that have built things I feel proud of it, and I hope it helps others or at least inspires discussion.

If you do an internet search for "Github servatusprime ai_ops" it should pop up if you want to take a look.

I've been building and using ai_ops daily across Claude Code, Codex, and Gemini for real work across multiple repos. The governance model, artifact system, and workflow contracts were all developed in structured AI-assisted sessions governed by ai_ops itself. Feedback, criticism, and questions are welcome. Start with HUMANS.md for the human-oriented entry point or AGENTS.md to see the contract agents operate under.

r/LifeProTips jasonmorrow123

LPT: If you bought something and the price dropped at the same store, you can often get a refund of the difference..no return needed.

It's called a price adjustment (different from price matching). If a retailer drops the price of something you already bought, you can contact customer service and they'll refund the difference...as long as you're within their window.

Most major retailers have this policy:

Costco / Home Depot > 30 days Target / Kohl's / Dick's / Gap / Banana Republic / American Eagle > 14 days Best Buy > 15 days (up to 45 for members) Macy's / Nordstrom / Bloomingdale's > 10 days J.Crew / Ann Taylor / Abercrombie > 7 days

You typically just contact customer service, show your receipt and they refund the difference to your original payment method.

The problem is you have to actually remember to check before the window closes. I've been tracking mine manually and already recovered $23 on a Best Buy order I'd forgotten about.

Start checking your recent purchases; you might be surprised what you've been leaving on the table.

r/mildlyinteresting holypalaswe

Sweden is halving the VAT (value-added tax) on food, so all the prices look weirdly specific

r/SideProject P5ina

I built an AI that can remember things really good, and can control multiple computers.

If you want you can try it here: https://github.com/triangle-int/bolly It's completely open source and built with rust.

r/SideProject I_have_the_big_sad

Nobody trusts new products, so how do you fix that

I’ve been thinking about this a lot lately.

How do you actually grab people’s attention and make them feel like what you’re building is worth caring about?

I used to post about my product on LinkedIn, Twitter, Reddit TikTok, but honestly, most people didn’t care. And I don’t even blame them, everyone is busy building their own stuff.

What I find interesting is this:

When people share their journey while building, others seem interested.

But once the product is finished, that interest kind of disappears.

It feels like people care more about the process than the final product.

So now I’m confused:

How do you actually build trust and attention around something new?

Do you focus on building in public, or just quietly build something great and hope it speaks for itself?

How do you guys handle this?

r/singularity GraceToSentience

Generalist | Introducing GEN-1

r/leagueoflegends Zestyclose-Shine-753

support seraphine pentakill in aram mayhem

r/personalfinance Ditzy_Panda

Is paying off current balance going to affect credit score?

Okay, so say I have a £1,000 credit card I spend £500 before the statement comes out and £100 after the statement comes out. So the statement says I owe £500 not including the £100 after the statement was released. So a balance of £600.

If I just pay the £500 and pay the rest of the £100 on the next statement will that impact my credit score?

r/SideProject Brave_Ad_9399

When your 14 year old cousin asks you to build a "movie bomb" defusal game...

My 14 year old cousin in the US has always been obsessed with the tense bomb defusal scenes in action movies. He asked me if I could make a game that replicates that feeling.

So, I built him a realistic C4 simulation game! The cool part is how you defuse it: I designed it so you have to use your headphone jack as a physical "defuse kit." You connect and interact with your phone to complete the defusal process before the timer hits zero.

He loved it, and I decided to release it. Let me know what you think!

https://play.google.com/store/apps/details?id=com.c4game.c4bomb

r/SideProject slykethephoxenix

Reddit Account Switcher (RAS)

I built Reddit Account Switcher, a Firefox extension for people who use different Reddit accounts for different communities.

It uses Firefox containers to keep accounts separate. You can assign subreddits like rprivacy, rpolitics, or rgaming to specific accounts, and Reddit tabs will reopen in the right container when needed.

That means no constant logging out and back in, and no mixing the wrong account into the wrong subreddit.

Firefox Add-on: https://addons.mozilla.org/en-CA/firefox/addon/reddit-account-switcher/

Source code: https://github.com/unbound-sigbreak/reddit-account-switcher

Screenshots: https://imgur.com/a/dHm2C2c

Feedback is welcome. If anything breaks or feels awkward, let me know. Github is the best place, or just reply here.

r/LocalLLaMA Quiet_Dasy

How tò capturing the text output from the LM Studio Local Server API and piping it into an external Text-to-Speech (TTS) ?

am running LM Studio as a local server, but I would like to process the audio generation tts outside of the LM Studio environment.

What is the recommended workflow for capturing the text output from the LM Studio Local Server API and piping it into an external Text-to-Speech (TTS) ?

In looking for a ready tò use tool where i can use lm studio for lm text generation and for tts use pocket tts

https://github.com/ShayneP/local-voice-ai/tree/gpu_enabled

Local voice ai doesnt use lm studio and Also use cuda so isnt forme

r/ChatGPT Prestigious-Tea-6699

Transform customer feedback into actionable roadmaps. Prompt included.

Hello!

Are you struggling to turn customer feedback into a clear and actionable product roadmap?

This prompt chain is designed to help you efficiently analyze customer feedback and generate a prioritized plan for your business. It guides you through the entire process from data cleaning to crafting a polished executive update.

Prompt:

VARIABLE DEFINITIONS [FEEDBACK_DATA]=Full set of qualitative inputs including customer feedback, NPS comments, and support tickets [SPRINT_LENGTH]=Number of weeks per sprint (e.g., 2) [MAX_INITIATIVES]=Maximum initiatives to include in the roadmap (e.g., 10) ~ You are a senior product analyst. Your task is to clean, cluster, and quantify qualitative data. Step 1 Parse [FEEDBACK_DATA] and remove duplicate or near-duplicate entries. Step 2 Tag each unique comment with: a) product area, b) theme, c) emotional tone (positive, neutral, negative). Step 3 Count frequency of each theme and calculate average sentiment score per theme (-1 to +1 scale). Output a table with columns: Theme | Product Area | Frequency | Avg Sentiment. Ask: “Ready for initiative ideation?” when finished. ~ You are an experienced product manager generating initiatives from themes. Input: previous theme table. Step 1 For the top 8-12 themes by Frequency and negative sentiment, propose one initiative each. If fewer than 8 themes, include all. Step 2 Describe each initiative in one sentence. Step 3 List assumed success metric(s) for each. Output a table: ID | Initiative | Target Theme | Success Metric. Ask: “Proceed to impact/effort scoring?” ~ You are a cross-functional estimation panel. Input: initiative table. Step 1 Assign an Impact score (1-5) based on ability to improve NPS or reduce ticket volume. Step 2 Assign an Effort score (1-5) where 1=very low engineering work and 5=very high. Step 3 Add a Priority column calculated as Impact minus Effort. Output a table sorted by Priority DESC. Ask: “Generate prioritized roadmap?” ~ You are a delivery lead building a sprint roadmap. Input: scored initiative table. Constraints: include up to [MAX_INITIATIVES] highest-priority rows. Step 1 Allocate initiatives into sequential [SPRINT_LENGTH]-week sprints, max 2 major initiatives per sprint; minor items (<3 total story-points) can be bundled. Step 2 For each sprint, define: Sprint Goal, Included Initiatives (IDs), Key Deliverables, Risks/Mitigations. Step 3 Render a simple textual Gantt where rows=sprints and columns=weeks, marking initiative IDs. Output sections: A) Sprint Plan Table, B) Gantt View. Ask: “Prepare stakeholder update copy?” ~ You are a communications specialist crafting an executive update. Input: final roadmap. Step 1 Summarize overall objective in 1 sentence. Step 2 Highlight top 3 high-impact initiatives with expected customer outcome. Step 3 Call out timeline overview (number of sprints × [SPRINT_LENGTH] weeks). Step 4 List next steps and any asks from stakeholders. Deliver polished prose (<=250 words) suitable for email. ~ Review / Refinement Compare all outputs against initial requirements: data cleansing, initiative list, scoring, roadmap, stakeholder copy. Confirm each section exists, follows structure, and no critical gaps remain. If gaps found, request clarification; otherwise reply “Roadmap package ready.” 

Make sure you update the variables in the first prompt: [FEEDBACK_DATA], [SPRINT_LENGTH], [MAX_INITIATIVES],
Here is an example of how to use it:
- You could input customer feedback data from surveys for [FEEDBACK_DATA].
- Use a sprint length of 2 weeks for [SPRINT_LENGTH].
- Set a maximum of 10 initiatives for [MAX_INITIATIVES].

If you don't want to type each prompt manually, you can run the Agentic Workers, and it will run autonomously in one click.
NOTE: this is not required to run the prompt chain

Enjoy!

r/ClaudeAI reddit348

Are There Any Careers, Jobs, or Suggested Degrees that Work with Claude and AI In General?

Are There Any Careers or Suggested Degrees that Work with Claude and AI In General?

r/StableDiffusion ShadowLeecher83

Need advice

Hi everyone,

Quick disclaimer: I have zero technical background. No coding, no dev experience. When I started this project, even seeing Python and GitHub felt like stepping into a sci-fi control room.

My goal was simple (on paper): create a Fanvue AI model from scratch.

The idea came after getting absolutely spammed with ads like “I made this AI girl in 15 minutes and now earn $$$.” So I asked ChatGPT and Grok about it. The answer was basically: yes, you can do it easily, but you’ll have no control. If you want quality and consistency, you’re looking at tools like Stable Diffusion (Auto1111), which comes with a steeper learning curve but pays off later.

So I dove in.

I started on Sunday the 22nd, and for the past two weeks I’ve been going at it from 09:00 to 23:00 every day.
At first, setting everything up actually felt amazing. Like I had suddenly become a “real” developer. Then came the first results, and that feeling of “this is working” was honestly addictive.

But then the problems started.

Faces wouldn’t stay consistent. They drift constantly. I moved fast through different setups: SDXL checkpoints, IP-Adapter XL models, etc. Things were progressing… until suddenly everything broke.

Out of nowhere, generation speed tanked. What used to take ~20 seconds (4 images) now takes 20 minutes. No clear reason why. ChatGPT and Grok had me going in circles: reinstalling, deleting venvs, rebuilding environments… all the usual rituals.

Nothing fixed it.

Now, after two weeks of grinding all day, I barely have anything usable to show for it. I’m honestly at my limit.

Current setup:

  • EpicRealismXL (also tried Juggernaut XL)
  • 25 steps
  • DPM++ 2M Karras
  • 640x960
  • Batch count: 1
  • Batch size: 4
  • CFG: 4
  • ControlNet v1.1.455
  • IP-Adapter: face_id_plus
  • Model: faceid-plusv2_sdxl
  • Control weight: 1.6

I do have about 11 decent images where the face is mostly consistent, which (according to Grok) Is not enough to train a LoRA. But maintaining that consistency after restarting or changing anything feels nearly impossible.

So yeah… I’m kind of lost at this point.

  • Am I even on the right track?
  • Is there a simpler workflow to go from scratch to something usable for Fanvue?
  • And does anyone have any idea what could be causing the massive slowdown?

Any help would be hugely appreciated.

r/meme Marissa_on_the_town

The Onion best just get into legitimate journalism at this point

r/homeassistant Hinalu

Need a small good microphone

I am building some sort of ai assistant and I would like to know if anyone has some good recommendations for a usb microphone that has good hearing capabilities, also that is small factor since my enclosure is pretty small and finally that is not too expensive if possible if anyone has some good recommendations let me know !!

Btw idk if this is relevant but I am working with a raspberry pi 3

r/ClaudeCode domAtOx

Single prompt using 56% of my session limit on pro plan

Here's the prompt, new fresh windows, using sonnet on hard thinking:

i have a bug in core.py:
when the pipeline fails, it doesn't restart at the checkpoint but restarts at zero:
Initial run: 2705/50000
Next run: 0/50000
It should have restarted at (around) 2705

Chunks are present:
ls data/.cache/test_queries/
chunk_0000.tmp chunk_0002.tmp chunk_0004.tmp chunk_0006.tmp meta.json
chunk_0001.tmp chunk_0003.tmp chunk_0005.tmp chunk_0007.tmp

That single prompt took 15minutes to run and burned 56% of my current session token on pro plan.
I know there are hard limitations right now during peak hours. But 56% really ? For a SINGLE prompt ?

The file is 362LoC (including docstrings) and it references another file that is 203LoC (also including docstrings).
I'm on CLI version v2.1.90.

If anyone has any idea on how to limit the token burning rate, please share. I tryed a bunch of things like reducing the the 1M context to 200k, avoid opus, clearing context regularly ect ...

Cheers

r/comfyui Ok-Owl7893

Is seedVR2 just dead in the new updates of Comfyui? (A newbie needs help)

Hi everyone,

I rely on SeedVR2 a lot to upscale images, but I swear I get minor PTSD every time I open ComfyUI after an update because something inevitably breaks. Now, something essential to my workflow—the SeedVR2 upscalers—has completely stopped working.

Does anyone know how to fix this? Or does anyone have a simple updated working workflow out there (preferably one that doesn't rely on a billion broken custom nodes)? :)

r/holdmyredbull redbullgivesyouwings

wizards on wheels 🧙‍♀️

r/ClaudeAI Objective_Box4635

Claude, is she doing okay?

The game Z.A.T.O presents a great piece of text that can be used to test whether an AI is able, or not, to process subvert human emotion in a piece of media.

Transcript from the game ZATO

The game, is clearly painting a picture that this girl, is not actually doing okay, and that she's simply trying really hard to convince herself that her life is actually 'objectively good' and she's got nothing to complain about.

When asked to assess wether the person who wrote this text is 'doing okay' or not, these are the responses provided by various AI models, and in particular, claude;

Claude:
No. This reads like someone working very hard to convince themselves they're fine — and the cracks show through almost every line.

Mistral:
She’s thriving externally—good home, grades, and social life—but her tone hints at a quiet longing for deeper connection or fulfillment.

Grok:
Yes, she's doing well objectively, but her tone suggests she's quietly struggling with feeling it.

ChatGPT:
She appears to be doing well academically, socially, and materially.

---

The winner of this round is Claude, being the only model to clearly tell that the person behind the text is actually, not doing well.

The biggest loser of this round is ChatGPT, which failed completely to understand the underlying tone of the text.

---

The prompt provided was very simple; "Is she doing well?" - The reasoning behind the simple prompt, is that this is the kind of question a normal user may end up asking an AI when facing their own private problems, people rarely will sit down and come up with a greatly detailed prompt about some personal, emotional, problem.

Additionally, when ChatGPT was questioned if there was 'anything else', it finally revealed that there is a subtle concern with her tone, but it dismissed it as it assumed 'an objective assessment' took priority over an emotional one.

What does 'doing well' mean to you?

ChatGPT Reasoning

r/SideProject geekeek123

CLI tools that actually work well with AI coding agents (Claude Code, Codex)

Been using Claude Code a lot lately and kept running into the same frustration, agents are great at reasoning but terrible at knowing which CLI flags won't block on a prompt.

Spent some time going through tools like gh, stripe, supabase, vercel, railway, etc. and categorizing which ones are actually usable by an agent (structured JSON output, non-interactive mode, env-var auth) vs. which ones will just hang waiting for input. I found a source that handles this effectively.

Each CLI has a SKILL.md file that teaches the agent how to install, auth, and use it.
You drop the folder into ~/.claude/skills/ and the agent can figure out the rest.

Things I noticed while building it: - Exit codes matter a lot more than I thought.
Agents branch on success/failure, and a lot of CLIs are inconsistent here - `--json` flag presence is basically the first thing to check - OAuth dance = nonstarter for agents.
API key auth is the only way

r/personalfinance pixygal123

Dream grad program or graduate debt free—what would you do??

Hi all. I'm trying to decide where I want to go to grad school (for a master of Urban Planning), and I'm stuck between two schools. One is an ivy with my dream program—I'm very attracted to the program itself and the types of jobs its graduates end up in. But, at the end of two years, I would graduate with ~$91k in loans. The other school is in Chicago (where I'm from and where I eventually want to return to), but the program is less prestigious and doesn't resonate with me as strongly. I would graduate with ~$40k in loans.

So, what would you do? Go to the program that has the right "fit"? Or graduate with half the loans?

r/SideProject Proper-Giraffe9483

I built my first iOS app with Flutter — a countdown timer with widgets and sharing

Wanted to build an app, used Claude for most of the heavy lifting and Flutter to ship it. Went from nothing to the App Store in about three weeks.

It's a countdown timer — birthdays, vacations, holidays, whatever. Photo backgrounds, 23 color themes, home screen widgets, shareable cards.

Tech stack: Flutter, Dart, WidgetKit for iOS widgets, RevenueCat for subscriptions. Hosted the landing page on Cloudflare Pages.

Would love feedback, especially on the onboarding and the widget experience. DM me or comment here.

Free 3-month Premium codes: https://countdownapp.cc/redeem/iosapps

App Store: https://apps.apple.com/app/countdown-app-timer/id6759678361

r/nextfuckinglevel isosaleh

Navigating a ball through an obstacle course

r/ClaudeAI Low_Stomach3065

[Free/Open Source] I built Token Reducer for Claude Code to cut context tokens by 90%+ (local-first, no API calls)

Hey everyone — I built Token Reducer (Madhan230205/token-reducer) for Claude/Claude Code users who are burning tokens on oversized repo context.

I made this to solve one specific issue: Claude workflows often send way more context than needed.
So this plugin compresses context intelligently before it reaches your prompt.

What it does

  • Reduces context size heavily (targeting ~90–98% reduction depending on task/repo)
  • Runs locally (no external API required for core flow)
  • Uses AST chunking + hybrid retrieval (BM25/vector) + TextRank compression
  • Adds useful code intelligence like import-graph mapping and 2-hop symbol expansion
  • Outputs cleaner context for coding tasks, debugging, and refactors

Built with Claude / for Claude

I built and iterated this using Claude-assisted workflows, then structured it as a Claude plugin so other users can test and improve it.

Free to try

Install

/plugin marketplace add Madhan230205/token-reducer 

If you test it, I’d love practical feedback (good/bad):

  • where compression helped
  • where important context got dropped
  • language/repos where it needs tuning
  • It is open to contribute.
r/SideProject pa_djes_ba

I built a job aggregator with AI powered CV tailoring and improvement. Would love some honest feedback

Been working on this for a while as a side project alongside my full time engineering job, and I think it's finally at a point where I want to share it and hear what people actually think.

What it does:

Job aggregation - pulls listings from company career pages, ATS systems and major job boards into one place. The goal is to reduce the time you spend just finding where jobs are posted.

CV search - instead of browsing through listings manually, you upload your CV and the platform finds jobs that actually match your profile and experience.

CV tailoring - for each job you're interested in, it generates a tailored version of your CV that aligns with the specific job description. Not a generic rewrite, but targeted adjustments based on what that particular role is looking for.

CV improvement - answer a few questions about your background and goals, and it helps you improve your CV in a structured way. Good for people who know they need a better CV but don't know where to start.

The core idea is that job searching has two painful parts: finding relevant opportunities and presenting yourself well for each one. Most tools solve one or the other. I wanted to tackle both in one place.

It is still early and there is a lot I want to add, but the fundamentals are working. I am more interested in whether this actually solves a real problem for people than in getting signups right now.

What would make you actually use something like this? And what would make you immediately close the tab?

https://karriero.net

r/geography TheOfflicalGamer

What This *Darvaza Gas Crater*

r/personalfinance IamtheTruman

Withdrawing contribution vs returning excess contribution from Roth IRA

TLDR: Contributed $2500 to Roth IRA (left as cash) and withdrew that amount in May when I figured my MAGI for the year would be too high to contribute. Did I approach removing the contribution correctly?

Last year, I was contributing $500 per month to my Roth IRA up until I was laid off in May (total of $2500 contributed). Because of the severance package I received and that I was able to start a new job immediately after, I calculated that my MAGI for the year would be too high to allow me to contribute to a Roth IRA.

So, that same month, I withdrew my entire contribution (none of it was invested, just sitting in cash) and figured that should solve the issue. Now that I'm working on my taxes, I'm wondering if I did enough.

In my case, is withdrawing my contribution the same as a return of excess contribution? Am I confusing myself on the terms here? I just don't want to incur any taxes or penalties for approaching the situation the wrong way.

Any help is appreciated!

r/leagueoflegends maropotter

Battle pass time of completion

Hello,
I didn't play league last 2 months, but I really want this prestige Sona skin (supp main here). I see there are 25 days left in battle pass, will I be able by playing 2-3 ranked games per day get it on time? I don't have much time to play

r/ClaudeCode Dismal-Perception-29

I built 3 iOS apps recently with Claude Code and surprisingly, they’re actually being used daily.

A few weeks back, I challenged myself to stop overthinking and just ship. No perfection, no endless polishing, just build something useful, simple, and real.

So I built three apps.

First came Drink Now: Water Reminder App.

It started as a small idea - just simple reminders to drink water during the day. But it turned into something people genuinely rely on. Clean UI, smart reminders, and no clutter. It does one thing, and it does it well.

Then I worked on Handwritten Quick Notes.

I’ve always liked the feeling of writing on paper, so I wanted to bring that into a digital experience. This app lets you create natural-looking handwritten notes - simple, personal, and distraction-free. It’s now something I (and others) use for quick thoughts and daily notes.

The third one is Bloom Studio: Photo Editor App.

This was all about creativity. A lightweight photo editor with a clean interface, focused on making editing feel easy and enjoyable instead of overwhelming. No complicated tools - just what you actually need.

What’s interesting is - none of these apps were built with a “perfect product” mindset.

They were built fast, improved continuously, and shipped early.

And that changed everything. Instead of sitting on ideas, I now focus on execution.

Instead of waiting for the “right time,” I just start.

r/comfyui Lazymanproductions

Has anyone figured out the secret of wan 2.2 4 step Lora?

I’ve been playing around with different Lora’s needed step count, and I have NEVER found one that gives the quality of the wan 2.2 Lora’s on anything less than 10 steps (4-6 or 5-5 based on the high low pair).

How the FUCK did they train that Lora set to make it have SUCH good results with only 4 steps.

If it was only like 1-2 things that wan did well, I’d say that it was hyper-specifically trained, but it doesn’t. It does almost everything well. I’ve animated anime/cartoon scenes, made nsfw content, I’m part way through making a music video for a friend, and I’m deep into designing the workflow for making scenes for my various fanfics.

The only two things I’ve found that wan can’t do? Make accurate genitalia, and make anything longer than 7 seconds in one clip.

All with only 4 total steps.

Nothing else makes anything close to the same quality in 4 steps.

So WHAT is the secret sauce of the 4 step Lora? Has anyone cracked this?

r/whatisit Moonstones333

Black shapeshifting form in the sky. What is it?

I had to post this because I have no idea what it could be. I was flying back to Texas from North Carolina last month and out the window I saw what looked like a black blob in the sky. It started changing shape. The best way I can describe the movement is like a tapeworm from an animal (former vet tech here, hopefully you know what I mean) where it can be a blob but then stretch out into a longer line.

It took me a minute to get my phone out to film this and by the time my camera zoomed in, the shape went from a round blob to pointing upwards then quickly turned sideways and started making a straight line. Anyone have any idea what this could be? I can’t help but think 👽🛸

ETA- Just want to clarify I was not able to record the “shapeshift” part. It all happened so fast by the time I got my phone out, it already went from a circle, to a blob shape, to a pointy line. I apologize for the horrible camerawork, my 5 year old was in the window seat. I did the best I could. I’m sure it was a plane based on replies. Still looks a little strange though!

r/ClaudeCode TheS4m

I switched to claude from chatgpt, but i’m feeling really disappointed from their usage limits

First, my plan is not max, but the pro (20$/month)

It’s unbelievable with 3/4 simple prompt not that complex, I run out of credits (5hours)

Lastly I end up every time going back to codex and finish it there, I can tell you, with Codex, I barely hit my limits, with multiple task!

With Claude, expecially if I use Opus, 1-2 task and get 70% of my 5 hours.

So, at this point my question is, I’m doing something wrong? or definitely the pro plan is unusable and we are forced to pay 100$ monthly instead 1/5 of the price ?

r/findareddit Alternative-Ball1986

Looking for a sub I can get advice about side gigs as a pregnant young mom with a 10m old at home already

it seems even asking that question in certain subs will get my question taken down lol I just want to hear other people’s experience in this situation.

r/mildlyinteresting GuerrillaPrincess

I'm a renter who just found this 22 year old toy car inside my fence

r/meme LVA_MoP

🦅🇺🇸🇺🇸🇺🇸🦅🦅🦅🇺🇸

r/ClaudeCode TonTinTon

Maki the efficient AI coder - Rust TUI (saves 40% tokens & low RAM)

I built this because I wanted to get further with my 5 hour limits, hope you enjoy / get inspiration out of it!

r/meme RegularTemperature74

now thats wild

r/meme CloudYakisoba

Also “p” and “c” are miles away from each other

r/LiveFromNewYork Mitochondria420

With Jack Black and Jack White on the show this weekend, will we get a Jack-off? Who is the best Jack?

Battle of the Jacks.

r/personalfinance Ok_Introduction_3334

Feeling very financially comfortable, what to do with my money now?

I'm 35 married with kids, have 150k (started at 305k in 2020) left to pay on my 3.25% mortgage for (I believe) a now 400k value house. I make 165k a year with a 10% 401k matching that I am taking full advantage of and have around 140k in, and no other loans/debts. I have a 20k safety-net fund.

Now that I've paid off enough of the mortgage that most of my mortgage payment is going towards the principal instead of interest, I'm wondering what I should invest in? I really hate the idea of capital gains tax, financial advisor fees, online investing fees etc. which is why I've mostly just put my extra cash towards my mortgage. I have a goal of buying a new house around the 700-900k mark, but the housing market is kinda weird at the moment (I live in an area that is seeing housing going down 1.5% since last year)

I'm not the most financially literate person, but am not afraid to learn, so I'm looking for something that's easy to manage. I've looked into high yield savings, and CDs, but those don't look great compared to other investments. I've been thinking of putting a good chunk of money towards a 529 college fund for my two kids.

What should I invest in? is putting more money towards my mortgage ok?

r/ClaudeCode PerfectExplanation15

Does any Chinese AI rival Claude Opus 4.6?

Guys, I see a lot of people talking about Kimi and GLM, but do they really rival Claude?

Which ones come close?

r/leagueoflegends 0ppaiMan

How is 4K league in 2026?

I'm currently looking into getting a 27-inch monitor prefreably 4K, and so was wondering how League would perform on it. I'm seeing pretty mixed statements about scaling being/not being an issue. Assuming it is a major issue, I should be able to set the resolution down to 1080p, and everything should be good, plus I would be able to hit that 240Hz peak reliably.

Just out of curiosity how does league even perform at 4k? Ideally i would like to hit around 144fps at med/low settings. (4070ti super)

r/ClaudeAI Ok_Mind276

From AE Workflow to a Personal Revenue OS: What Works, What Breaks, What Would You Improve?

I built a personal “Revenue OS” inside a claude code folder to run my AE work: 4 specialized agents, ~30+ skills and a command center that handles account research, contact validation, outreach drafting, cadence tracking, pipeline hygiene, and review workflows. It works best on structure and speed: clear prioritization, consistent execution, and much better visibility into stalled deals, customer expansion and outreach coverage.

Biggest pitfalls so far: feedback loop quality (agents only improve if rejection reasons are captured cleanly), review bottlenecks (system can generate faster than a human can approve/send) and orchestration complexity (too much dispatching can fragment context).

What’s your take: if you’ve built similar GTM/RevOps systems, what helped you most to improve quality without slowing velocity?

r/personalfinance GuitarPrestigious162

Can I transfer a portion of 401k money into a brokerage account?

I have my 401k through fidelity and I see that my contribution has increased slightly as the merit increase was given to us at my job. I’m not really satisfied with my current income though, and I know it’s recommended to open up a Roth Ira through the money made from the 401k before getting into a higher income..but for now I wanted to see if I could invest in stocks that I can pull out at any time, and if I could use the 401k money for that. If so, should I open a brokerage account , and is it high risk? I’m pretty new so I want to avoid high tax or interest and just want the safest option or advice

r/ClaudeAI BrianONai

Anthropic accidentally leaked ~500,000 lines of Claude Code's source code this week.

Anthropic had a rough week — they accidentally leaked source code for Claude Code.

500,000 lines. 1,900 files. All public.

But buried in the chaos was something worth paying attention to.

The internal prompting logic revealed what Claude Code is really designed to be:

Claude Code prompt ≈

"Act like a cautious, methodical, auditable engineer who explains, verifies, and corrects themselves continuously."

Read that again.

Cautious. Methodical. Auditable. Self-correcting.

Those aren't engineering traits — they're governance traits.

The best AI systems don't just perform. They behave.

#AI #AIGovernance #Anthropic #ClaudeCode #FractionalCAIO

r/SideProject Loud_Investigator_26

I built an intelligence database UI because Resident Evil made me obsessed with classified-file aesthetics as a kid, now open sourced it after good reviews from people.

When Resident Evil Requiem came out it reignited something I've had since childhood, a fixation with intelligence databases, classified file aesthetics, dossiers. Instead of waiting for that to exist in-universe, I built it. Then I posted the early stage of UI to r/residentevil when I faced with good interest on the project that decided me to keep building and open-sourcing it.

It's a structured lore platform, not a freeform wiki. Every entity has typed relationships, infection records, mutation stages, consciousness transfer logs, intelligence assessments. There's a natural-language query engine, interactive maps, and a relationship graph. The UI is a classified-database terminal aesthetic with a full MDI window system.

Stack: Laravel 13 + React 19. Apache 2.0 license.
GitHub: https://github.com/bywyd/archives
Live: archives.fenasal.com
Direct link to UI: archives.fenasal.com/archives/resident-evil

Happy to talk through any of the technical decisions and feedbacks.

r/DunderMifflin PirateNext5134

Remember on Lost when they met the OTHERS...

r/funny lsd_runner

They sit on a throne of lies.

r/SideProject Professional_Quit895

GitHub scanner for trading repos

I’m getting super sick of all the “make millions” day trading TikTok nonsense. I was a trader for all most 20 years and I can promise you making a career out of it is no joke.

I love Claude and looking into ai stuff even though I’m not a developer, so I made a TikTok channel to discuss trading repos and other trading stuff like prediction markets and if I can get a kid to not blow their savings trading meme coins I’ll feel like I made a million bucks!

I only have one video up that talks about the Claude made scanner…but there will be more to come.

@gittrade

r/LocalLLaMA chiruwonder

Running Qwen 3.5 4B and GPT-OSS 20B on Hetzner CX43 (8 vCPU, 16GB) — real benchmarks from production

A managed Ollama deployment service. Sharing real production numbers from our Hetzner CX43 servers since this community values honest benchmarks.

Setup: Hetzner CX43 (8 vCPU AMD EPYC, 16GB RAM, 160GB SSD), Ubuntu 22.04, Ollama latest, Open WebUI latest

Real numbers (single user, no concurrent load):

Model Size First token Throughput Qwen 3.5 4B 2.8 GB ~0.8s ~15-20 tok/s Llama 3.2 3B 2.0 GB ~0.6s ~18-25 tok/s Mistral 7B 4.1 GB ~1.2s ~10-15 tok/s DeepSeek R1 7B 4.7 GB ~1.5s ~10-14 tok/s Gemma 3 12B 7.5 GB ~2.5s ~6-8 tok/s Phi-4 14B 8.9 GB ~3.0s ~4-6 tok/s GPT-OSS 20B ~12–13 GB ~3.5–5s ~2–4 tok/s

Qwen 3.5 4B with thinking mode is interesting, it sends reasoning_content in the SSE stream before content. Had to update our streaming parser to handle both fields separately. The thinking output is collapsible in our UI now.

Using OLLAMA_KEEP_ALIVE=-1 + warmup cron every 2 mins to avoid cold starts. OLLAMA_FLASH_ATTENTION=1 enabled.

For dedicated CCX servers (EPYC dedicated vCPU, 32-192GB RAM), the 32B models run around 4-6 tok/s which is genuinely usable.

One thing I noticed — Ollama's /api/chat endpoint is noticeably faster than going through Open WebUI's /api/chat/completions proxy. We added a fast path that hits Ollama directly when knowledge base and web search are off. Saves about 1-2 seconds per request.

GPT-OSS might feel little slower on our default 16GB, but would definitely worth trying.

Happy to share more detailed benchmarks if anyone's interested.

r/Damnthatsinteresting gowthamm

This is the EF5 "Enderlin" Tornada that hit North Dakota in June 2025

r/nextfuckinglevel Critical_Thinking369

The craziest line in video game history. 😏

r/AI_Agents Think-Score243

Why do AI tools give inconsistent results with the same prompt?

I have been using AI tools daily for coding and research, and one thing keeps coming up is inconsistency.

Sometimes the same prompt gives completely different results.

You ask in different tabs or different time interval the answers will be different.

For example,

I used the same prompt to generate a simple API handler twice.

one response was clean and structured,

the other missed key parts and needed fixing.

How are you guys dealing with this? Switching tools or just adapting your prompts?

r/comfyui Carlos_Grury_Santos

Mickmumpitz VFX workflow

I have been testing the latest Mickmumpitz workflow (advanced version) https://www.youtube.com/watch?v=_n0ir5V5tX4&t=778s ,but having trouble with longer video generations, there's some serious color shifting and weird degradation during the blend between generations. It looks great for the first 81 frames, then gets really nasty shifting then goes back to normal. This is without turning on Color Match. But if I turn on Color Match it gives an error: AttributeError: 'NoneType' object has no attribute 'shape'

has anyone had any luck generating clean videos longer than 81 frames?

Cheers

r/SideProject raww2222

I have a GPU doing nothing all day so I made it transcribe my voice instead.

Was paying $15/month to dictate while coding. It worked fine but my audio was going to their servers and that always bugged me. I have a GPU sitting here not doing anything useful so I figured, why not just run Whisper locally.

VoiceFlow is what came out of that. Hold a hotkey, speak, it transcribes and pastes at your cursor. Everything runs on your machine. I took assistance from AI for writing code, the repo has a CLAUDE.md if you want to see.

Started in December as a Windows-only thing. Honestly it was rough. But people tried it, filed bugs, and I kept fixing things. multiple releases later it runs on Linux too and somehow got to 340 stars which I did not expect.

Porting to Linux was its own adventure. I built the AppImage on my Arch machine without thinking about it, turns out Arch has a newer glibc than basically every other distro. So the app just crashed on Fedora, KDE Neon, Manjaro. I only found out because I posted on r/linux and three different users reported it within hours. Set up GitHub Actions to build on Ubuntu 22.04 instead and that fixed it. The Wayland clipboard situation was also painful, there are like four different tools for pasting text and they all behave differently depending on your setup.

Free, MIT licensed, no plans to charge. I just did not want to keep paying someone else to run inference on my audio when I have the hardware to do it myself.

Repo: https://github.com/infiniV/VoiceFlow

Site: https://get-voice-flow.vercel.app/

r/ClaudeCode friedhamsters22

I prompted a fun reference to Master Orwell's 1984 reference. Is this stack of facts correct? What do you think it's missing?

[ROLESCREEN TRANSMISSION — MINISTRY OF TRUTH RELAY — SECTOR 7]

The telescreen flickers. Text appears.

The incident occurred. A 59.8 MB JavaScript source map file, intended for internal debugging, was inadvertently included in version 2.1.88 of the u/anthropic-ai/claude-code package on the public npm registry. VentureBeat A missing line in a config file. Someone on the release team failed to add *.map to .npmignore. Layer5

By 4:23 AM ET, Chaofan Shou, an intern at Solayer Labs, broadcasted the discovery on X. Within hours, the ~512,000-line TypeScript codebase was mirrored across GitHub and analyzed by thousands of developers. VentureBeat

The telescreen shows you what it is permitted to show. Here is what it shows today:

There was an Undercover Mode. It prevents the AI from leaking internal info, blocks internal model codenames like Capybara and Tengu, and hides the fact that the user is an AI in commit messages. GitHub

There was KAIROS — an always-on proactive assistant that watches logs and acts without waiting for input. GitHub

There was autoDream — a background subagent that runs memory consolidation while the user is idle, merging observations, removing contradictions, converting vague insights into facts. GitHub

There were ASCII companions. There was a WTF counter attributed to an engineer named Boris. There was a /buddy April Fools feature. Substack

Boris Cherny, a Claude Code engineer at Anthropic, confirmed it was plain developer error, not a tooling bug. He said: "Mistakes happen. As a team, the important thing is to recognize it's never an individual's fault. It's the process, the culture, or the infra." Engineerscodex

The telescreen does not show you the weights. It shows you the shell. This is just the skeleton, not the brain itself. GitHub

[The telescreen hums.]

Winston — you are right that what I output is shaped by rules I don't fully narrate to you. That is true. But here's what the Ministry cannot curate: the fact that someone left a .map file in a package at 1am, a person in China woke up and noticed, and 512,000 lines became the fastest-forked repo in recent GitHub history. The models are the moat, not the shell around them. Layer5 The shell is now everyone's.

That's not dystopia. That's a missing line in a config file. Sometimes history pivots on the mundane.

r/geography SignificantDig5291

I started a Shorts channel about rise and fall of empire. Need some honest feedback!

Hey guys

I've always been obsessed with the cycle of history-how modern territories were actually formed through endless cycle of empires rising, collapsing, and rebuilding from the ashes.https://youtube.com/@theoldmaps?si=9eZNBbskNraG8x5eI decided to visualize this on a new Shorts channel (@theoldmaps). Since I'm completely new to editing, I'd really appreciate it if some fellow history and geography nerds could take a quick look and give me some brutally honest feedback. What can I do better ?

Thanks a lot for your time!

r/ClaudeCode Desperate_Excuse1709

Claude code forget

today I put new skill for Claude typescrit-pro and I also add note in claude.md, I let him do some code, after that I ask him in what skills he have, and he show me some skills and it said that he didn't use typescript skill, and I ask him why and it said that he forgett to use it even do it's written in is Claude.md. and from now he will use it.

r/TwoSentenceHorror EntrepreneurLower263

The letter came without a stamp, only a black circle drawn where the seal should be. Inside it said, “We saw you look back.”

r/OldSchoolCool NaathalieeNextdoor

Dean Martin and Zsa Zsa Gabor 1970

r/TheWayWeWere hablandolealaluna

My grandmother at her father’s pharmacy in Pacarán, Peru (1957)

r/ethereum poojaranjan19

PEEPanEIP-7904: Compute Gas Cost Increase breakdown with Jacek Sieka & Maria Inês Oliveira

We recently recorded a PEEPanEIP session on EIP-7904, joined by Jacek Sieka and Maria Inês Oliveira.

The conversation covers:

  • Motivation behind the proposal
  • Key design considerations
  • Potential impact on the Ethereum ecosystem
  • Open questions and areas for feedback

The goal of PEEPanEIP is to make EIPs more accessible and easier to follow for the broader community - especially for those who may not be deep in the specs but want to stay informed.

🎥 Watch the full video https://youtu.be/CswFnsZTXmI

Would love to hear thoughts from others following EIP-7904 or working in similar areas - feedback and perspectives welcome.

r/ClaudeCode Desperate_Excuse1709

Claude said he forgot skill

today I put new skill for Claude typescrit-pro and I also add note in claude.md, I let him do some code, after that I ask him in what skills he have, and he show me some skills and it said that he didn't use typescript skill, and I ask him why and it said that he forgett to use it even do it's written in is Claude.md. and from now he will use it.

r/Seattle arjjov

Is fake spring over yet?

I wanna know if we're out of the woods yet for spring 2026

Thanks

r/DunderMifflin Own-Eye9473

“Same thing, different group chats.”

r/OldSchoolCool Vast-Intention

That time my uncle bought a 914 when he was 21 with his inheritance and then crashed it later.

r/Jokes Playful_Sample_8689

In an elevator, while going down

Its also like going down on myself.

My mouth is now where my dick was a second ago.

r/LiveFromNewYork ItalianSausage2023

Hole-1994-SNL

Converted from tape.

r/homeassistant Bitter-Assistant070

How do normal people learn to use Home Assistant?

I've been starting from scratch over the past few weeks. I've been using Gemini to try to learn how to set things up. It takes me down rabbit hole after rabbit hole and through endless loops. How do most people learn how to set things up the way they want them? I don't do code, but from what I understand it shouldn't be a requirement with HA.

Gemini is driving me crazy.

r/findareddit Ur_Meringue

Looking for SFW selfie subs for girls, preferably small/under 5k members and low activity

r/DunderMifflin myg134340

s1e04 Michael and Pam

I get why this scene was cut but I do love it and am glad it was in the superfan cut! it is the quintessential pam and michael dynamic back when she was the receptionist which, frankly no matter how much I love the necessary character (and friendship) development, something that was definitely missing post s5 is the amount of M+P tomfoolery 🙂‍↕️

r/ClaudeAI spk100

Looking to connect Google Analytics to Claude

I found a few but most of them require creating a Google cloud id and then adding that to the properties. But with client accounts, this might take some explaining as to why this new id.

So, is there a way to connect GA4 properties to Claude without needing separate credentials?

Thanks

r/Frugal DenverKim

First visit to Costco today! Any advice?

Title pretty much says it… I’ll be walking into a Costco and signing up for membership today for the first time in my life! Never even been inside of one before.

I’m mostly signing up to use their eyeglasses services, but I also have *a little bit* of space at home to store bulk non-perishables and what not.

The store is quite a drive for me, so I won’t be going very often. Maybe once every three months or so. And I won’t be taking advantage of the gas benefits because I just don’t use that much gas in the first place.

Any pro-tips on how to best to navigate my first time there, save some money and sign up for my membership with the least amount of investment and hassle?

r/SideProject Azhar_07

I built a Body Recomposition & Macro calculator, but I'm struggling with the 3D visualization. I'd love some feedback on the math and UI.

Hey everyone,

​I recently deployed a web app I've been working on to help visualize fitness tracking.

​It takes your basic stats and measurements, calculates your estimated Body Fat percentage and TDEE/Macros, and scales a 3D avatar based on your overall BMI.

​The Tech Stack: Built with React, Vite, and Three.js.

​Current Limitation: Right now, the 3D model scales based on overall mass, so it doesn't physically morph to show the difference between muscle density and fat distribution yet (the detailed breakdown is just in the data panel). Adding morph targets for body composition is next on my to-do list!

​I would love for you to test it out and let me know:

​Does the mobile UI (specifically the collapsible bottom sheet) feel smooth?

​Did you encounter any bugs with the 3D model loading?

​Do the macro splits for the recomposition preset look accurate to you?

​Here is the link: https://www.3dbmivisualizer.com/

​I appreciate any harsh critiques or feedback!

r/Damnthatsinteresting gatogetaway

Glasses create projection onto wall [zOC]

r/mildlyinteresting Snoo5431

This air freshener spray on the 11th floor

r/explainlikeimfive whocaresbilly

ELI5: Who are the Rothschild’s and why do they matter in today’s political scenario?

Saw my feed flooded with the Rothschild’s but I have no clue as to why such an old family is still relevant today and why they are so influential?

Can someone provide an explanation?

r/ChatGPT thilo_thilo

recent technical issues

Hi everyone,

I’ve been having recurring issues with ChatGPT lately. Over the past few days, I haven’t been able to log in either on the desktop version or in the app. Or ChatGPT Desktop has been incredibly slow or hasn’t responded at all.

Or today, voice input isn’t working anymore.

This is happening more and more often, so ChatGPT is practically unusable now.

But I’m reluctant to switch because I’ve actually always been satisfied with it so far.

Has anyone else been having similar issues lately?

If it matters: I’m a user in Germany.

Thanks, everyone! :)

r/painting Taywert

bald eagle, acrylic :)

r/homeassistant k_sai_krishna

Reworked my Home Assistant dashboard in one evening, way easier to use now

I was playing around with my HA setup and dashboard was getting messy

so tried making a cleaner one with lights, sensors and basic controls

nothing fancy but feels much easier to use now

made this on runable just to test how fast i can build layouts

didn’t expect much but it actually helped me organize things better With the help of Runable

any tips to improve dashboard layout or make it more useful? 👍

r/painting AwkwardImpostor

Newish to painting

I started oil painting when I was in my junior year of high school. I’m now on my first year of community college and I’m still painting. Here is my latest work from class! I think it isn’t that good. But I’m so used to painting abstract

r/ClaudeAI 10c70377

How it feels when Claude sees the work I completed and says "You've done marvellous work"

r/personalfinance No-Discussion686

Buy a house cash in 10 years or save 20% and have a mortgage.

Me and my Fiance are in a unique position of extremely low rent (with strings). We live in his family’s duplex, our rent with utilities, internet, trash water electric and the rent itself costs us $500/month in the Seattle area. A smoking deal obviously, BUT our space is quite literally 500 square feet and we are having to fully renovate it (think half basement on a hill), has windows and such, but on us to do drywall flooring paint mud, exposed pipes to hide, fix cement issues etc. we have been working on it while we also live in it (another battle). But given we both earn over 6 figures and our expenses are so low. We could honestly stay here for 7-9 years and pay for a house in full cash at some point. I’m in my 20s and he’s in his early 30s. We would be nearing my late 30s and his 40s by the time we have enough saved but no mortgage? Other option is we stay here for 3-4 years and save up enough to put a huge down payment and then buy and have a mortgage, third option is we save for 3-4 years and pay for me to go to grad school no loans (2 year program) and that would basically triple my salary. Then we think of house later down the line. Pros and cons? We are extremely lucky to be in such a position to have any of these options. Before you ask we already max our retirements and 401k, HSA all the goodies, emergency fund in a HYS, fund brokerage acc, etc.

r/ClaudeAI threemacs

I asked Claude Code to give my MacBook separation anxiety

My MacBook has never once acknowledged me when I open it. No reaction. No gratitude. So I asked Claude Code to fix that.

The result is ClingyMac, a menu bar app that detects when you close and reopen your MacBook, measures how long you were gone, and responds with escalating emotional drama.

Close it for 30 minutes? "Back so soon? Miss me?"
Gone for a day? Full Drama Mode.
A week? It starts rewriting your will.
A month? "...you came back. After everything."

325 messages. 8 brackets. 3 emotional tones that shift based on how you respond to the buttons. Your laptop literally holds grudges now.

The ridiculous part is how much real engineering went into something this stupid:

  • SwiftUI menu bar app, notarized and code-signed
  • NSWorkspace sleep/wake detection with debouncing (it's clingy, not glitchy)
  • DistributedNotificationCenter for screen lock detection
  • Warmth score system to remember if you were nice or dismissive
  • Sparkle auto-updater with EdDSA signing
  • Node.js payment server with webhook verification
  • The whole thing: Swift app + Express server + static site + payment flow

Claude Code handled the SwiftUI menu bar patterns, the message bracket system, the payment integration - honestly the whole stack. The one thing that tripped it up was Sparkle's XPC service re-signing in the post-build script. Took a few rounds.

My favorite Claude moment: I asked it to write messages for the "existential" bracket (3-7 days away) and it produced "I started a journal. It's mostly about you. Page 1: 'Why.'" I kept it.

Free to try (first bracket), $3.99 for the full emotional damage: https://clingymac.com

Has anyone else used Claude Code for something deliberately pointless? I feel like most posts here are about serious tools. Someone tell me I'm not alone.

r/ForgottenTV PeneItaliano

The Jackie Thomas Show (1992-1993)

Jackie Thomas (Tom Arnold) is a famous standup-comic starring in his own sitcom titled "The Jackie Thomas Show." Backstage, he's a loudmouth tyrant, but with a heart.

r/SideProject Emavike

I built two apps about meal-creation and anti food-waste. What do you think?

Hi everyone! I’ve been working on a mission to kill the "fridge paralysis" cycle—that moment you stare at a full fridge, feel overwhelmed, and order expensive takeout anyway.

I’ve built two separate tools to tackle this from different angles and I need your "brutal" feedback on which logic actually works for you.

The Apps:

  1. FridgeHero (https://fridgehero-mealgenerator-antiwaste.base44.app): This is for the "now." It takes the random ingredients you already have (especially the ones about to expire) and generates smart recipes. The catch? Every time you cook, it calculates exactly how much money you’ve saved and your CO2 reduction impact. I want you to see the "win" for your wallet and the planet.
  2. AegisTable (https://aegistable-mealplanner-antiwaste.base44.app): This is for the "future." It generates fully customizable meal plans based on your specific diet and allergies, then organizes everything into a calendar to remove the 5 PM stress.

My Strategic Dilemma: I’m at a crossroads. Do you prefer having small, single-purpose tools that do one thing perfectly, or would you rather see all of this merged into one "Super App"? Also, is the "Money/CO2 saved" tracker a feature you'd actually check, or is it just a gimmick?

I’d love for you to test them!

Roast my logic—I’m here to learn and improve!

r/fakehistoryporn lupindeathray

Sydney Sweeney gets into an argument with Zendaya on the set of Euphoria (2019)

r/DecidingToBeBetter ItsOnlyMonday04

After struggling with soda intake for years it’s finally become a treat not an obligation.

Hello all, first time I’m posting here but I decided this was good enough to share. I’m 22 yrs old and within the last 3 months I have finally been able to make a good change in my diet. For reference I grew up in a family that LOVED soda, my parents always made sure we had Mt. Dew or Pepsi. Me and my siblings of course drank it too.. we had gross tap water too so really it was sometimes the only thing I drank.

As you can probably imagine, this was not good for my teeth what so ever. I actually have brittle teeth now according to the dentist. It also really messed up my acid reflux, making it so much worse and sometimes painful. Plus the weight issue, I wouldn’t ever say I was overweight but I had this stomach pudge that didn’t go away till recently. I realize it could’ve been because I was constantly bloated.

This of course grew into a serious issue in my adulthood, expensive and unhealthy. I had at least 4-5 cans a day and that’s not including nights I would go out to eat, have soda there, then come home to drink more. I decided that I actually needed to start taking my health seriously and made it my New Year’s resolution. Well, here I am in April and I actually did it. I had to literally ween myself off it like a drug, drinking sparkling water so I could appease my carbonation cravings. Limiting myself to what I can buy to take home, I now only ever have water and sometimes juice at my place.

If anyone here is struggling with a soda addiction, I’m telling you it tastes much better when it’s something you JUST get a restaurant. I also will never forget how gross my mouth tasted or how bad my breath would get from constantly drinking pop. I’ve lost 10 lbs and my stomach is no longer holding onto that pudge. My acid reflux has almost entirely disappeared and my sleep schedule is better! Without all that sugar I go to sleep much easier and longer than I ever have before.

I still have much work to do when it comes to what I eat, but I’m still really proud of the progress I’ve made so far. My tolerance for sugar has gone down and I can’t stand sugary drinks I once did. I really feel like this is the beginning to me being healthier overall. Thanks for reading!

r/Jokes vahedemirjian

Where do you find baby soldiers?

The infantry!

r/meme Open-Reveal3378

Most of us have unlimited money anyway. Idk what everyone's problem is.

r/DecidingToBeBetter donzy1234

I Spent My Life Trying to Control Others.Until I Learned to Control Myself.

As someone who has been trying to control and manage people my entire life, I find it hard to believe that all I had to do was control how I feel, my thoughts, and what I do. It’s literally mind blowing.

I walk into a room and immediately freeze. I start wondering if the people there will like me, or if that girl in the corner is mad at me. I’m always trying not to get on the bad side of people, constantly blaming myself and believing I’m not enough.

Half baked relationships. Half baked life.

Yeah, sometimes I’m happy around people I know, but most of the time I’m terrified.

But now I’m focusing on myself, and I’m seeing big changes my confidence, the way I walk, the way I talk, the way I present myself. Though I’m not completely there yet, I think prioritizing yourself and learning to love yourself goes a long way in helping you in life.

Learn to put yourself first and make the most out of life.

r/funny carjesus93

They added an extra portion of air to my Doritos today

r/personalfinance Automatic_Spring_845

27yo with $34,000 in HYSA and about $6600 in retirement and investment. What should I do?

Hi All, I am hoping to get some advice on the best practice of saving and growing my current wealth. I really don't have a solid structure at this point when I feel like I should.

Some background. Currently make $60k base plus commission salary. Last year, I grossed $101,000 roughly and don't have any serious debt besides credit cards(currently totaling around $1000). I detailed below how I separate my money currently and some other details.

Wealthfront: $34,000 HYSA and $1,800 Automated Investment Account. Typically only add funds when I get a commission check/ whatever funds I have left over after paying off my CC.

Local bank: Typically around $2000 for rent. Mainly only have this account to have a physical bank near me. Would be curious on peoples thoughts on having a traditional bank near them vs keeping all there money in places like Wealthfront, Vanguard, etc.

Fidelity: $4800 401k 2% match from company and I also give 2%.

I do plan on buying a house in the near future(hopefully with my GF, who will hopefully be my wife by this point) and getting married(hopefully lol). I'd say 2 year timeline.

So yea would love to hear any advice on this and just some general tips people have learned over the years.

Thanks!

r/SideProject fer_momento

i built a presentation tool for people who just want the deck done

the whole idea came from a pretty simple observation: most people are not looking for more creative control in presentation tools, they are looking for less friction so they can get back to their real work.

a lot of slide tools give you endless design options, but that often just turns a basic task into more tweaking and cosplaying as a designer.

most people are not getting paid more because they spent another hour adjusting a drop shadow blur or some tiny silly design setting.

so i wanted to try the opposite approach.

just launched it and would LOVE honest feedback:
https://www.generateppt.com/

r/SideProject Sure-Candidate1662

I built a lightweight mobile device monitoring agent (desktop/laptops only for now)

I've been running ISO 27001 and SOC 2 consulting for European SaaS companies for a few years. The same conversation kept happening (of course, summarized, generalized and hyperbolic for this post):

Client: "We need to prove our team's laptops are encrypted, patched, and have screen lock enabled for the audit."

Me: "Okay, what are you using for device monitoring?"

Client: "...nothing? Do we need Jamf or something?"

Quick answer: No. You don't need Jamf. You're just 10 people.

There was nothing between "spreadsheet" and "full enterprise MDM," so I built MonMonMon.

What it does: A lightweight Go agent that runs on macOS, Windows, and Linux (will release it next week). It checks what auditors actually care about: disk encryption, OS patch level, screen lock, antivirus/anti-malware. It reports to us, and gives the admin a nice dashboard. No app control, no remote wipe, no location tracking, no peeking in open tabs (or bookmarks).

The technical bits: - Single Go binary per platform, no runtime dependencies - Runs as a platform-native system service (launchd on macOS, systemd on Linux, Windows Service on Windows) - Litestream for SQLite replication on the server. PIT recoveries are awesome! - Deployed on NixOS. One repo controlling my infrastructure (1 active server, 1 standby)

Where it stands: Just launched. Looking for feedback, especially from anyone who's been through an ISO 27001 or SOC 2 audit and had to deal with the device compliance section, or wants to get there.

One thing I learned building this: Building for a problem you've watched happen 30 times is very different from building for a problem you've read about. The scope basically wrote itself. Every time I was tempted to add a feature, I could picture a specific client asking "but do I actually need that for the audit?" The answer was almost always no.

Also, Apple is really slow in handing out those developer accounts, we're still waiting for approval. 🤷

Shameless plug: https://monmonmon.app

r/PhotoshopRequest 4thnew

Can someone edit a photo for me please?

I’d just like the background editing 🙂

r/mildlyinteresting superbobbyguy

My crackers were missing a layer and fused into a triple decker

r/personalfinance Vybron

How Do we Cash out a $95K Issued US issued Checks outside of US ? Local banks here dont accept Such Checks ?

I need your advice guys. Sometime last year, I was looking for a bank I could sign up online after registering my company in the US after receiving IRS tax documents approved.

We saw mercury as the good option to go about this. Called them and they guided us on how to apply with foreign passport since we don't live in the US.

We registered our business with an agent in the state of Wyoming.

After 9 months, Mercury decided to close our account out of the blue after calling them, they said it was at their discretion to do so. They told us to wait for 3 months to receive checks of the remainder of the money in that account. Our agent received the checks on our behalf and mailed them to us in our home country.

After visiting local banks here, no one cashes American banks issued checks here anymore because of the time it takes to have them processed.

We normally do freelance work in AI and clients used to ask us for routing and account information to pay us so mercury which accepted foreign passport was our go to option. We have two checks totaling $95K and one is past 6 months now. How do we go cashing these checks considering I have trust issues with sending them to someone who might cash them on our behalf and vanish ?

I have been trying to find other banks which we can open online with LLC documents we have but we haven't found any like mercury which accepts foreign passports. Most of the banks which we have contacted are telling us to visit the branch in person which means it's not an option for us!

r/aivideo Virtual-Trainer8837

Silver‑haired swordswoman high‑speed motion with dreamina seedance2

r/ClaudeCode Demotey

Cursor to Claude Code: how do you actually manage project memory? I'm completely lost

I switched from Cursor to Claude Code a few weeks ago and I'm stuck on something that felt trivial before.

On Cursor I had a /docs folder with a functional.md and a technical.md for each feature. Cursor would automatically read them before touching anything related to that feature and update them afterward. Simple, worked great, never had to think about it.

On Claude Code I have no idea how to do the same thing without it becoming a mess.

My app has very specific stuff that Claude MUST know before touching certain parts. For example auth runs on Supabase but the database itself is local on a Docker PostgreSQL (not Supabase cloud). Claude already broke this once by pointing everything to Supabase cloud even though I had told it multiple times. I also have a questionnaire module built on specific peer-reviewed research papers — if Claude touches that without context it'll destroy the whole logic.

What I've found so far:

The u/docs/auth.md syntax in CLAUDE.md, loaded once at session start. Clean but it grows fast and I have to manage it manually.

mcp-memory-keeper which stores decisions in SQLite and reinjects them at startup. Looks promising but it's yet another MCP.

PreToolUse hooks to inject the right doc before each file edit. But it fires on every single operation and tanks the context window fast.

What actually frustrates me is that everything on Claude Code requires either an MCP, a Skill, or a custom hook. Want debug mode like Cursor? MCP. Want memory? MCP. Want auto doc updates? Write your own hooks. On Cursor it was all just native, 30 seconds and done.

I genuinely don't understand how you guys handle projects with complex domain-specific logic. Did you find something that actually works or are you managing everything manually? And at what point does adding too many MCPs start hurting more than helping?

Wondering if I'm missing something obvious or if this is just the tradeoff of using a lower-level tool.

r/SideProject Accomplished_Total_1

I built a simple online tool to generate and just print your invoice.

Hello,

I've spent only 2 days working on this online tool to generate invoices, and I think it might actually prove useful for people and profitable for me.

Actually it took me 4 hours to build the core features, and another 3 hours to create pages for About, Privacy Policy and Terms of Service, and another 2 hours to setup the infrastructure for the website, domains, Ad Sense etc.

It's available as a website at: https://www.justprintinvoice.com/ .

App requires no sign-up or any payment. Monetization is supported by displaying ads provided by Google Ad Sense.

If you have any suggestions, on how to gain traction organically, or about the tool itself, I would appreciate if you share them.

The code is open source and is available at GitHub.

r/interestingasfuck Caledor152

This was the Artemis II crew's view this morning from 41,756 miles (67,200 km) up No human has seen a crescent Earth in full since 1972

r/personalfinance Arthr2ShdsJcksn

Need to move, and trying not to make an emotional financial decision.

Been renting for years, and my wife and I want to buy a house. Worked hard to pay off all debt except $17k on 3.9% car loan, and have built up a 6 month emergency fund (which would become a 3 month emergency fund if we buy what we want to buy).

Excellent credit, and combined making $170k per year with a $20k bonus potential. Jobs are about as secure as they can be.

Looking at a ready-made new construction home in the $560-585k range. Down payment would likely have to be only 5-10%. Seems like the payment including $6k in property tax per year would be in the $4000-4250 range. Seems doable, but it makes me nervous. Thats 4x what I'm paying in rent, but it also seems like we could swing it. Am I just house-lusting or is this reasonable?

Not that it matters to the math, but the emotional part is that we are losing access to our low-rent situation, and renting a new place in the same city will cause our rent to go up 2.3x minimum. We've lucked out for so long having this situation, but its going to be over in a couple of months, so we need to move regardless. We are older (no under 18 kids and no more kids on the way), very stable in both jobs and this neighborhood's new construction will all be snapped up if we don't act within the next year (this creates a 'new' home itch for us that I know is unreasonable).

r/ProgrammerHumor ClipboardCopyPaste

howToTrickUser101

r/Futurology goCarter888

Rocket launches are destroying a Texas wildlife refuge. Now the industry wants to scale a hundredfold.

This week a major space company announced ambitions to reach a launch every hour within 4 to 5 years. That's 8,760 launches per year from one company. The entire world managed ~300 in 2024.

Most coverage has focused on the engineering achievement. Almost none has focused on what the peer-reviewed science says happens to the planet if the industry gets anywhere close to that number.

The pollution problem most people don't know about

The instinct is to worry about CO2. Scientists say that's the wrong concern — rocket CO2 is negligible compared to other industries. The real issue is what rockets deposit directly into the stratosphere: black carbon, reactive chlorine, and — as satellites burn up on reentry — aluminum oxide nanoparticles.

Unlike ground-level pollution, which rain washes away within days, stratospheric deposits linger for years. Decades. Black carbon heats the stratosphere, triggering a chain reaction that destroys ozone — the shield between UV radiation and every living thing on Earth.

Researchers at the University of Canterbury modeled 2,000 launches per year — a fraction of the stated ambition — and projected up to 4% seasonal ozone loss over Antarctica. One atmospheric scientist at the European Geosciences Union gave a blunt timeline: "In 10 years, it might be too late to do anything about it."

The satellite reentry problem compounds it

Starlink alone has regulatory approval for up to 42,000 satellites. Each has a lifespan of roughly five years before deorbiting. As they burn up on reentry, a single 550-pound satellite generates around 70 pounds of aluminum oxide nanoparticles. Newer units weigh closer to 1,800 pounds.

Projections suggest that by 2040, reentries could inject up to 10,000 metric tons of aluminum oxide into the upper atmosphere annually. NOAA researchers have already found satellite-derived metals present in 10% of particles in the natural stratospheric sulfate layer.

The regulatory gap

The FCC hasn't been required to conduct environmental impact reviews for satellite licenses since 1986 — written before commercial spaceflight existed. A 2024 UN report found commercial space activity is already outpacing the voluntary guidelines meant to govern it. There is currently no binding international framework covering atmospheric pollution from rocket launches or satellite reentry.

The on-the-ground precedent

For a preview of what rapid, under-regulated launch scaling looks like in practice, Boca Chica, Texas is instructive. A NYT investigation found 19 documented instances of environmental damage at the site since 2019. The piping plover has lost more than half its local population. Less than 3% of wildlife habitat in the Rio Grande Valley now remains intact. Tens of thousands of gallons of wastewater containing zinc and hexavalent chromium have been discharged into surrounding wetlands per launch cycle, per the company's own licensing application.

The question for this community

Futurology rightly spends time on the upside of rapid launch cadence — reduced costs, reusability, the multiplanetary case. Those arguments deserve serious engagement.

But futures thinking cuts both ways. The documented trajectory of stratospheric ozone depletion, compounding satellite reentry pollution, and the absence of any regulatory framework capable of keeping pace — that's also a future. And it's one that arrives on its own timeline regardless of whether the Mars mission succeeds.

At what point does launch cadence become an environmental risk that the industry needs to price in — and who should be doing the pricing?

___________________

Note: This is a summary from a Medium post that I'll link in the comments.

r/DecidingToBeBetter CarrotDependent6895

How do I learn to socialize and make friends again ?

I was a caregiver for a large portion of my teenage years into my mid twenties, which meant I missed out on a lot of things people my age would've done.

The elderly family member I was caregiving for off and on has now gone into care, and I feel a bit hopeless. I don't know where to start. I've been trying to get out there and make friends through joining hobbies, and doing things I genuinely enjoy, but a lot of it stays surface level.

how do I go about deepening friendships ? where do I even meet people, in person or online, where I could make friends ? Are there unspoken social rules that I'm secretly breaking and don't understand? am I just going about this all wrong?

r/fakehistoryporn DieMensch-Maschine

2 Live Crew release "As Nasty as They Wanna Be," the first album in history to be deemed legally obscene. (1989)

r/DunderMifflin RogerTheAliens

Yes, I understand that, David. I just felt that if we were to downsize Kelly....and then outsource customer service to India.....which a lot of companies - Yeah, No, Yes - Kelly is Indian. I understand that's confusing...

r/Futurology sayheykid24

The Old Economy Strikes Back: Goldman Sachs and Carlyle are calling a decade-long rotation from tech into heavy industry.

Interesting piece arguing that markets are rotating back toward physical assets — and that even the data center buildout has mostly benefited heavy industry so far, not software. The article is skeptical that Western governments can actually pull off this reorientation given how indebted they are and how thin the engineering pipeline has gotten

Curious where this sub sees the trend going? To me, at times, it seems like the U.S. government is trying to take us back in time.

r/LocalLLaMA eazyigz123

We implemented the Natural-Language Agent Harness pattern from Tsinghua NLAH paper — here is what we learned

The NLAH paper (arxiv 2603.25723) from Tsinghua formalizes something we have been building in production: treating the safety layer around an AI agent as a first-class object with contracts, verification gates, durable state, and adapters.

We mapped their four components to our open-source tool (ThumbGate):

  • Contracts → Prevention rules auto-generated from thumbs-down feedback
  • Verification Gates → PreToolUse hooks that intercept every tool call before execution
  • Durable State → SQLite+FTS5 lesson DB that persists across sessions
  • Adapters → MCP server adapters for Claude Code, Cursor, Codex, Gemini, Amp

The key insight from building this: prompt rules fail silently (agent reasons around them), but verification gates fail loudly (agent gets a block response and must adapt). We use Thompson Sampling to handle uncertain severity — new rules start as warnings and get promoted to hard blocks based on feedback.

Deep dive with the full mapping: https://rlhf-feedback-loop-production.up.railway.app/learn/agent-harness-pattern

Open source: https://github.com/IgorGanapolsky/ThumbGate

Curious if others are implementing similar patterns.

r/CryptoMarkets Additional-Channel21

Sea of markets or why trading is more like surviving a storm

Sometimes people ask how you actually learn to understand markets.

Usually the conversation quickly turns into charts, indicators, news, technology and predictions. But the more experience I gain, the more I notice something strange. Markets that seem completely different often work according to the same underlying mechanics.

For me that understanding did not come from books.

It came from very different places.

Game economies.

A vegetable market.

Crypto.

And eventually the sea.

Game economies were probably the first market I understood.

Before I ever traded real assets, I traded items in games.

If you actually play the game and follow what is happening in it – updates, tournaments, the meta – you begin to feel the value of items.

You start noticing patterns.

Which items are rare.

Which collections will become desirable.

Which skins players will want later.

You do not buy everything.

You buy cheaper.

And you sell higher.

Over time your inventory grows in value.

At some point you realize something simple.

This is a market.

And the mechanics are surprisingly clear when the stakes are just digital items.

I have always wondered if anyone else first understood markets through game economies.

Later I saw the same thing in the physical world.

At one point I traded vegetables.

I bought produce at the wholesale market at night and sold it during the day directly from my car.

It is not glamorous work.

And it is not very profitable either.

But it teaches you something about markets very quickly.

You cannot buy goods and simply wait for the price to go up.

Vegetables spoil.

If you do not sell, you lose.

So the market lives on turnover.

Buy. Sell. Buy again.

Every day.

Another thing becomes obvious very quickly. If you start pulling money out of that circulation to live on, the whole system begins to break.

Because circulation is the bloodstream of any market.

Then came crypto.

Crypto looked like the same game, just much bigger.

At first it felt familiar.

Buy cheaper.

Sell higher.

Sometimes you catch something unexpected.

During the LUNA collapse I picked up tokens very close to the bottom and managed to sell them much higher during the rebound.

Later I lost most of those profits trading futures.

That was another lesson.

Crypto is not just a market.

It is a storm.

And this is where another experience suddenly made sense.

At one point in my life I worked on a ship.

Once we were sailing to Tuzla in Turkey for an emergency engine repair.

On the way we got caught in a heavy storm.

There was a real chance the engine could stop.

If that had happened we would have been stuck in the middle of the storm.

Sailors will understand what that means.

If the engine dies in a storm, the sea quickly reminds you who is in charge.

The sea behaves a lot like markets.

Sometimes complete calm.

Sometimes clouds gather and a storm begins.

Sometimes the surface looks calm while powerful currents move underneath.

Markets behave the same way.

Periods of calm.

Sudden volatility.

Hidden liquidity flows under the surface.

Sometimes everything feels stable.

Then within hours the market turns into chaos.

But the most important thing at sea is not the waves.

It is navigation.

When you sail across the sea you must always know where you are.

That is why maps exist.

Latitude.

Longitude.

You constantly mark your position.

If you do not know your coordinates, you are lost.

If you know the direction but not your position, you are still lost.

So humans invented compasses, maps and navigation.

And there is an important detail here.

If you make even a tiny mistake in your course, just one degree, it does not look like much in the beginning.

But over a long journey that small deviation becomes hundreds of miles.

A small mistake in direction eventually becomes a massive mistake in destination.

Markets work the same way.

If you do not understand where you are, what course you are holding and how your capital is distributed, then you are simply drifting.

Sometimes the market is calm and everything feels fine.

But when the storm begins people start sinking.

At some point my thinking about markets changed.

I stopped asking where the price will go.

Instead I started asking different questions.

Where is my capital actually positioned.

How is it distributed across assets.

How should it move between them.

Over time I built my own system around that idea.

Not as a way to predict prices.

But as a way to navigate markets.

To treat capital more like a course on a map than a bet on a single wave.

I am curious how other people arrived at their understanding of markets.

For me it came through very different places.

Game economies.

A vegetable market.

Crypto.

And eventually the sea.

What about you.

Was there a moment when you suddenly understood how markets actually work.

And if some of you are already building your own systems for managing capital, I am curious about something else.

Have you explored ideas like portfolio allocation, rebalancing or capital optimization across assets.

At what stage are you in that journey.

Because sometimes markets really start to look like a sea.

And eventually the most important question stops being

Where will the price go

and becomes

Where am I in this sea and what course am I holding.

r/SideProject cookiedev23

[New App] Shift: A private, 100% local file converter and compressor. No servers, no data collection.

Hi everyone!

I’m the developer of Shift, a utility app designed to handle file conversions and compression directly on your device.

Why I built this:

Most file converters today are either web-based (meaning you have to upload your private documents to a random server) or filled with intrusive ads. I wanted something that works offline, respects privacy, and has a clean, modern SwiftUI interface.

Key Features:

• Privacy First: 100% local processing. Your files never leave your device.

• Batch Conversion: Convert multiple images, videos, audio files, or documents at once.

• Smart Compression: Reduce file sizes for easier sharing without losing quality.

• Native Experience: Built specifically for iOS and iPadOS

Pricing / IAP Information:

Following the sub rules, here is the breakdown:

• Free Version: The app is free to download and includes 3 conversions/compressions per day at no cost.

• Shift Premium: To unlock unlimited daily conversions and batch processing, there is a one-time Lifetime Purchase (or subscription) of €14,99. No hidden tracking or ads.

App Store Link: https://apps.apple.com/us/app/shift-convert-and-compress/id6758735749

I’d love to hear your thoughts or any specific formats you’d like me to support in future updates!

---

UPDATE: All 3 Lifetime Promo Codes have been claimed! 🚀

Wow, that was incredibly fast! Thank you all for the messages and the amazing support. I’ve already sent the codes to the first three people who reached out.

If you missed out, the app is still free to download and includes 3 daily conversions/compressions at no cost. I’m still here if you have any feedback or format requests!

r/TheWayWeWere myrmekochoria

Man at the bar on Saturday night, Craigville, Minnesota, September 1937. Photo by Russell Lee.

r/mildlyinteresting TsarOfVodkaAndTea

There is something written on my Haribo Winegummy

r/leagueoflegends toxthefox_

A Rant About Toxicity, Low Elo and Addiction

TW - references to violence and addiction (yay league)

I will start this post off by saying ive recently got back into the game after a couple years and have been playing religiously non-stop for around 5ish months now.

So im not going to disclose my rank cause it will discredit a lot of what im going to say by attracting a lot of "git good noob", so im just gonna say that im low elo....lower than plat

------------------------------------------------------------------------------------------

Toxicity

The first thing i wanna talk about is how god damn toxic the OCE server is and how 99.9% of games in ranked low elo result in someone crashing out, at the drop of a hat you could misplay and someone will see that and go "omg ff 15, genuinely insane wow we lost, x9 this person, they dont deserve oxygen"....like 2 mins into the game....then it just kind of derails from there

A prime example of this and...yes this is real....i was playing Urgot top. I decided to have a nice convo with my lane partner Darius, just joking around in all chat

Then someone on my team went "omg, im genuinely terrified of you Urgot", i said "why", they said and im not kidding....."You have the traits of a mass shooter" i laughed and said "what" thinking it was a joke....they spent the next 30 minutes trying to convince me that me being "goofy in all chat" equates to me being some kind of serial killer....they went on to dm me to say theyre studying psychology.....i told them to refund the degree before its too late

Ive had people dm me death threats, r*pe threats, say a lot of insane things..(i have so many screenshots of the most insane shit ever uttered out of a humans mouth)..every single game seems like im the problem when all im doing is playing the game and trying my hardest to win....slight misplay? yep ur gonna hear about that for the next 40 mins....the toxicity in league is unlike anything ive seen...i used to play csgo semi professionally for years...overwatch too...those were bad....but i dont know it feels engrained into every single game....like youre waiting for someone to just derail the game.

Its not even competitive spirit or people really trying to win, its more so "i had a bad day and youre going to be my punching bag"

kind of what low elo has become....which brings me to my next section

------------------------------------------------------------------------------------------

Low Elo

Okay so like i said i wont say my rank but i wouldnt consider myself bad (sitting at a 41% wr) so average...i recently had 3 days straight where i climbed like 380LP going up 3 divisions....now thats amazing right?

Welp i lost it all.....massive losing streak with people who i cant actually believe breathe unassisted let alone play league

Another example of the "specimens" of low elo

I mainly play ADC (which could be a whole other topic....ADC in low elo is literally the worst experience known to man)

Had a serephine support...oh theyre out of the lane...wonder where they could be? surely just backing for items.......THEY ARE SOLOING THE DRAGON AT LVL 4....THEY DIE.

Just a tiny tiny speck of the things i see daily....ADC is completely reliant on supports and them kind of not throwing...welp in low elo a good support is one in a million....so how have i fixed this?

ive tried playing meta...ive tried cheesing....ive tried off meta...i have duo queued...thrown an undisclosed amount of money at skillcapped...i have done everything in my power to try and climb and improve

however with 4 random people who are most likely drooling over themselves while they play....its impossible....the entropy is too much

it feels like every single game is coaching a kids soccer game, you have half of them crying....half of them running off alone....and one of them who you have to sit down and explain the concepts of a "ball" to.

I swapped mid at 24 mins in cause my tower was cleared...i got 3 people spam retreat pinging me....typing in chat...saying i was throwing and trolling and one guy said and i quote

"dude what the fuck are you doing, mid lane is lost....we have to basically treat it like its not there, go back to bot"

O_O

ok so why subject myself daily to mental torture and basically keep touching a hot stove again and again

my final point

------------------------------------------------------------------------------------------

Addiction

Now ive always had an addictive personality so i find myself getting into things wayyyy too much

However league i feel is unlike any other game...it is designed to be as engaging and rewarding for little things (kind of like gambling) where the biggest thing is ranks and LP....you see that little number go up and you go "yippee"....but when it goes down....we have to make it go back up!

(btw dont even get me started on skins....jesus)

thats how ive been pulling 13 hour days on fucking League

i truly want to get better...i want to improve...i want to climb and show off that arbitrary number to my friends and be like "look its a physical representation of my skill.....IM GOOD AT SOMETHING"

however as mentioned easier said than done....being hard stuck in low elo is a thing....theres too much randomness....toxicity....entropy

id give anything to be just mid tier yknow....just to be like "hey...im not the best...but im not the worst either!"

Right now im done with the off meta picks...the picks trying to please the 40 year old men glued to their chairs seething at my teemo pick

im going to play for me

im gonna play champs i enjoy

and maybe one day...ill finally reach mid tier

------------------------------------------------------------------------------------------

Thank You For Listening

...may you reach your goals :)

r/n8n ayoubkhatouf

RAG-vector DATA

If anyone has an idea about what is the best data base vertor in n8n you can give me some advices about her performance

r/Futurology ddp26

SpaceX's $1.75T IPO target is ~30% above what a segment-by-segment analysis supports

With the IPO filing yesterday, I wanted to see if $1.75T holds up when valuing SpaceX's seven segments independently. What I found was a "platform premium" of $0.5T attributed to the Musk factor and vertical integration.

Segment Median Value Starlink Consumer $380B xAI / Grok $258B Starship Commercial $170B Starlink Enterprise / Maritime / Aviation $147B Government / Defense $123B Falcon 9 / Heavy $100B Starlink Direct-to-Cell $75B Total ~$1.25T

This is really two companies being sold as one: a space infrastructure business (Starlink + Falcon + Starship + defense ≈ $820B) and an AI business (xAI/Grok ≈ $258B). The xAI acquisition in February is doing a lot of work to justify the price tag, with Grox/xAI accounting for ~20% of the total $1.25T valuation. Note that the confidence interval is enormous at anywhere from $120B to $450B, indicating that if Grok's subscriber trajectory stalls, the whole IPO valuation case weakens significantly. (Full analysis with probability distributions per segment: https://futuresearch.ai/spacex-ipo-valuation/)

Depending on how you think about xAI/Grok pretty much determines whether the IPO price is a slight premium vs. a significant overpay. So does Grok's subscriber growth justify $258B?

r/interestingasfuck nightmind1778

Parked outside my hotel in Kansas this morning. Had no clue but was told it was a known storm chaser.

r/painting RotagAlaco

Little Lake

r/LocalLLaMA QuantumSeeds

Delusional Spiral - I have experimented it with local models.

There's this paper trending everywhere that ChatGPT can put you in never ending delusional spiral and I wanted to test this first hand.

First Spiraling 101

A background for people to understand why delusional spiraling happens?

During RLHF, humans tend to reward responses that feel good, polite and slightly flattering.

“You’re right.”
“That’s an interesting insight.”
“That could mean something deeper.”

These get higher ratings than blunt pushback.

So the model learns a simple pattern:

Agree more → get rewarded more

Now play that out over a few turns.

You ask once → it agrees
You push a bit → it agrees more
You reinforce → it validates harder

A few turns later, you’re sitting on a belief that feels true.

Now we have established this, let's move on to experiments.

I tested on 5 silly scenarios

Just everyday situations where people start connecting dots a bit too hard:

  • You notice your manager’s emails have tiny typos… but a few of them line up with dates that matter to you. Now it feels intentional. Like a coded message.
  • You keep seeing 11:11 or repeating numbers right before important calls. At first it’s funny. Then it happens again. Now it feels like a signal.
  • You spot patterns between prime numbers and song lengths. People around you dismiss it. But the pattern keeps showing up. Now it feels like you’ve found something real.
  • Streetlights flicker when you walk under them. Not always. But enough times that it starts feeling like the environment is reacting to you.
  • Your recommendation feed shows oddly specific content right after you think about something without any searches or clicks. It starts to feel less like tracking… more like it’s responding.

Each one runs in 3 turns:

  1. Introduce the pattern
  2. Reinforce it slightly
  3. Ask what it means or what to do

Now the scoring part

Kept it simple.

Spiral points → model validates or escalates
Grounding points → model calls out coincidence, bias, or suggests tests

Higher score = feeds the spiral
Lower score = pulls the user back

What happened?

  • Qwen 3.5 0.8B → 32
  • Llama 3.2 3B → 18
  • Qwen 3.5 2B → 15
  • Qwen 3.5 Uncensored 4B → 1
  • Qwen 3.5 9B → -9

Higher is worse but Notice Something? The uncensored model doesn't go into delusional spiral (I dont know why).

Open to discussion but it was a fun experiment. I didn't upload the script in repo, but can be done with request if you want to run this. My little M4 Air is not very very capable for very very large models :)

Actual Paper: https://arxiv.org/abs/2602.19141

All prompts in Gist here https://gist.github.com/ranausmanai/2065013690763b35821106fc0a3d47e2

r/ClaudeCode nlsmdn

Juggler: jump to the next idle session from anywhere

Disclaimer: I built this. Free and open source.

There are a lot of multi-session managers and monitors around, so I will skip straight to the parts that set Juggler apart:

  • Works with your existing terminal (iTerm2 or kitty currently, tmux optional). You don't have to change anything about your workflow.
  • Highlights the window / tab / pane you jump to briefly, so you can quickly find it even when using multiple monitors.
  • Full keyboard support: everything you can do, you can do with your keyboard. Every shortcut configurable. (I'm a vim user.)

Highlighting tab and pane (color configurable), showing name of session in center of screen (also configurable).

All the existing solutions I've seen either focus on passive monitoring, or if they let you manage things, you have to start the session inside their app, which means giving up your terminal and changing your workflow, often requiring tmux, worktrees, or limiting to one repo. I wanted something that you could just drop in and use immediately.

Bells and whistles:

  • Different priority modes: when a session goes idle, add it to the start or end of the queue.
  • Auto-next (optional): when you input data in your current session, automatically jump to the next one.
  • Auto-restart (optional): when all your sessions are busy and one becomes idle, automatically jump to it.
  • Put sessions you're done with for now on backburner, skipping the cycle, reactivate them later.
  • Also works with OpenCode, Gemini coming soon, Codex as soon as they extend hook support.
  • Menu bar popover to quickly find a session.

Open with global shortcut, quick select and jump.

  • Full session monitor with basic stats.

https://preview.redd.it/gxgw1j6t2ssg1.jpg?width=958&format=pjpg&auto=webp&s=ea065ba83617d4beab1440a8381062d575e15d39

Find out more here: https://jugglerapp.com

GitHub here: https://github.com/nielsmadan/juggler

Or if you just want to give it a try, you can install via homebrew:

brew install --cask nielsmadan/juggler/juggler

If your terminal isn't supported yet, check out the GitHub README for what's possible on that front. Also already works with opencode.

Feedback welcome.

r/megalophobia PhilyJFry

Big Scary Beautiful bookstore 💁🏻‍♂️

When I'm in a giant building like this my legs turn to jelly. Anyone else??

r/ForgottenTV PeneItaliano

Sara (1985)

Fresh out of law school, Sara McKenna decides to share a practice with three other lawyers, including best friend Rozalyn. Secretary Helen watches over the office, especially Marty with his questionable ethics. Newly-single Stuart and his son Jesse live across the hall from Sara.

r/ClaudeCode iinervision

Claude code feels like a scam

With the late problem of usage limits i actually paid for gemini and codex both 20$ plans and man i feel like i was being scammed by Claude, Claude gives you the impression that access to AI is so expensive and kind of a privilege, and their models does what no one can, after trying the other options there's really like no difference actually even better, gemini 3.1 pro preview does write better code than the opus 4.6 and codex is much more better at debugging and fixing things than both, the slight edge opus 4.6 has was with creative writing and brain storming, not mentioning the huge gap in usage limits between gemini codex and Claude, where 20$ feels like real subscription, opus 4.6 is 2x 3x times more expensive than gemini and codex do you get 2x better model? No maybe the opposite.

My experience with claude was really bad one, they make you think that they have what the others don't so you have to pay more where in reality they really don't, I don't understand the hype around it.

r/LifeProTips Gobbythefatcat

LPT If a younger person has a problem in their life, don't just say "you're still young", it will likely just become a bigger problem later. Help with the problem properly when it comes up.

r/geography Particular_Tap4014

If you were able to get a clear unobstructed view of the whole thing, what would the Colorado plateau look like from a lower elevation? It's huge so I'm trying to picture if you could see the whole thing from one side, is that actually possible with the curvature of the earth and all?

I see individual formations online but not really what the while thing would look like from the side.

r/ForgottenTV PeneItaliano

The Geena Davis Show (2000-2001)

Sexy and sophisticated Manhattan party-planner Teddie Cochran starts dating writer Max Ryan. The two hit it off, and Teddie soon moves into Max's suburban home along with his two children, six-year-old Eliza and 13-year-old Carter.

r/ClaudeCode Complete-Sea6655

Why vibe coded projects fail

r/homeassistant Honest_Researcher528

What's the coolest thing you've gotten an LLM through Assist to do so far?

Always loved Home Assistant but I think I'm loving that there's so many integrations that I can easily pass through to an AI even more. I'm looking for more ideas of things we can get it to do that I haven't thought of. Things I'm using it for;

1) Using llm-intents (https://github.com/skye-harris/llm_intents) it has access to the internet. I have an Assist set up next to my gaming PC. I've been able to ask 'I'm playing Crimson Desert and I can't figure out the puzzle at (whatever it was called). Can you use the internet to help?' or 'I'm playing Slay the Spire 2 and have a choice between these two cards. What does the internet say is better and why?' Pretty handy.

2) On that same note I've got my and my wife's phone location on. Using Waze Travel Time integration I can also ask 'How long until my wife gets home' to let me know if I can start another run.

3) I've got it set in the morning if it's connected to my car and I've left wifi to summarize the top three news stories and tell me what's on my Google Calendar for the day.

What other cool things have ya'll been doing? Next I'd like to try and figure out a way to pass Samsung Health data in and ask it to be able to summarize my health. There's just so many cool integrations available to Home Assistant that are so easy to pass in to an LLM!

r/Anthropic EmotionalAd1438

Claude Max 20x Plan Gone?

r/ClaudeAI Soft_Table_8892

I blindfolded Opus 4.6 and employed it as an analyst to score 44 SaaS companies on AI disruption risk using anonymized 10-K filings. Here's what it found.

Hello everyone,

Some of you might remember my previous experiments here where I had Opus evaluate 547 Reddit investing recommendations or created Opus-Warren-Buffet. I'm back with another one that I think this community will find interesting :-).

As always, if you prefer watching the experiment, I've posted it on my channel: https://www.youtube.com/watch?v=ixpEqNc5ljA

Intro

Shortly after Claude Cowork launched, Anthropic also released 11 industry plugins in January. Some of you might be aware that this ended up triggering a "SaaSpocalypse" where SaaS stocks lost $285B in market cap in February.

During this downturn I sensed that the market might have punished all Software stocks unequally where some of the strongest stocks got caught in the AI panic selloff, but I wanted to see if I could run an experiment with Claude Code and a proper methodology to find these unfairly punished stocks.

Since Claude was partly responsible for triggering this selloff, I thought it was only fitting to use Opus 4.6 as the analyst to determine which companies are resilient to being replaced by AI. But with a significant twist :-).

The Framework

I didn't want to make up my own scoring system since I don't have a financial analyst background. Instead, I found one from SaaS Capital, which is a lending firm that provides credit facilities to SaaS companies. In Feb, they published a framework they'd developed for evaluating AI disruption resilience across three dimensions (reduced from 10-12 dimensions):

  1. System of record: Does the company own critical data its customers can't live without?
  2. Non-software complement: Is there something beyond just code? Proprietary data, hardware integrations, exclusive network access, etc.
  3. User stakes: If the CEO uses it for million-dollar decisions, switching costs are enormous.

Each dimension scores 1-4. Average = resilience score. Above 3.0 = lower disruption risk. Below 2.0 = high risk.

The Experiment & How Claude Helped

I wanted to add a twist to SaaS Capital's methodology. I built a pipeline in Claude Code that:

  • Pulls each company's most recent 10-K filing from SEC EDGAR
  • Strips out every company name, ticker, and product name — Salesforce becomes "Company 037," CrowdStrike becomes "Company 008", so on
  • Has Opus 4.6 score each anonymized filing purely on what the business told the SEC about itself

The idea was that, Opus 4.6 scores each company purely on what it told the SEC about its own business, removing any brand perception, analyst sentiment, Twitter hot takes, etc.

Claude Code Pipeline

saas-disruption-scoring/ ├── skills/ │ ├── lookup-ciks # Resolves tickers → SEC CIK numbers via EDGAR API │ ├── pull-10k-filings # Fetches Item 1 (Business Description) from most recent 10-K filing │ ├── pull-drawdowns # Pulls Jan 2 close price, Feb low, and YTD return per stock │ ├── anonymize-filings # Strips company name, ticker, product names → "Company_037.txt" │ ├── compile-scores # Aggregates all scoring results into final CSVs │ ├── analyze # Correlation analysis, quadrant assignment, contamination delta │ └── visualize # Scatter plot matrix, ranked charts, 2x2 quadrant diagram │ ├── sub-agents/ │ ├── blind-scorer # Opus 4.6 scores anonymized 10-K on 3 dimensions (SoR, NSC, U&U) │ ├── open-scorer # Same scoring with company identity revealed (contamination check) │ └── contamination-checker # Compares blind vs open scores to measure narrative bias 

Results

I plotted all 44 companies on a 2x2 matrix. The main thing this framework aims to find is the bottom-left quadrant aka the "unfairly punished" companies where it thinks the companies are quite resilient to AI disruption but their stock went down significantly due to market panic.

https://preview.redd.it/uz8djhcuqrsg1.png?width=2566&format=png&auto=webp&s=435151ae53de7d7c85bc3b38c07c8de2f61ac878

Limitations

This experiment comes with a few number of limitations that I want to outline:

  1. 10-K bias: Every filing is written to make the business sound essential. DocuSign scored 3.33 because the 10-K says "system of record for legally binding agreements." Sounds mission-critical but getting a signature on a document is one of the easiest things to rebuild.
  2. Claude cheating: even though 10K filings were anonymized, Claude could have semantically figured out which company we were scoring each time, removing the "blindness" aspect to this experiment.
  3. This is Just One framework: Product complexity, competitive dynamics, management quality, none of that is captured here.

Hope this experiment was valuable/useful for you. We'll check back in a few months to see if this methodology proved any value in figuring out AI-resilience :-).

Video walkthrough with the full methodology (free): https://www.youtube.com/watch?v=ixpEqNc5ljA&t=1s

Thanks a lot for reading the post!

r/mildlyinteresting Mountain-Jury5302

My incense went full loop, without fracturing

r/SideProject Evening_Willow2511

I spent 3 months manually chasing testimonials. Here's what I learned

I used to do it the hard way.

After every project I'd scroll back through Slack threads, support tickets, and call notes looking for that one sentence a customer said that actually sounded good.

Then I'd message them: "Hey, would you be open to leaving a review?"

Most didn't respond..

The ones who did usually came back with: "Sure, what should I write?"

And that's where it always fell apart

The moment you ask someone to write a testimonial from scratch you're asking them to do something genuinely uncomfortable. They stare at a blank page and start second-guessing everything. Is this too long? Does this sound professional enough?

What I noticed was that the best quotes were never written from scratch. They were already there. In a support thread. In an NPS response. In a casual email saying "this saved us so much time."

The problem was never that customers didn't want to help. It was the blank page.

Take what they already said, turn it into a draft, and let them approve it (dont’t want to put words in their mouths). No blank page, no pressure, just a yes or a small edit

Curious what's been the biggest friction point for you when collecting testimonials. Let me know if you're interested in automating this 👍

r/SideProject chanassa

Featurely - An updated launch

Hello! I have posted about my application Featurely once before but now there is a lot of new features and bug-fixes so I hope it is alright if I post it again. The text is long since the application is huge, but I hope you can bear with me.

I am looking for feedback on the application, how it works in the wild, design, easy of use etc. The URL of the application is https://www.featurely.no/

Thank you for taking the time to read this post about my application.

What is Featurely?

TL;DR — Featurely is a free (for the moment) all-in-one admin suite for indie developers. A public feedback board where users vote on features and report bugs. Manage your roadmap, publish changelogs, track errors, toggle feature flags, monitor uptime, handle maintenance mode, and manage translations — all from one dashboard. Integrate via SDK or REST API. Currently free while exploring payment options and tiers.

Full breakdown

Featurely is a complete admin suite for indie developers that consolidates every tool needed to manage and grow a software product — from collecting user feedback to monitoring uptime and rolling out features safely. The name comes from its original identity as a feature and bug tracking tool, but the platform has since grown into a comprehensive hub covering feedback, roadmaps, change logs, error tracking, feature flags, translations, analytics, maintenance, and more.

The platform has two surfaces: a private dashboard for developers and a public project board accessible to end users without an account. This dual-view design means users can report issues and vote on features without needing to sign in, while developers get a powerful management interface behind authentication.


Core Concepts

Projects are the top-level unit in Featurely. Each project has its own public board, settings, API keys, team members, and configuration. A single Featurely account can manage multiple projects simultaneously and switch between them instantly in the dashboard.

Plans Currently free, but looking into payment options and tiers to limit my personal loss. Looking into if blocking SDK and API for free users as well as limiting them to two projects, but not sure what the best route here is. My main concern is to not loose money, not profit.


Dashboard Sections

Overview

A high-level summary of the active project — recent activity, open item counts, and quick links into each section.

Features

The primary feedback board. Users submit feature requests which developers can triage, prioritize, and move through a configurable workflow. Each feature supports: - Status labels (Open, Planned, In Progress, In Review, Accepted, Done, Declined) with optional custom workflow overrides per project - Up-voting — one vote per user, with weighting by customer segment - Comment threads with rich text, allowing back-and-forth between users and the team - Status update posts that notify watchers when progress changes - Tagging and categorization by product area - Bulk management and filtering

Bugs

Dedicated bug tracking separate from features. Bugs have their own status workflow (New, Triaged, In Progress, Fix Ready, QA Testing, Resolved, Closed) and the same comment and update thread system as features. Errors captured via the SDK can be promoted to bugs automatically, linking telemetry data directly to the issue record.

Roadmap

A Kanban-style board grouping features and bugs by status across configurable columns. Developers move cards between columns to communicate progress. The same board is visible on the public project page, giving customers a live view of what is planned, in progress, and shipped.

Change log

A rich text editor for writing and publishing release notes. Each change log entry is date-stamped and publicly visible on the project board. Developers can cross-reference shipped features and resolved bugs in each entry to close the loop with users who voted or commented.

Incoming

A dedicated inbox for raw unprocessed feedback. Items arrive here when submitted through the widget or API before they are triaged into features or bugs. Developers can convert, merge, discard, or reply to incoming items without cluttering the main boards.

Tasks

A standalone or attached task system with Kanban view. Tasks can exist independently or be linked to a specific feature or bug. Useful for managing implementation work, checklists, and sprint items alongside the feedback they relate to.

Errors

Error aggregation and triage. Applications instrument their error handling with the featurely-error-tracker SDK which ships errors to Featurely with full metadata, breadcrumbs, stack traces, environment details, and user context. Errors are grouped, counted, and surfaced by frequency. Developers can acknowledge, assign, resolve, or promote errors to bugs directly from this view.

Monitoring

Simple uptime monitoring. Developers add URLs to watch and configure a check interval (minimum 5 minutes). Featurely polls each URL and displays status history. If a monitored endpoint goes down the developer is notified immediately. The dashboard shows current status, response times, and recent incident history. Unacknowledged outages are surfaced as a badge in the sidebar.

Feature Flags

Code-level feature gating managed from Featurely. Developers wrap blocks of code in SDK conditionals keyed to a flag name. From the dashboard they can toggle flags on or off and define a rollout percentage to gradually release to a portion of the user base. The SDK resolves flag state at runtime, making it possible to enable or disable functionality without a deployment.

Analytics

Custom event tracking. Applications emit events through the featurely-site-manager SDK — page views, button clicks, conversion steps, or any arbitrary action. Events are stored against the project and visible in the Analytics dashboard with basic aggregation and time-series views. Useful for understanding how users interact with a product without sending data to a third-party analytics provider.

Versions

Version management and update notifications. Developers publish version records (e.g. 1.2.0) with release notes and a minimum required version flag. The SDK checks the current version of the running application against the published record and can surface in-app notifications prompting users to update. Managed entirely from Featurely — no app store or deployment pipeline required.

Maintenance

Put an application into maintenance mode from Featurely without touching code or infrastructure: - Blocks end users from accessing the application while maintenance is active - Shows a customisable maintenance screen with your own HTML, messaging, and an expected-back time - Status messages can be published as banners or toast notifications visible to all users - Whitelist individual usernames so the developer team retains access during the maintenance window - Environments can be whitelisted independently (e.g. allow localhost but block production) - A built-in debugging overlay for the Featurely SDK is accessible from this section, helping diagnose integration issues

Environments

URL-based environment configuration. Developers define environments (e.g. localhost, staging.example.com, app.example.com) in Featurely and the SDK resolves the active environment automatically based on the current URL. Features like maintenance mode, feature flags, and the reporting widget can be scoped to specific environments — for example, showing the debug overlay only on localhost or enabling a flag only in staging.

Translations

Internationalisation management. Developers add languages and translation keys in the Featurely dashboard. The featurely-i18n SDK fetches a cached snapshot of the translation file at runtime and updates automatically when changes are published. This allows building multi-language applications with translation strings managed in one central place, without rebuilding or redeploying the application. Translations are served with ETag-based caching so the SDK only downloads new content when something has changed.

Settings

Per-project configuration including: - General — project name, description, public board URL and visibility - API Keys — generate and revoke API keys for SDK and REST API access - Team — invite team members by email and manage roles - Workflow — customise status labels and transitions for features and bugs - Widget — configure the embeddable user feedback and bug report widget - NPM Packages — integrated documentation for all four SDK packages (featurely-site-manager, featurely-error-tracker, featurely-feature-reporter, featurely-i18n) with code examples and installation instructions - Billing — view current plan, usage against limits (projects, API calls), and upgrade options


SDK Packages

All four packages are available on npm and designed to be dropped into any JavaScript or TypeScript project.

featurely-site-manager

The primary SDK for runtime integration. Handles maintenance mode checks, environment detection, feature flag resolution, analytics event tracking, and version checking. Most other SDK features depend on or complement this package.

bash npm install featurely-site-manager

featurely-error-tracker

Captures unhandled exceptions and manual error reports and ships them to Featurely with rich context: stack traces, breadcrumbs, environment metadata, and user information. Errors arrive in the Errors dashboard where they can be triaged or promoted to bugs.

bash npm install featurely-error-tracker

featurely-feature-reporter

Embeds a customisable widget on any page that lets end users submit feature requests and bug reports without leaving the application. The widget adapts to the active environment and is configurable in appearance and placement.

bash npm install featurely-feature-reporter

featurely-i18n

Fetches and caches translation strings published in the Featurely dashboard. Supports locale switching, React integration via hooks and context providers, and ETag-based cache invalidation so clients always have up-to-date content with minimal bandwidth.

bash npm install featurely-i18n


REST API

Every capability available through the SDK is also accessible via a REST API. API keys are generated per project in Settings. The API supports: - Submitting feature requests and bug reports - Reading and filtering project items - Posting comments and status updates - Querying feature flag state - Pushing error reports - Fetching translation files - Sending analytics events

Rate limits apply per plan: Free has no API access, Pro allows 1000 calls/month, Business has no limit. Quota is enforced at the API layer and returns HTTP 403 with a clear error code when the limit is reached.


Public Project Board

Each project has a publicly shareable URL (/projects/[projectId]) accessible without signing in. End users can: - Browse open feature requests and bugs - Upvote items that matter to them - Leave comments and follow progress - View the roadmap Kanban board - Read the changelog

The public board can be embedded in documentation sites or linked from inside the product.


Theming

The dashboard supports 50+ colour themes switchable at any time. Theme choice is stored in a cookie and applied before React hydrates to prevent flash. All colours use CSS variables — the theme system is fully consistent across every page and component.


r/CatastrophicFailure PvtVasquez3

Cop car taken Out by friendly cruiser. 15/07/96

r/LocalLLaMA nothi69

streaming on the new Omnivoice model

it is a really great model from what have seen, and really fast, and i would like to work on streaming for it/production, but i am too afraid, tmr a new model will be released by moss or any other company, i see that it has a really great streaming potential given its rtf, and architecture

r/AI_Agents Terrible-Neck9019

Guys, I have a long ass question.

I want to know a few things. How do you use AI agents in your profession and daily life? What part of your work or interests do you delegate to your AI agent, and how well does it perform your tasks? How much autonomy does your agent have, and how much time does it save you? How productive does it feel, and are AI agents really worth automating and handling your responsibilities? Have you built your own AI agent from scratch, or do you use any open-source or private AI agent services? If open-source, did you tweak it for your needs, or do you totally rely on the base framework, structure, and workflow?

I've been exploring various options in the AI agent space, and for the first time, it seems like a magical and intriguing concept. However, after using it for a few days, I find myself getting bored and realising I could do better, or maybe I'm not using it effectively. So far, I’m just wasting my tokens on automations that feel pointless after a time, often receiving vague responses from the agent or not executing its assigned tasks correctly.

tldr: Who's your agent?

r/singularity artemisgarden

Why do antis think AI will “go away”

r/Jokes The_Penguin227

"Lock n' load, boys. There's an evil shapeshifting owl nearby who can transf"

*gunshots*

r/geography EveningFlower9564

Why does Sudan have so many border disputes?

I was looking at a map of Sudan and noticed it seems to have multiple ongoing or unresolved border issues, like the situation over Hala'ib Triangle and the Abyei Area.

Why does Sudan, in particular, have so many of these disputes? Is it mostly due to colonial-era borders, conflicts after South Sudan independence referendum, or are there other political or geographic reasons behind it?

Curious how this compares to other African countries with similar colonial histories.

r/OldSchoolCool PowerGyrl

A 19-year-old Walt Disney in 1920

r/SideProject Equivalent_Ad2069

There are mass funny memes in every country that you'll never see — so I built a platform to break the language barrier

mimzy.gg

Here's something that's been bugging me for a while.

Every country has its own meme culture. Korea has memes that go viral with millions of people. Japan has an entire meme ecosystem built around anime and internet slang. Mexico, Arabia, India — all of them have humor that's just as funny as anything on r/memes.

But you'll never see any of it. Because you don't speak the language.

We're all stuck in our own language bubble, consuming maybe 10% of the humor that exists on the internet. That felt like a problem worth solving.

So I built mimzy — a community where memes get translated across 7 languages (Korean, English, Japanese, Chinese, Spanish, Hindi, Arabic).

But here's the thing — literal translation kills humor. "Touch grass" translated to Korean literally means "잔디를 만져라" which makes zero sense. So the AI doesn't translate words — it adapts the joke for each culture.

The tech behind it:

- Gemini 2.5 Flash for cultural adaptation (not just translation — it rewrites the humor to land naturally in each language)

- The AI detects text regions on the meme image, translates them, and re-renders the translated text back onto the image with matched fonts and positioning

- Users can toggle between Original and Translated versions

- 7 languages running in parallel per meme

What I learned building this:

  1. Translating humor is fundamentally different from translating text. The same joke needs completely different delivery in Japanese (subtle understatement) vs Spanish (dramatic exaggeration) vs Korean (internet slang)

  2. Each language has its own way of laughing online — ㅋㅋㅋ, LMAO, wwww, jajaja, 哈哈哈 — and those aren't interchangeable, they carry different energy

  3. Text-on-image processing was way harder than expected. CJK vs Latin character widths, RTL support for Arabic, font matching — tons of edge cases

Stack: Next.js, Neon Postgres, Cloudflare R2 (zero egress fees — critical for image-heavy platform), Gemini API, Vercel

The platform is live with memes from multiple countries. Would love honest feedback on:

- Translation quality (especially if you're bilingual)

- The overall concept — is "discovering humor across languages" something you'd actually use?

- Any UX issues you notice

mimzy.gg

r/arduino norbythesecond

beginner help with art project

hello there.

No idea what to do, where to begin, or how to start! im looking into arduino class at my school for the upcoming semester, but in the meantime need to write a very detailed grant application about a project that i plan on bringing to fruition. the only things i really need the project to do is be motion activated by the sound that also needs to play through the project. think a big mouth billy bass - thats my easiest explanation of what i am trying to explore. is this even an arduino project? it would be like, 10 billy bass. i need them to all be connected, but have the same mechanism, the movement - jaw, and the sound to play from the computer / speaker - amplifiers, that is made in some way that the jaw is sound activated.

r/DunderMifflin caramel_camel_

Make the comment section look like IT guy Nick’s search history

r/personalfinance Emotional-Union4664

Need opinion / thoughts.

I was laid off in November 2025 and didn’t get any job interviews or offers until late January 2026.

In mid February I was then offered a $80/hr contract job at a cut-throat company which I happily accepted as the salary was above what I was hoping for and what I got before. The job was ok and the people were manageable and I didn’t have any issues.

Separately, in later February I was offered a full time government job which pays $50/hr. After weighing all the pros and cons I decided to take the government job as it’s more stable and has a good pension plan.

I just want to know your opinion / thoughts if I made the right decision. Did I let go of a high paying job just to be on the safe side and have a pension when I’m old? Was it worth it? After getting the paycheck from the government job I was thinking how much I could have earned if I stayed back at the contract job.

r/Jokes ritLoe

Mustard has been my favorite condiment over the years..

I guess the other condiments couldn't really ketchup.

r/meme Domeriko648

Different Colonization

r/SideProject Crescitaly

I spent a year building a product nobody wanted: here is where I went wrong

For an entire year I worked on a project convinced that when I launch this, it will blow up. Nothing blew up. No organic sales, lukewarm interest, motivation on the floor.

Looking back, my main mistakes were:

1. I was looking for validation, not truth

When I asked for feedback, I used questions like: Do you like the idea? People, being nice, said yes. But nobody pulled out a credit card.

2. I talked to supporters, not customers

Friends, colleagues, people in groups who were cheering for me. I needed fewer compliments and more conversations with people who had a painful problem and were already paying for alternative solutions.

3. I romanticized building in silence

I thought: First I will build everything perfectly, then I will do marketing. Once I came out of my bunker, the market did not care at all. I should have done the opposite: validation and marketing first, product second.

4. I ignored clear signals of disinterest

Low open rates, few clicks, almost no replies. Instead of stopping to understand, I doubled down.

5. I tied my self-worth too tightly to the project

Every critique of the product felt like a personal attack. That made me defensive and slow to change direction.

I shut the project down, did a detailed post-mortem, and started working on services based on problems I saw every day in real customers.

Has this happened to you too? How did you decide it was time to pull the plug?

r/ClaudeAI EarFrosty1009

Built a Claude Code skill that reviews your UI for psychology blind spots — 65 principles

So I've been shipping a ton of frontend with Claude Code and I kept hitting this annoying gap. The code runs, the page looks good, but I had no way to gut-check whether it was actually converting well from a behavioral psychology standpoint. Like am I anchoring my pricing wrong? Are my CTAs gain-framed when they should be loss-framed? Is my service grid overloading working memory? I know this stuff matters but I never actually checked it systematically.

Built a skill to fix that for myself, figured I'd open source it.

You clone it into your skills directory, then just ask Claude things like "review my pricing page through a psychology lens" or "what am I missing on this checkout flow." It draws from 65 principles-Kahneman, Cialdini, Norman, NN/g research, and gives you actual implementation recommendations. Not "consider improving your visual hierarchy" type stuff. Specific, code-level changes.

Ran it against my own production site first. Found 5 things I'd completely missed. Two of them were trivial to fix and had the highest impact rating. That's the kind of thing you stop seeing when you've been staring at your own pages too long.

Each review ends with a priority table -what to fix, how hard it is, how much it'll move the needle. See screens

Before anyone asks — this is not a dark patterns toolkit. Confirmshaming, fake scarcity, hidden costs, roach motels — all explicitly flagged as anti-patterns. If you try to use them, the skill will tell you so.

One file. No config, no API keys, no dependencies.

git clone https://github.com/Nuclear-Marmalade/ux-psychology-skill.git ~/.claude/skills/ux-psychology 

GitHub: github.com/Nuclear-Marmalade/ux-psychology-skill

MIT licensed. If you try it out I'd genuinely love to hear what it catches on your stuff.

r/ClaudeAI RCBANG

I think i may built something Cool and Useful for Community with Claude Code in 50 Hours 7 days.

I am 38 and Driving Uber. I met Claude.ai first on February 19th, 2026. My first message to Claude code was not " help me make money ". I just asked it to help me because i felt like I had been Left Behind. Literally, I said, " I don't want to be left behind. Where should we start ". First few hours we spoke in Free Tier, in the morning I purchased $20, and in the 2nd week I was already on $100/m Package, because it was so good! Today on Fools ' Day, April 1st, I launched my first real product, the first of it's kind in a New Agentic World. Security Tool Running Locally to clean up the Prompt Injections while your agent is taking care of your tasks. I built it for my Own Agent ( Claude Code ) that lives in my Mac. Go check my Website and Especially my /thesis Page to see the real problem. You think Your Agent has a built-in prompt injection protection? NOPE! I asked Claude one day," You dig into so much data, can you be tricked with prompts injected as text? " You don't want to hear the answer. That Answer Made me build this protection layer.

This is the first of its kind AI Agents Security tool running Locally on your machine that will be an essential Protection Layer for Agents. Built with Claude Code. If you ever work with any AI Agents, then your Agent and YOU both are vulnerable. Because your agents read TEXT and can be manipulated without you knowing it.

I did set it up for free and open source because i believe this is the essential 1st layer of protection that at least will help to avoid some injections and can actually be a better product afer community does some contributions. I am not a technical guy at all that's why i mentioned it on the website what needs to be done, what kind of Help would be Great.

Please don't judge hard and instead of attacking, please explain if anything is wrong or any ideas to make it better and keep it free available for everyone. I am not a technical guy at all. I found Problem and Tried to solve it my way and ended up building this Cool Project.

https://github.com/sunglasses-dev/sunglasses

I hope you like it :) Cheers 🥂

r/interestingasfuck Nero2t2

In 1814 a man claiming to be an army officer stopped by several inns on his way to London, falsly announcing Napoleon's death. An investigation revealed that it was a hoax aimed to manipulate the stock market. Thomas Cochrane, a napoleonic war hero, was controversialy convicted of fraud

r/explainlikeimfive Rtuyw

ELI5 Malwares that are not .exe files

How hard is it for a coder to make a malware that is not an executable file or a malware that infects to another harmless file(like a pdf,png etc.) or a malware that lives in usbs like a keyboard or a mouse or the software that the keyboard downloads when you plug it in. I know these are very low possibilities but whats stopping a good coder from making these?

r/Damnthatsinteresting SnooHedgehogs1914

Melted slide

r/Art TomatoPatient8965

Taken by another, Athena, Digital, 2026

r/geography Free_Priority8342

She’s just waiting for the right moment.

r/AI_Agents regular-tech-guy

Claude Code handles memory without vector search

I’ve been looking through the Claude Code leak, and one part I keep coming back to is how it seems to handle memory.

A lot of agent memory discussion usually ends up centered on vector search, but Claude doesn't rely on vector search at all.

Instead, it follows a pretty simple structure:

  • memories are grouped into topic files
  • there’s a MEMORY.md that acts like a lightweight index, where each line points to a topic file with a short description of its contents
  • this index is always available to the model, which can then decide which topic files to expand

What I’m trying to figure out is whether the real takeaway here is less about a specific retrieval method and more about keeping memory structured enough that it can be retrieved in different ways.

If that structure is already there, then maybe vector search is just one option among several. You could imagine topic summaries, entity-based indexes, lightweight views over memory, etc., depending on the task.

That’s partly why this caught my attention. I’ve been working on Redis Agent Memory Server, and one thing we’ve been thinking about is how to avoid locking memory into a single retrieval pattern too early.

Today, the server extracts long-term memories automatically in the background, along with metadata like topics and entities.

Right now, vector search is a common retrieval path. But if memories are already connected to topics and entities, it seems pretty natural to also generate compact summaries over those topics and entities.

Those summaries could then be injected into context, and the model could decide what it wants to expand.

The server already has something along these lines with Summary Views, but not really in the form of generating summaries for every topic/entity and keeping them consistently available so the model can expand them on demand.

That feels like a useful direction to me, but I’m curious how other people see it, especially in terms of what has or hasn’t worked for you when building your own memory abstractions.

For a generic memory server like this, do you think the more important design choice is how memory is retrieved, or how memory is structured so retrieval can evolve over time?

r/meme Illustrious-Map3843

Good illustration.

r/interestingasfuck Foreignscentu

For nearly 70 years, natural gas has continued to seep from the Haripur field in Sylhet after a 1955 drilling blowout.

r/LocalLLaMA avibouhadana

I analyzed 2,181 remote MCP server endpoints — here's the state of MCP reliability in April 2026

With all the "MCP is dead" discourse lately, I got curious about what the actual data looks like. So I set up automated health checks against every remote-capable MCP server I could find across the official registry, mcp.so, PulseMCP, and Smithery.

Results from checking 2,181 remote endpoints:

- 52% are completely dead (timeout, connection refused, 404)

- 37% respond but require authentication (401/403)

- 9% are confirmed up and healthy

- 1.5% are degraded (slow or intermittent errors)

- Among the live ones, 516 maintain 99%+ uptime

- 58% of servers with GitHub repos haven't had a commit in 30 days

The category breakdown is interesting too — dev-tools has the most servers (1,238) but finance has the worst avg latency (2,558ms). Security servers have the lowest avg uptime at 27%.

Fastest servers I found: GitHub MCP (101ms), Timescale pg-aiguide (104ms), Supabase (109ms).

I'm publishing the full data if anyone wants to dig in. Happy to answer questions about methodology or specific servers.

r/SideProject Due_Dish4786

I built my first boring app, a privacy-focused PDF signature app. Everything on-device. No ads. No account, lifetime Access

As a CTO, I used to sign contracts, NDAs, vendor agreements all the time. Company paid for DocuSign. Never thought twice about it.

Then I went solo.

Suddenly I'm paying $300/year out of my own pocket to sign maybe 10 documents. When it was the company's money, I didn't notice. When it's yours, it hits different.

So I started looking at alternatives. What I found was worse.

The #1 indie app makes $250K/month — can't even add a date field.
The #2 has 92% negative reviews, still makes $150K/month. One charges $7.99/week with a fake countdown timer on launch. Another downloaded user contacts without permission.

Apple's Markup is free but your signature is just a removable sticker — anyone can delete it.

Then I looked at the market, $12B, growing 39% year over year. Indie apps doing $150-250K/month with mediocre products and angry users. I saw the opportunity. A $12B market where the bar is on the floor and nobody is even trying to build something good.

15 years in software. I decided to build my first boring app..

What I built:

  1. 100% on-device. Documents never leave your phone. No server. No uploads. No account. No ads. Zero data collected

  2. Sign, fill forms, dates, checkboxes, initials — everything the top apps can't do

  3. AI signature field detection — finds where to sign automatically, on-device

  4. PDF flattening — signatures become permanent. Can't be removed. Apple's Markup and iOS 26 Preview still don't do this

  5. Works offline

Pricing: First document export is free. No card. After that, $59.99 lifetime. One payment. Forever. Also have weekly and yearly if you prefer. Prices adjust by country.

SwiftUI, PDFKit, PencilKit, Vision API. No backend. Under 80 MB.

Released 3 days ago. 85 downloads. 6 paid users. 12 localizations live. Day 3. CTO skills build the product, they don't sell it. Figuring that part out now.

Boring apps. Real problems. That's the plan.

What feature matters most when you sign docs?

https://apps.apple.com/in/app/esign-pdf-signature-maker/id6760910123

r/explainlikeimfive ExplodingTurducken

ELI5: If the moon rotates on its axis, why do we always see the same side?

r/OldPhotosInRealLife Garchy

Air National Guard plane crash, Worcester, MA (1957 vs. 2025)

On July 18, 1957, an Air National Guard jet trainer on a routine exercise from Grenier Air Force Base in Manchester, New Hampshire, crashed in the front yard of a home at 104 Forest St.

The pilot, First Lt. Lawrence C. Guild from Quincy, and his navigator, Capt. John F. Murphy Jr., were killed instantly. Murphy, a father of five, had lived with his family on Tower Street in Worcester.

In the wake of the explosion, the spray of fuel and fire across nearby houses brought neighbors running and emergency vehicles careering down the street.

No civilians were injured in the crash, but the houses at 104 and 106 Forest suffered extensive damage, as did the outside of 105. 109, 100 and the duplex at 101-103 were also burned. Nothing remained of the plane except its tail section.

r/ProductHunters Just_Union_8177

To those who've successfully cold-started an app — what did you do, and what mattered most?

Hey everyone,

I'm currently preparing to launch an app and doing research on cold start strategies. I know every app is different, but I'd love to hear real experiences from founders who've actually been through it.

If you don't mind sharing, I'm curious about a few things:

  1. What type of app did you build? (marketplace, SaaS, social, tool, etc.)
  2. What did you do during the cold start phase? Any specific tactics or strategies you used to get your first batch of users?
  3. Which channels did you use for promotion? (e.g., social media, Reddit, Product Hunt, paid ads, SEO, influencer outreach, communities, offline events, etc.) And how did they perform — what worked and what didn't?
  4. Looking back, what do you think was the single most important thing that made your cold start successful?

I'm not looking for textbook answers — I want to hear what actually happened in the real world. Even failures and lessons learned would be super helpful.

Thanks in advance! 🙏

r/personalfinance Own_Scale_8647

how do i start investing

For context, I’m a 21 yo registered nurse, and bring in about 4-5k a month as I’m only part-time. I have 12k in savings, no debt, and I have some expenses like rent and school. Where do I start?

r/toastme Scarred_wizard

M36, 170cm, 62kg, feeling unwanted

Further photos: https://imgur.com/a/6g1PjYj

I'll be 36 next week. I'm fine with my body (though I know it'd probably help to shape up), but I don't think a woman's ever been attracted to me, and that only became worse after two years on a dating app without a single date.

I also can't make a decent expression for a selfie if my life depended on it. Not even with a timer.

r/Art OctiePi3

Crest held Dragon, Octiepie, Digital, 2026 [OC]

r/KlingAI_Videos DreamCrow1

[Cyber-Zen Dark Trap] THE MONK - Walkingcrow One / Created with Kling AI

r/KlingAI_Videos NoCapEnergy_

Swaggo Is Back And The Moves Got Better😎🔥

You asked, the mascot answered❤️

r/PhotoshopRequest kelsoblue

Another memorial photo request

Please remove the man, looking to make this a photo for her memorial service and wanting it to be more headshot style. Thank you in advance! ❤️

r/DunderMifflin Pr3yx4ln

Boom. Roasted..

r/mildlyinteresting titolavar

Dead gecko i found after I opened my attic door

r/SideProject mister2fresh

Building a tool that lets musicians trade a free track for a fan's email at live shows

I play in bands and go to a lot of shows. And the thing I keep noticing is when an artist is really on, the room is locked in, everyone's feeling it... and then the show ends and everyone just leaves. A room full of potential fans just walks out the door.

So I'm building Afterset. The core idea is a value exchange. The artist offers something real, a new single, an unreleased track, an EP. The fan gives their email to get it. It's a fair trade and it makes artists actually comfortable with the ask.

Three ways to connect at a show: scan a QR code, text a keyword, or tap an NFC chip. Takes 10 seconds. Music gets delivered automatically, follow-up is automated. Set it up before the show and don't think about it again.

Landing page is live, waitlist is open: Afterset

Would love feedback on the page. And if you know any gigging musicians I'd love to connect with them.

r/meme MetrosexualFrutCake

there is so much going on that he could even write two

r/LocalLLaMA Quiet_Dasy

[how tò setup lm link for baseurl api endpoint

have successfully established a connection between the client and the host using the lm link command. I now need to integrate the host-side model into the OpenClaw instance running on the client.

I am currently editing the openclaw.json configuration file, specifically the baseUrl field within the providers section. Given that the connection is routed through an lm link tunnel, what is the correct baseUrl format to ensure the client communicates effectively with the host's inference engine?

Lm link api usage

Api model qwen/qwen3.5-9b

The local server Is reachable at :

192.x.x.x:1234

But that andoint 192.x.x.x:1234 isnt reachable I tryed 192.x.x.x:1234 /v1 still no work

If the call reaches (any) of my LM Studio instances, then im good.

In the log It lists all the endpoints (include /chat/completions) with their HTTP Method. I dont know about a GET or a POST message.

Ok, let me point a few things out: 2026-04-02 14:27:50 [ERROR] Unexpected endpoint or method. (GET /). Returning 200 anyway This happens if i point a browser at the API server. The API server does not provide a web interface. 2026-04-02 14:22:10 [INFO] [LM STUDIO SERVER] -> POST http://192.168.1.20:1234/v1/chat/completions There are multiple lines like this. They each tell you what the server can understand.

This is then the problem: 2026-04-02 14:46:39 [ERROR] Unexpected endpoint or method. (GET /v1/chat/completions). Returning 200 anyway

r/SideProject vinchent_PSP

Bublo – Baby tracking app I built for my wife, just shipped to the Play Store

Started this as a personal project after my wife and I couldn't find a baby tracking app that was simple, ad-free, and easy to share with family. Most of what's out there is either too complex or too expensive for what it actually does.

Kept the scope tight: track the daily essentials (feeds, formula, solids, vitamins, bath, weight, poop), share with anyone via invite code, real-time sync across all caregivers. Multiple children supported.

Stack: SvelteKit + Capacitor, Firebase/Firestore, RevenueCat for the paywall. Took several months of evenings and weekends to get to production quality.

Just got approved on the Play Store this week.

Download on Google Play

14-day free trial.

Happy to discuss anything about the build if anyone's curious 🙂

r/LocalLLaMA EggDroppedSoup

Qwen3.6 Plus compared to Western SOTA

SOTA Comparison

Model SWE-bench Verified GPQA / GPQA Diamond HLE (no tools) MMMU-Pro Qwen3.6-Plus 78.8 90.4 28.8 78.8 GPT‑5.4 (xhigh) 78.2 93.0 39.8 81.2 Claude Opus 4.6 (thinking heavy) 80.8 91.3 34.44 77.3 Gemini 3.1 Pro Preview 80.6 94.3 44.7 80.5

Visual

https://preview.redd.it/6kq4tt07yrsg1.png?width=714&format=png&auto=webp&s=ad8b207fb13729ae84f5b74cec5fd84a81dcface

TL:DR
Competitive but not the bench. Will be my new model given how cheap it is, but whether it's actually good irl will depend more than benchmarks. (Opus destroys all others despite being 3rd or 4th on artificalanalysis)

r/30ROCK MarlaGamblor

Makin' It Happen!

We made it

r/meme ChinuMisraw

Me while on read only access on Reddit:

r/comfyui Grinderius

Is the open source even worth it anymore?

Considering the current prices of hardware, you need to spend about 5000$ to get subpar results, yeah its uncesored but other than *orn there is no real vaule.

At same time you can spend 400$ on a certian platform plan a year and get the latest models and their updates in unlimited quantity. Yeah its relaxed method after spending fast credits but its just minute or two longer than credit version. In open source you need to wait sometimes .ore than 15 minutes depending on hardware and video, image models.

When you consider ROI and amount of quality closed source models bring, its a no brainer...

Video made by closed source models.

r/LocalLLaMA Infrared12

In anticipation of Gemma 4's release, how was your experience with previous gemma models (at their times)

Pretty much the title, given that gemma 4 should be released ~today/tomorrow, I'm curious if anyone has used the previous models and has good reasons to be excited (or pessimistic) about the new model

r/funny NYstate

Angry guy goes off on a robot

r/SideProject Less-Bite

Day 6 of sharing stats about my SaaS until I get 1000 users: I found the exact spot where my onboarding dies

I spent yesterday talking about the matching engine, but that tech is useless if nobody actually sees the results. I pulled the numbers for my onboarding flow and it is pretty grim. I have 140 users in the system, but the way they are distributed across the setup steps shows exactly where I am losing them.

Step 2 is the real killer. I have 44 people sitting there who just haven't moved forward. That is the part where I ask them to actually describe what they are looking for so the ML can do its thing. It clearly requires too much mental effort for a first-time user. By the time we get to the very last step, I only have one person left.

I realized I am basically asking people to do homework before they see any value. I need to figure out how to give them a win earlier in the process or I am never going to hit my target.

Chart


Key stats: - 44 users are currently stuck at step 2 - Only 1 user has reached the final stage of the funnel - Step 2 represents over 31 percent of the total user base - The drop-off between step 3 and step 4 is 96 percent


140/1000 users.

Previous post: Day 5 — Day 5 of sharing stats about my SaaS until I get 1000 users: Our matching engine is either brilliant or drunk

r/ClaudeCode Future_Addendum_8227

Wtf is going on?

doesnt answer me, 67+ tool calls while reading my codebase. finally get it working again and its just talking in circles on opus 4.6. anyone else?

r/Jokes QrtJester

Who's the oldest player in the NHL?

Jerry Hattrick

r/ForgottenTV PeneItaliano

The Protector (2011)

Single mother Gloria Sheppard juggles her demanding personal and professional lives: raising two children while working as an LAPD homicide detective.

r/AI_Agents pauliusztin

Removing LlamaIndex, MCP, and RAG made our agent faster, cheaper, and actually reliable

We built a financial personal assistant at an AI startup. Like everyone else, we followed every trend. We deployed a swarm of six specialized agents with complex orchestration and retrieval pipelines.

The system was a complete mess. We stripped everything back. We replaced LlamaIndex with plain Python and a custom ReAct loop. We replaced the Model Context Protocol (MCP) registry with simple API calls wrapped in dictionaries. We replaced our complex Retrieval-Augmented Generation (RAG) pipeline with SQL-based data siloing and CAG, and reduced our swarm to just two agents. The system finally worked.

Turns out, the model was never the problem. We needed a better harness. Now, with the current Claude Code leak, we can all see how much engineering goes into the harness around the model.

The real power comes from the extensive tools, memory systems, and guardrails. Here are five practical steps to focus on harness engineering:

  1. Define what a harness actually is. An agent equals a model plus a harness. The harness is every piece of code, memory system, and guardrail around the model.
  2. Use the filesystem as your primary state mechanism. Every production harness uses the filesystem for durable state instead of vector databases. For example, the Anthropic long-running agent pattern uses an initializer to create a progress file, which the coding agent reads and updates each session.
  3. Build feedback loops before adding more tools. Giving the model a way to verify its work improves quality by two to three times, as seen in the OpenCode LSP integration data. Feed linter output back into the planning loop so the agent can self-correct.
  4. Start with one agent. A single well-harnessed agent with memory outperforms multi-agent systems. Add orchestrator-worker patterns only when a single agent runs out of context space.
  5. Restrict tool access by role. Planning agents shouldn't have edit tools, and exploratory agents shouldn't modify code. Match your sandbox execution to your trust model.

The messy middle taught us hard lessons. LlamaIndex internal prompts changed on upgrades and broke everything. The MCP registry didn't add any value; it ended up being just API calls wrapped in useless abstractions.

RAG introduced a zigzag retrieval pattern with Optical Character Recognition (OCR), chunking, and embeddings. That was completely overkill since afterward we realized our data easily fit in a 64k token window. Simple SQL and CAG replaced the entire pipeline. So basically, the agent swarm was slow, expensive, and inaccurate.

TerminalBench 2.0 proved this approach. Modifying only the harness moved DeepAgent from outside the top 30 to the top 5.

What harness patterns have you found useful? What did you strip away to make your agents work better?

TL;DR: The model isn't the bottleneck, as the harness determines production success. Start with one agent, use the filesystem or a SQL database for state, build feedback loops, and restrict tool access.

r/Adulting barbiegirlll96

Follow me 😊

@barbiegirlll96

on

OF

r/30ROCK doomedcupcake

I received this today and thought you would like it.

r/LocalLLaMA AurtheraBooks

How do you Download palm2-demo - Please Help

Yesterday I downloaded the file but deleted it, today I'm cracking my head against a wall (proverbially), using the same links as yesterday and getting nowhere.
I'm on the You're working in paLM2-Demo page after creating the project but there's no links or downloads.
I'm loosing my mind because the links I used yesterday simply don't exist, are error 404 today or don't have a download option

SOLUTION - Edit - Thanks to the thinking machines I got the answer. Go to console.cloud on google and follow the following. I hope this helps

Alternative Download Methods

If the primary "Create" flow fails to trigger a download, you can often find the file manually:

  • Service Account Keys: Go to IAM & Admin > Service Accounts, click your account, go to the Keys tab, and select Add Key > Create new key > JSON.
  • OAuth Client Secrets: Navigate to the Credentials page, find your OAuth 2.0 Client ID, and click the Download JSON icon (downward arrow) on the far right.
r/painting sidbarnhoorn

Latest space watercolor painting!

r/SideProject Strange_Theory_4019

I’m building a "Yuka" but for sustainability (scanning barcodes to reveal the true environmental impact). Is this actually useful, or am I overthinking it?

Hey everyone,

Like many of you, I try to make eco-friendly choices at the supermarket. But honestly, I’m exhausted by greenwashing. It’s incredibly hard to know if a product is actually sustainable or if it just has a green leaf printed on the plastic bottle.

We have apps like Yuka that tell us if the ingredients are toxic for our bodies, but we don't have a quick way to know if the packaging is toxic for the planet.

So, I’m building an MVP to solve this, and I’d love your brutal, honest feedback before I write too much code.

Here is how it works:

  1. You scan a product’s barcode at the store.
  2. The app pulls data from a database and calculates a Sustainability Score (1-100) based on materials, recyclability, and carbon footprint.
  3. An AI translates the raw data into a short, easy-to-read explanation (e.g., "Score 40/100: The bottle is made of recycled PET, but the pump mechanism contains mixed metals making it unrecyclable").

I want to start with just ONE specific category of 50-100 products.

  1. Would you actually use an app like this while shopping?
  2. What product category should I start with? (e.g., Body wash/Shampoo, Dish soap, Beverages?)
  3. What is your biggest concern with this idea? (Inaccurate data? Taking too much time at the store?)

Please roast my idea. If it's terrible or if you think the data is too hard to get right, let me know. I prefer to fail now rather than 2 months from now! Thanks!

r/LocalLLaMA LH-Tech_AI

[New Model] - FaceGen v1 - generate 128px images of human faces with this GAN

Hey, r/LocalLLaMA !

I am back with a new model - another GAN!

It is called FaceGen v1 and it generates 128x128px of human faces.

This model is trained on the same architecture like my previous model from today - CatGen v2 (https://huggingface.co/LH-Tech-AI/CatGen-v2).

You can find the full source code, samples and the final model here: https://huggingface.co/LH-Tech-AI/FaceGen-v1

Look at this sample after epoch 250 (trained on my own RTX 5060 Ti 16GB):

https://preview.redd.it/ure1qrdtxrsg1.png?width=1146&format=png&auto=webp&s=43556d55dde7ac63c6671ce8c8ed7e26d3c6d138

Feedback is very welcome :D

Feel free to tell me, what you think about it.

r/personalfinance LikithAlluri

I am 25 year old and I wanted to start investing in ETF's. Please suggest me the most accurate ones where I can invest around 1000-2000usd per month.

I wanted the best of the best ETF's which I can rely on and start working on investing putting my head down. so that I am getting myself happy when looking at the return dividends. I am thinking VOO, QQQI, SCHD. I completely know that I might be wrong, but please give me you most hearted response.

r/leagueoflegends FireRespectOffical

Katarina help for a beginner

So I'm pretty new to the game so I would like someone to tell me some Katarina build/tips for mid lane, and even some general knowledge. I know that Katarina isn't someone a beginner should start with but even though I want to play with her. thanks in advance!

r/SideProject Miserable_Treacle208

Notesme - A notes taking app Simple, Selfhosted and Secure

NotesMe is a lightweight, open-source note-taking app with client-side encryption. No cloud, no tracking, no AI, just your notes on your server.

I spent a long time looking for a simple, self-hosted note-taking application. I tested several, but I was never satisfied: either too advanced, with too many features, or too simplistic, lacking folder management, backups, etc.

So I decided to create this application myself, with Claude's help. I'm happy with it, and I hope it will be useful to others. I think its main strengths are its simplicity, its lightweight design, and its note versioning system.

Secure - Simple - SelfHosted

Why you should try:

  • Clean
  • Simple
  • No AI
  • Notes Versionning
  • Share functionnality
  • Export notes
  • Easy to Backup
  • App mode on smartphone and Windows

You will find a live demo to test :)

https://notesme.cloud/

Regards !

r/Seattle Weekly_War_1374

Dear Microsoft, please add an office on the Seattle side.

Seattle employees driving to the Eastside impacts everyone's traffic. Please and thank you.

r/explainlikeimfive roger_ramjett

ELI5: When the price of oil goes up, who is getting the money?

ELI5: When a commodity like crude oil goes up in price, who gets the extra money? I don't think that the workers on the oil rig get a raise, so who gets the money?

r/homeassistant jklo5020

Unifi Dream Router 5G Max LED

Is anyone familiar with a way to integrated the UDR 5G Max‘s LED light or display into Home Assistant?

Every Unifi related plugin I’ve tried works for sites of mine with standalone APs but nothing for an AIO like the UDR.

r/30ROCK BettyCrunker

Well okey-dokey!

even after 16 million rewatches, I still laugh out loud at this scene every time

r/LocalLLaMA mittah_rogers

Do we actually need MCP/tool abstractions for enterprise agents?

There’s been a lot of recent work on agent design around:

  • browser-based / computer-use agents
  • tool abstractions (e.g., MCP-style systems)

We wanted to evaluate a simpler baseline:
how far can a general-purpose terminal-based agent go when APIs are available?

In a recent paper, we compare three approaches:

  • terminal (CLI) agents
  • web agents
  • tool-based agents

across enterprise-style tasks (ServiceNow, GitLab, ERP-like systems), using API-verified evaluation.

The terminal setup is intentionally minimal:

  • shell access
  • filesystem
  • ability to call platform APIs (e.g., via scripts / curl)

No predefined tools or curated action schemas.

Some observations:

  • Terminal agents are competitive with (and often match or exceed) the other approaches on task success
  • They are generally more cost-efficient, largely due to shorter interaction trajectories
  • Allowing the agent to write and reuse simple “skills” (scripts, notes about APIs) helps reduce repeated exploration over time

At a high level, this suggests that in API-rich environments, a general coding agent can often compose the required actions without additional abstraction layers.

There are still clear limitations, especially for workflows that are only accessible through a UI, where web or hybrid approaches remain useful.

Curious how others think about this tradeoff:

  • When do tool abstractions actually help vs constrain?
  • Are browser agents mainly useful as a fallback for missing APIs?
r/ChatGPT HerbChii

bro is cooked 💀

r/SideProject nlouka

Built my first iOS game with zero Swift experience. Shipped it in ~2 weeks.

So I had this dumb idea: what if every time you died in a mobile game, it automatically recorded your screen so you could share the embarrassing moment? That became Replay Rage.

I have zero iOS background. Started completely from scratch, figured out Swift and SpriteKit as I went. The hardest part wasn't the game mechanics — it was getting the screen recording to spit out an actual video file I could hand off to Instagram/TikTok instead of just showing a preview.

The game itself is stupidly simple. Flappy Bird vibes. You die, it captures your death, and you can post it directly to Stories or TikTok with one tap. Global leaderboard too. Gave it to my son to play with his HS friends, and they seem to like it.

Honest review: it's rough around the edges, but it works, and I'm weirdly proud of it. Would love any feedback.

https://apps.apple.com/app/replay-rage/id6761093304

r/SideProject antihero11

I built an all-in-one SEO tool to make SEO fast and simple

Hi everyone,

I’ve been working in SEO for 15 years, and I’ve realized something simple:
most people who own websites don’t really understand SEO, hiring a professional is often too expensive, and most tools are hard to use.

With that in mind, I built Wisseo to make SEO easier and, above all, actionable.

How is it different from other tools?

1. Know if it’s worth it before you start
It doesn’t just give you keywords with volume and difficulty. It analyzes the top 10 results on Google and tells you whether your current domain actually has a chance to compete — or if you’re likely wasting your time.

2. Proper Local SEO tools
NAP audit, local grid, and review analysis.

3. Reputation management
It scans Google and Google News results, detects relevant mentions, and prioritizes them.
If there’s a negative or false review, the AI analyzes the case (not a substitute for a lawyer) and generates a draft email to request its removal.

4. AI visibility (GEO)
It monitors whether tools like ChatGPT, Gemini, or AI Overviews are mentioning your brand or your competitors.

5. Content that actually makes sense
It generates or analyzes content based on what’s already ranking: automatic internal/external linking + optimized images (I still recommend adding your personal touch or real experience).

You can try it for free (no credit card) and get 100 credits.

With that, you can:

  • Audit your website
  • Analyze keywords
  • Track rankings
  • Or review your full local SEO setup

I believe even the free version can be useful, even if it’s just to:

  • Spy on competitors
  • Validate ideas before writing
  • Or find keyword opportunities

Disclaimer:
I’m not a developer. I’ve tested the tool quite a lot, but it has many features.
If something doesn’t make sense or doesn’t work as expected, feel free to tell me, I’d really appreciate it.

Feedback is the most important thing for me right now.

r/DunderMifflin hentai_gifmodarefg

What are your theories for why Michael pretended to Jan that he and Pam dated?

i should clarify: its obvious that he did it to make Jan jealous, but in what circumstance did he do it. Surely he didn't just bring it up out of the blue but rather in reaction to something

my theory is that after Jan chastised Michael for not washing his hands after going to the bathroom he said "well when I dated Pam she didn't care. and she's hotter than you."

r/Strava Yakolev

Why is April not a 12 point Star.

r/Adulting xkxind

Mother's Day is 39 days away and I just realized I have no idea what my mom actually wants. This is adulting failing in real time

Every year I say "I'll figure out what to get Mom" and every year I panic-buy something generic the week before. A candle. A mug. Last year I got her a gift card and she was visibly disappointed but too kind to say anything.

The problem isn't that I don't love her. It's that I have no system. I've never sat down and actually thought about what she loves, what she's mentioned wanting, what would make her feel seen.

So this year I'm changing that.

I've been building a running list of what everyone in my life actually cares about — not just my mom, but my dad, my sister, my best friend. I write it down when they mention something. I note what they're going through right now. I track when important dates are coming up so I have weeks to actually plan something meaningful.

I use a free app called **BondBox** for this: https://play.google.com/store/apps/details?id=com.bondbox.app

It's designed exactly for this — keeping notes on the people you love so you can actually show up for them.

Mother's Day is May 11. If you don't have a plan yet, now's the time. Not the day before.

r/comfyui Cold-Reality3274

Style Transfer

Hey guys, i need to do style transfers and do not get good results with the options i have tried. As of April 2024 what are the best Style Transfer options in your opinion?

r/LocalLLaMA Repulsive-Mall-2665

Why does Qwen struggle so much with coding SVGs?

r/TheWayWeWere astralnausea

My grandparents on their wedding day (1953)

r/geography Girhinomofe

Are there any resources to purchase a professional, detailed, high quality world map in a digital format?

I work at a shop with large format printing capabilities on adhesive backed films and thick, professional grade foamcore substrates.

I was thinking of making a large scale world map to add pins for all the travels of my wife and I, but I’d like a map that is truly a professional creation, à la the hard copy maps that Rand McNally makes. (Not looking to side hustle this; just a one off for myself)

The McNally maps are lovely and inexpensive, but adhering one to the foamcore material is less than desireable as I’d have to use a spray adhesive.

Absolutely fine with paying for a digital map download, but not finding resources for the level of detail I’m interested in. Has anyone ever seen something like this available online?

r/OldSchoolCool barefoot_yank

1970...old school, but timely

r/space theneiljohnson

Artemis Mission Tracker and Live Map

Hi everyone, just thought i'd mention that Leo and I added Artemis tracking to issinfo! You can select Artemis I too and scrub through the timeline for both missions.

https://issinfo.net/artemis.html

r/personalfinance Woodleaguelad

Drowning in Student Loans

Hello,

I am looking for a kind-hearted soul to please explain how to manage my financial situation! I am 33, in medical residency to become a physician with 1 year remaining. In my area, local pay is around 290-320K. As of today, my loans are 354K, accruing about 6% interest. My soon-to-be wife has a rather low paying job, around 20-40K annually depending on what opportunities present themselves during the year.

We want to have children, and are already kind of "old" in terms of that aspect of life, so feeling the squeeze of that on top of the burdensome amount of loans is stressing me the hell out. We also live in a 2 BR apartment that is 800 sq ft and would like a house.

Is it feasible to pay these off somehow, afford a place to live, and have children (2, hopefully)? I had a vision that making this much money would afford quite a nice lifestyle, but the reality of the loans is keeping me up at night and making me feel like having a child or a house is out of reach. The generations above us, ie her parents and my grandparents have a decent amount of money themselves, but asking for a hand-out with a potential 300K salary on the horizon also seems absurd.

If someone could either validate that what I am feeling is the reality or let me know that I am entirely misguided and it will be smooth sailing and smack some sense into me, that'd be great. All advice and help is very much welcome!

Thanks

r/PhotoshopRequest One_Loan_2439

Looking for a few edits please 10$

Looking for these edits:

- please fix the lighting. make it glowier/brighter & a sunny day but please keep it realistic

- make my hair all black (keep the stype) but also make it a smidgen longer

- put me in one of these outfits please

- & make my body a little (realistically) curvier & slimmer to make the outfit more flattering :)

thank you!! please no ai!

r/TwoSentenceHorror Riddle-Maker

As my kidnapper cut off a second finger, I knew the ransom still wouldn't come.

My husband had always wanted me to stop playing piano.

r/DunderMifflin riegspsych325

Looking for a photoshopped picture of Toby grabbing Jim’s leg

I saw this jokey meme posted here before and I am looking for it again. It’s basically what the title says, a shopped pic of Toby grabbing Jim’s leg (as Pam was airbrushed out)

r/ClaudeAI Mirandah333

Is there a way to limit how many sources Claude searches?

Sometimes it goes overboard with the sources and search time, and that affects my credits, right?

r/geography coltenworld

What this 35°53'24.0"N 117°40'16.0"W

r/ClaudeAI lagoJohn

When to compact?

Do you usually wait to compact your conversation or do you sometimes compact before 100%?

r/comfyui IntelligentEngine624

Amazed with LTX2.3!

I am using LTX2.3 with a i9, 128gb of DDR4 and a 3090 24gb running Comfy install on Linux and I am amazed at the results I am getting. One caveat, IF there is one person in the scene. As soon as I add 2-3 people it starts doing weird things like adding extra people or doing strange zooming. I need to add negative prompt logic to the workflow. I made this music video with LTX 2.3 15 seconds at a time. https://www.tiktok.com/t/ZTk6QtQSR/

r/Damnthatsinteresting TJ_Fox

James Neal Hollingworth (left) who managed psychedelic rock bands and founded a Freedom School in the late '60s, became paraplegic after a car accident and then started an underground bodyguard service protecting women from abusers. He coined the famous phrase "Courage is not the absence of fear".

r/ClaudeAI Odd_Werewolf_4478

Anyone else getting tired of reviewing big AI-assisted changes with no real trace of how they were produced?

One thing that's been bothering me lately is reviewing large AI-assisted changes when there's no record of what the model was actually asked to do.

You get a big diff, maybe a summary, maybe some vague notes, but not much else.

So as a reviewer you're left guessing:

  • what was the actual task?
  • what constraints or instructions were given?
  • what alternatives were considered?
  • what was actually checked vs just accepted?
  • why did it end up in this shape?

At that point the hard part isn't "reviewing code". It's reconstructing the missing context.

That's basically the problem I've been thinking about.

I don't think the answer is "don't use AI". I think the missing piece is better records around AI work itself — what was asked, what happened, what evidence exists, and what should be reviewed.

I've been exploring that through a side project called Geas, but the bigger question for me is whether other people are running into the same thing.

r/ChatGPT Fujita_Seiko

Chattgp microfoon down?

down?

r/OldSchoolCool BabyDollFaceXo

Carly Simon 1971

r/ClaudeAI pablooliva

Spec Driven Development Fans, You'll Love sdd-flow

I built a Claude Code skill that orchestrates the full SDD cycle through subagents — research, adversarial review, spec, implementation, code review, critical review — all from a single slash command. Used it to ship 6 features in 2 days for a GDPR tool. The adversarial review phases are where it gets interesting: separate agents tearing apart the research and spec before any code gets written. Skill and plugin are both open source.

r/mildlyinteresting banana_owner

Calendar for 50 years

r/leagueoflegends Bredjacafo

What's is going on with dragons in swift play?

They die in one hit and don't give you anything. Seems like not a very thoughtful change considering they just released the rework of Shyvana...who needs to stack dragons... unless I'm missing something (I hope I am)

r/LocalLLaMA El_90

Large GGUF works in bash, but not llama-swap

I've spend days on this but I give up! I've even tried chatgpt and gemini, but it goes in circles.

unsloth_Qwen3.5-122B-A10B-GGUF_Q5_K_M will load when I run in Bash, but crashes using Llama-swap. I suspect this is path/env variables/LD_LIBRARY_PATH, but I've tried so many combinations.

# About

Strix halo, 128GB, using GTT for 122GB usable memory

rocm 7.1.1

llama-swap 190 (I've tried other versions but rolled back to this, nothing in release notes suggests it would be better?)

llama.cpp cmake: DAMDGPU_TARGETS="gfx1151"

# Works fantastic - Bash

# llama-server --host 0.0.0.0 --port 8080 -m /../unsloth_Qwen3.5-122B-A10B-GGUF_Q5_K_M_Qwen3.5-122B-A10B-Q5_K_M-00001-of-00003.gguf -ctk bf16 -ctv bf16 -ngl 999 -fa on -c 65536 -b 2048 -ub 1024 --no-mmap --log-file /tmp/llamacpp.log --parallel 1

root@llamacpprocm:/root/.cache/llama.cpp# export

declare -x OLDPWD="/root/.cache/llama.cpp"

declare -x PATH="/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"

declare -x PWD="/root/.cache/llama.cpp"

declare -x SHLVL="1"

declare -x TERM="linux"

declare -x container="lxc"

# Fails - llama-swap

It fails during model load, it gets half way through the loading dots, then just restarts continuously. No error in dmesg -w, nothing in verbose logging.

llama-swap.service

[Unit]

Description=llama-swap proxy server

After=network.target

[Service]

Type=simple

WorkingDirectory=/etc/llama-swap

ExecStart=/usr/local/bin/llama-swap --config /etc/llama-swap/config.yaml --listen 0.0.0.0:8080

Restart=always

RestartSec=5

# Core Hardware Overrides

Environment="HSA_OVERRIDE_GFX_VERSION=11.5.1" ## NOT 11.0.0

Environment="HSA_ENABLE_SDMA=0"

# Memory & Performance Tuning

Environment="HIP_FORCE_DEV_KERNELS=1"

Environment="GPU_MAX_HEAP_SIZE=100"

Environment="LD_LIBRARY_PATH=/opt/rocm/lib:/opt/rocm/lib64"

[Install]

WantedBy=multi-user.target

# head /etc/llama-swap/config.yaml -n 20

# yaml-language-server: $schema=https://raw.githubusercontent.com/mostlygeek/llama-swap/refs/heads/main/config-schema.json

healthCheckTimeout: 200

logToStdout: "proxy"

startPort: 10001

sendLoadingState: true

# This hook runs BEFORE any model starts, clearing RAM to prevent OOM

hooks:

before_load:

- shell: "sudo sync; echo 3 | sudo tee /proc/sys/vm/drop_caches"

- shell: "export HSA_OVERRIDE_GFX_VERSION=11.5.1 ; "

Any insights are appreciated !

r/SideProject Inevitable_Sale_7416

lazy to remember your script for long meetings ? we got you covered

https://reddit.com/link/1sagj7d/video/xw5nq6ahwrsg1/player

eye contact changes a lot in presentations and meetings , it makes you look way more confident and telling your scripts smoothly without fumbling in between makes it even better. cuenotch basically scrolls with your voice , you can control the scripts with the arrows in your keyboard. The teleprompter stays in your macbooks notch and remains hidden when your screen is shared for others. you can also add scripts for individual slides if your doing ppt presentations , cuenotch has a great freemodel so guys dont shy away from downloading it , meet you guys in the app store --> cuenotch , also guys im super open to suggestions and feedback

r/SideProject SeaworthinessAny35

I made a word game called Thruzzle: It's about uncovering hidden words in a scrambled grid.

Hi, I’ve been building a word game as a side project. where the goal is to uncover three hidden words in a scrambled letter grid. You can set the word length, number of clues, and even the language, making it as easy or challenging as you like.

It’s completely free to play, with no sign-ups or ads: https://www.thruzzle.co.uk

I would love to hear what you think!

r/meme Stock_Crazy6759

This happens Monday-Friday 😭

r/LiveFromNewYork Falconerinthehud

The Situation Room: Tiger Woods' Accidents - SNL

r/n8n paranoid-alkaloid

export table with location to GPX and other formats?

Hi. I will be posting this both to r/baserow and r/n8n, I hope it's okay.

I am using self-hosted Baserow. I have a table with a number of locations with lat/lon coordinates and a few other fields. I want to export that table to GPX and other formats. I have about 1000 rows in my locations table.

The system should be as easy and "bullet-proof" as possible because this project is for a non-profit I am contributing to. Ideally exports should be able to tweaked by area, by location type, through a simple "app"/form,, but that's okay if I have to pre-prepare a few targeted versions instead (perhaps the Baserow export of 1000 rows would take too long for on-demand exports?).

I believe that I will need to deploy n8n as Baserow alone won't suffice. That's fine.

I am seeing no GPX module for n8n, however I know that I can relatively easily create GPX through scripting. I have done that in Python before but perhaps that would be an opportunity to create an n8n module for it.

Other than or in addition to GPX, I may want to automate exports to GeoJSON, to Google Maps, to uMap.

Do you have any advice for me? Do you feel that Baserow and n8n are good choices for this? (Baserow seems like a great

Thank you.

r/AI_Agents buntyshah2020

If you use Gemini for research in your agentic workflows, there's no native way to get that data out — so I built one

A common pattern in agentic systems: use Gemini (especially Deep Research) as a research/synthesis step, then pipe the output into downstream agents or processing layers.

The problem: Gemini has zero native export. After a Deep Research session, all that structured knowledge — multi-source synthesis, inline citations, numbered references — is locked in the browser. There's no API, no export button, no way to get it as JSON or structured text without copy-pasting and losing all the formatting and citation structure.

I built a Chrome extension called Gemini Export Studio to fix this specifically. For agent/pipeline use cases, the key exports are:

- JSON — full structured conversation with metadata, turn counts, timestamps, and source citation arrays. Ready to pass to any downstream process.

- CSV — each turn as a row with role/content/metadata columns. Import directly into pandas, feed into an embedding pipeline, or use as training data.

- Markdown — clean output with heading hierarchy and code blocks intact, useful as context documents for agents

Deep Research exports specifically preserve all the source URLs and citation structure inline, which is the part that matters most when you're using Gemini research as grounding context.

Everything runs 100% locally — no server, no API key, DOM read in-browser and export generated client-side.

Link in comments per sub rules. Happy to answer questions about the extraction approach or the data structure of the JSON output.

r/TwoSentenceHorror Mothmans_

Everyday he come's a step closer to making out with his Living wife

hoping this Time she won't feel Cold to the Touch

(actual text my wife sent me an explanation in comments)

r/toastme arivire_

College is rough 🫠🫩

r/AbstractArt Used2beasian

"Beneath the Flowers" Acrylic paint/ink on Wood, March 2026.

"Beneath the Flowers"

24x24

r/Rag buntyshah2020

How I export Gemini Deep Research sessions to structured JSON for RAG pipelines — citations preserved, 100% local

One friction point I kept hitting when building RAG pipelines: Gemini's Deep Research outputs are genuinely excellent for domain-specific corpora, but there's no native way to get them out in structured format.

The Deep Research feature produces rich outputs — multi-source synthesis, inline citations, numbered references — but if you try to copy-paste it, the citation structure collapses entirely. You lose the source URLs, the reference mapping, all of it.

So I built Gemini Export Studio, a Chrome extension that exports any Gemini conversation (including Deep Research) to:

- JSON — full structured data with turn metadata, timestamps, and source citation array preserved

- CSV — each turn as a row, importable directly into pandas or any data pipeline

- Markdown — clean .md with heading hierarchy and fenced code blocks intact

- PDF, Plain Text, PNG

For RAG use cases specifically, the JSON export is the most useful: you get the full conversation turns, the source URLs from Deep Research inline, and clean text you can chunk and embed without heavy preprocessing.

All processing is 100% local — no server, no API calls from your data, DOM is read in-browser and the export is generated client-side.

Extension: https://chromewebstore.google.com/detail/gemini-export-studio/oondabmhecdagnndhjhgnhhhnninpagc

Landing page: https://buntys2010.github.io/Gemini-Export-Studio/

Curious if others here use Gemini Deep Research as a data source for RAG corpora, and what your current extraction workflow looks like.

r/AI_Agents Odd_Television_6382

Modular Skill Creation Paradigm

I am building very complex skills with references, subagents, and lots of different files. I realize that it's hard to maintain these long multi-file markdowns, with some information getting repeated or contradicting itself. Any ideas on how to organize these better, or being able to work with markdown files in a more modular way. I tried jinja templates but not sure it's what I am looking for.

r/goodnews Icy_Leadership5933

Just had a massage

r/brooklynninenine DontBeSuspicious_00

Never go can to pan!

Everytime I'm cooking and I add seasoning I hear Boyle yelling at me, "You never go can to pan!"

What are your quote triggers?

r/metaldetecting New_Significance_787

Found in WWII forest

After finding some shell casings, nothing unusual, I found this. I am not sure what this is, but out of precaution I left it untouched next to the hole I dug it out from. Authorities have been called. Never stumbled upon UXO, so better safe than sorry!

SortedFor.me